var/home/core/zuul-output/0000755000175000017500000000000015153033746014534 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015153042046015471 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000171416015153041765020267 0ustar corecoreCikubelet.log_o[;r)Br'o -n(!9t%Cs7}g/غIs,r.k9Gfͅt|ᳲi.߷;U/;Yw?.y7W޾n^X?|𒆷7̗8zTY\].f}嗷ovϷw_>on3cvX~egQBeH,nWb m/m}*L~AzHev_uαHJ2E$(Ͽ|/+k*z>p R⥑gF)49)(oՈ7_k0m^p9PneQn͂YEeeɹ ^ʙ|ʕ0MۂAraZR׏!#Šgv cXk?`;'`&R7߿YKS'owHF6":=3Ȑ 3xҝd){Ts}cZ%BdARO#-o"D"ޮrFg4" 0ʡPBU[fi;dYuȓE$ޠYx3JtQJFjc 9G8MOY:GTMce0hTYF;B6@ c$Ⱦ֠N+fD>%vz_. o~I|3j dF{ "IΩ?PF~J~ ` 17ׅwڋًM)$Fiqw7Gt<LJ|iZ~hal\t2Hgb*t--ߗ|Hp(-J C?>:zR{܃ lM6_Oފ?O1nԝG?ƥF%QV5pDVHwԡ/.2h{qۀK8yUOdssdMvw`21ɻ]/ƛ"@8(PN_,_0;_x+Vy<h\dN9:bġ7 -Pwȹl;M@n̞Qj_P\ Q]GcPN;e7Vtś98m1<:|a+.:a4nՒ,]LF0);I$>ga5"f[B[fhTRk׿kb8_b|r wFuRI%T۩Ѭza\_/2vw>- MR9z_Z;57xh|_/CWuU%v[_((G yMi@'3Pmz8~Y >hl%}Р`sMC77Aztԝp ,}Nptt%q6& ND lM;ָPZGa(X(2*91n,50/mx'})')SĔv}S%xhRe)a@r AF' ]J)ӨbqMWNjʵ2PK-guZZg !M)a(!H/?R?Q~}% ;]/ľv%T&hoP~(*טj=dߛ_SRzSa™:']*}EXɧM<@:jʨΨrPE%NT&1H>g":ͨ ҄v`tYoTq&OzcP_k(PJ'ήYXFgGہħkIM*򸆔l=q VJީ#b8&RgX2qBMoN w1ђZGd m 2P/Ɛ!" aGd;0RZ+ 9O5KiPc7CDG.b~?|ђP? -8%JNIt"`HP!]ZrͰ4j8!*(jPcǷ!)'xmv>!0[r_G{j 6JYǹ>zs;tc.mctie:x&"bR4S uV8/0%X8Ua0NET݃jYAT` &AD]Ax95mvXYs"(A+/_+*{b }@UP*5ì"M|܊W7|}N{mL=dC' =MS2[3(/hoj$=Zm Mlh>P>Qwf8*c4˥Ęk(+,«.c%_~&^%80=1Jgͤ39(&ʤdH0Ζ@.!)CGtGLS0l/LKcQ.os2% t)Eh~2p cL1%'4-1a_`[Zz㧦|k˭c ĚOρ_} Ewt3th?tvͪ{~;J0= |JUԍ;Iw}/9nh7l%>'ct Հ}a>-:(QxPyA Z UcÖgڌ:8cΗ|U1,-N9 dI [@3YN%:ò6PT:”QVay 77ĐrX(K&Y5+$wL#ɽ 4d-bbdAJ?:P>n^2] e}gjFX@&avF묇cTy^}m .Ŏ7Uֻ󂊹P-\!3^.Y9[XԦo Έ')Ji.VՕH4~)(kKC&3[n )ܗKj/jUSsȕD $([LH%xa1yrOpY]Q4`Iz_*2coT'ƟlQ.Ff!bpRw@\6"yr+i37Z_j*YLfnYJ~Z~okJX ?A?gU3U;,ד1t7lJ#wՆ;I|p"+I4ˬZcն a.1wXhxDI:;.^m9W_c.4z+ϟMn?!ԫ5H&=JkܓhkB\LQ"<LxeLo4l_m24^3.{oɼʪ~75/nQ?s d|pxu\uw?=QR -Mݞίk@Pc n1æ*m$=4Dbs+J \EƄզ}@۶(ߐ/ۼ𹫘qݎt7Ym݃|M$ 6.x5 TMXbXj-P\jА޴y$j`ROA"EkuS#q * CƂ lu" yo6"3껝I~flQ~NCBX`]ڦÞhkXO _-Qy2$?T3ͤEZ긊mۘ$XD.bͮW`AީClСw5/lbl[N*t*@56."D/< {Dۥ sLxZn$N(lYiV =?_e^0)?]{ @| 6+#gPX>Bk2_@L `CZ?z3~ }[ tŪ)۲-9ֆP}b&x Uhm._O 4m6^^osVЦ+*@5Fˢg'!>$]0 5_glg}릅h:@61Xv` 5DFnx ˭jCtu,R|ۯG8`&ו:ݓ3<:~iXN9`2ŦzhѤ^ MW`c?&d.'[\]}7A[?~R6*.9t,綨 3 6DFe^u; +֡X< paan}7ftJ^%0\?mg5k][ip4@]p6Uu|܀|Kx6خQU2KTǺ.ȕPQVzWuk{n#NWj8+\[ ?yiI~fs[:.۽ '5nWppH? 8>X+m7_Z`V j[ s3nϏT=1:T <= pDCm3-b _F(/f<8sl, 0۬Z"X.~b٦G3TE.֣eմi<~ik[m9뀥!cNIl8y$~\T B "2j*ҕ;ێIs ɛqQQKY`\ +\0(FęRQ hN œ@n|Vo|6 8~J[,o%l%!%tyNO}}=ʬ-'vlQ]m"ifӠ1˟ud9)˔~BѤ]һS8]uBi( Ql{]UcLxٻa,2r(#'CDd2݄kTxn@v7^58þ Ţ&VY+yn~F8I !6WB3C%X)ybLFB%X2U6vw8uUF+X|YukXxVO(+gIQp؎Z{TcR@MSRδ~+1æ|mq՗5$B᲋eY(|*磎\Dži`dZe j'V!Mu@ KV{XץF .Jg< ƜINs:b zĄu3=Az4 u5'og^s7`Rzu-anOIq;6z( rx߅ euPvIɦ7聀t>G;_H;2ʗ6 h6QװxmR JQUbTP2j˔Ni)C)HKE"$ӝ!@2<Bq 2oh80,kNA7,?ע|tC3.㤣TiHEIǢƅaeGF$ u2`d)/-st{E1kٌS*#¦۵_Vu3ЩpRIDr/TxF8g4sѓ{%w .ʕ+84ztT:eEK[[;0(1Q@ET0>@wY)aL5ׄӫ A^%f+[`sb˟(]m`F3 W((!5F-9]dDqL&RΖd}})7 k11 K ;%v'_3 dG8d t#MTU']h7^)O>?~?_ȿM4ə#a&Xi`O}6a-xm`8@;of,![0-7 4f kUy:M֖Esa./zʕy[/ݩqz2¼&'QxJE{cZ7C:?pM z*"#窾+ HsOt۩%͟A498SwWv|jNQ=-[ӓI]iSCQ&s~In/SZ % 'I Ƿ$#stV \'xMgaSZNg8>e!^f%cYr]qs:"̊;isXa]d+"v=x7p.fZCg_Ys;pE&\U}ܫSh])qKYAـhhdEnU14&G * QIQs;rԩ.k83֖8Muqu_48dHܥlWW q>fu6+'}xu\Veelz`Zbym gp8펠ˋֆ:1IC8qٞ\vXçL ]X/r}7O}Wh,h ;RQ=]u00yiC۔I^3!?H6iUH:ô 4P$rT`%2Aq-֢׍qt=@x#~0)p# ы9'iri]ͪ/@繁qVGCڤr,DihB ,m 9 _$q3= A$IC"6g^4e`Xo(D*6"^eTh'4xpFڜe'fVQ7~'c L^ԯwIڣA.}H;Ë*׬=`^ 9]r鐃 -Dfi2|QwZk‹u^6DQ1&H凎c!n[mi3)WfsF:M"uҷs.1!뾧1%s,hQs|hx̗3%*v9(I;:'>uQ+v)vR/egBhAAdh]4H:nV$tHI98/)=mͭ ڐn}}~ק?g_6WĩDRc0]rY9'z .(jHI :{HG}HDN`h7@{jnE#[dz;n#y 9D*A$$"^)dVQ.(rO6ӟZw_Ȣaޒu'- ^_,G;U\cAAz7EtlLuoXuA}bT2H_*kIG?S(קjhg 5EF5uKkBYx-qCfqsn[?_r=V:х@mfVg,w}QJUtesYyt7Yr+"*DtO/o۷~|hw^5wE of7cꃱ.)7.u/}tPTGc 5tW> l/`I~>|灹mQ$>N |gZ ͜IH[RNOMTq~g d0/0Љ!yB.hH׽;}VLGp3I#8'xal&Ȑc$ d7?K6xAH1H#:f _tŒ^ hgiNas*@K{7tH*t쬆Ny497ͩ KVsVokwW&4*H'\ d$]Vmr달v9dB.bq:__xW|1=6 R3y^ E#LB ZaZd1,]ןkznxtK|v+`VZ3JϧC^|/{ś}r3 >6׳oƄ%VDSWn 0,qh! E-Z%ܹpU:&&fX+EǬ.ťqpNZܗÅxjsD|[,_4EqgMƒK6f/FXJRF>i XʽAQGwG%mgo 恤hˍJ_SgskwI\t`ﶘ080ƱQŀllKX@116fqo>NrU Ѣ9*|ãeeH7.z!<7zG4p9tV|̢T`˖E ;;,tTaIUle*$!>*mBA2,gJIn_kSz)JC]?X(OPJS3.}clݨ{e!MB,cB߮4af祋,1/_xq=fBRO0P'֫-kbM6Apw,GO2}MGK'#+սE^dˋf6Y bQEz}eҏnr_ ^O^W zw~Ȳ=sXअy{E|!4ӥ2 ]8â6 U`V%`!c%؎ʨTzrKh! c.}.D>)d_ 8rcu,wf2?Ǡ*_lDn}rauyFp*ɨ:UiM2r:9ct X1lmĪ o玓,R%!`hGT LYF#g<cm${|Xdu4tmtїUJ\~dc0KcMlf2?mμQ ߉J4WrSHTdp"ӹ'cJq2zPlX̯.0H!ND@UapVoGڧD5>H]f@!=߸2V%Z 0"G4ȇʩ@]>Y$ًF_Mm_Tt)ib+q&EXFu򾬳ǝ/RS>r,C2NfOjpcm{Ll9vQOT>9U;])>6JdbXԠ `Z#_+D[7IIjJɟUh ҙ"`"a ߒ"G̾H`6yiCk(OA/$ ^%K^+(Vr[RR1"u4A.1X0=7f/"(o9/L1X{]q`Ȝ/; 9a>E)XOS K9mUxBa"'4T[Jl /K/9,rlCAj_TiǘP,:4F%_0E5IE'rX-|_W8ʐ/=ӹjhO%>| :S Px„*3_y.g9| ;b`w NtZtc> ײ1KĴ{3Gl& KT1ZWX8?C]~We$9; -.D087?1a@P5B,c}jcGȱ WW/ @a#LA4.ٹ^XڋXٝ:^Izq. ٽƎDn6ٹBc5Lt;3#i3RAٽ9| cbpcTfp> 6L/_x 'ۙz7~w~);qU9GDT! 6]c_:VlnEUdn6UˇKU;V`JUݵޙEO[)ܶCy*8¢/[cչjx&? ՃJȚ9!j[~[' "ssTV2i sLq>z@JM->=@NỲ\쀜*/) ̞r21.y? bO]3?C!yw3ޯL_Su>o>&lrw&i"< :]_<<7U_~z5є/rfn͝MLmc 6&)e+n7cyy{_~궼07R7wPuqpqo{ߟ+[w_uOq?u-|?WS_tOq?Eu-L_p?Cz .e ϿO*3 `Ђ6a-`kIf-s,RL-R`1eL~dپ&+IhYRczr?㐟,v~,b6)up)3K,RLW"Qd9JgT\1f3@Kh% a4x,kA k ^d kYj5Ah𚄓vXZhX1xҖ51Y +Id ZZ\C| fD>hB֡#-$+Jpሟ,Cg:6 3 xH "}C[`ӨOAFn5ʬLHϰ:N@VcyBI#Dr. "h hg ۃm-qu>V&൘ G7qi#^tҒ[JI!{q*lrD܇Gk@;oI<5xZ4xM"؇'k!>V|lk'{d+ :sXӄc)?W`*|\v aVT0"tMًcΒVz]T.C$cEp._0M`AlF̤@U' u,—rw=3}resLV&ԙy=Ejl1#XX۾;R;+[$4pjfљ lݍ3)`xvcZRT\%fNV Q)nsX }plMa~;Wi+f{v%Ζ/K 8WPll{f_WJ|8(A ä>nl"jF;/-R9~ {^'##AA:s`uih F% [U۴"qkjXS~+(f?TT)*qy+QR"tJ8۷)'3J1>pnVGITq3J&J0CQ v&P_񾅶X/)T/ϧ+GJzApU]<:Yn\~%&58IS)`0効<9ViCbw!bX%E+o*ƾtNU*v-zߞϢ +4 {e6J697@28MZXc Ub+A_Aܲ'SoO1ۀS`*f'r[8ݝYvjҩJ;}]|Bޙǖߔ 3\ a-`slԵ怕e7ːزoW|A\Qu&'9~ l|`pΕ [Q =r#vQu0 M.1%]vRat'IIc(Irw~Z"+A<sX4*X FVGA<^^7 vq&EwQű:؁6y\QbR9GuB/S5^fa;N(hz)}_vq@nu@$_DVH|08W12e_ʿd{xlzUܝlNDU j>zƖݗ&!jC`@ qэ-V Rt2m%K6dX)"]lj齔{oY:8VmS!:Wh#O0} :OVGL.xllT_oqqqLec2p;Ndck[ Rh6T#0H Q}ppS@ώ@#gƖ8sѹ e^ CZLu+."T#yrHhlكʼE-X'I^=bKߙԘ1"+< gb`[c1髰?(o$[eR6uOœ-m~)-&>883\6y 8V -qrG]~.3jsqY~ sjZ+9[rAJsT=~#02ݬf¸9Xe>sY~ ae9} x* zjC.5Wg󵸊y!1U:pU!ƔCm-7^w]斻~[hW$k sE0ڊSq:+EKٕ|dvvjjy6 æ/ML-yz,ZlQ^oAn-})xǺǍ--qcl:WLg ӁvJ[ǧc~Of+8qpçco#rCtKӫce0!Y-+cxMK-H_2:Uu*corD~@N`#m~R:ߙ歼!IZ5>H;0ޤ:\Tq]_\_>e˲\oUQ\Wߋ47WwߋKpwSSۘF,nC.\UߋoVEuY]^VW0R=<ު˜˻ x}[ێ'|;c^ M7 >5\-> m-8NJ\ALd!>_:h/NAC;?_ξqĎ6xMY(=ͯl~l8V0٨T zL{Ac:&$ ^CpH*DW\r2aR|=(L X1|wrO_g ux1^^V2޲jMi^b``Q#dBxV#NBk1;DAV$"*1]Y~ d->'I`cZF_q|lJUnzC ٪Fn[!88ֻY0L hA'FIҘUg{l(}UL6s@`_mR$<ߓkU }ՅG$ QJW|GEWnWھ)WEd"HZ'Q2hʪ^7wgYr< <$}U 75{q*SZGeJ1,y N{O b_&,6?¢`QKHxɠDw d__ͯ Az3/|2q<^2=jiBgB/)`~rC%J NI;:%e#ON0C?24s3X ^zhxY\̍LW/+`V$]䍀3eqeLZ^W̏vl)Ba8wN tHX7fX6o }m1y\z9vtWk?/yJ4_Əm.,VA(l< w{l5+ sq}; #iʋtRku Io s3^OcC7Kݸ4KݹK:,׎K4fqX{-(| ~c ]oakB~~i'x/}e0!~c8~d/|%_*t_^{2Kc,šB,?f+KRd2ٽ hL/aI]_缩Y,Y5zDF^`R~(c0,ƅS+48P@&az]zqWQX(~g5|=^.VyP,fY<Ӗ x& Џt4-b1ۖy̽f~2M'i y g]4|=(ewGQ&[SIdY 2K^`<2O ?{X祼CBXe>iڊ5SLy <Rf#! `R9QKbj7 OP֨*d^_G0WQoGRfu]T'tWa5yuUMUk]r^Rڷ,'i y/OADߌ:Bk;Kx<ڞύ-y1Qw`<_}< 2*D"F,`"I/'S莞͹>wŜw䨍SXqѴϰ8ugۣ#|'1,LˌúlP43}{;o$qRJ!A<9GN  qFL=fӵۓw+UiEa[tQd)/jPònR u^^,vn ör2k ,>&ap1tg' o 41WZe_474P֠Ѐft6w鬮>v'oeϼчלcCB^P9 yORЩg{g:"|~%x}!j|y` 4,b8"'JcoƽLz>4-q<* 9M u;,2a'>r3j4YcCF5!2!z٭;usv1gkwW$g/q\NPIP( /*A y;xSBP$D7<s|Mj9@ow_YI(BdsKip~+4L #d`dGc%cg7#&O+XG;zzi\L$j$ZUF} 9:ASV ]cj_ < ),X#!#$!?~|mH 1=-_YKyR{8rm-0ܔRKUqwjf1Az{LlLM\k@L՚7ıvyl,%VZY^ʕT4`79☞V"'}haUj)V׼PӘ\}IIÒ4T&Z}ZT/Ч6Py[`,o*y%e]:4 8Tse% q6Xb XLi+l*mʿZgⶆ1qkC:ay4"ClTQZNȂe$fj3՗&S1ZTbK6Zz;e7FE'0D iLgogyNg$CԚۄ#z֕TbzVt};pPyI^`z&LhTi!eKw6|aʀ!{Tݯ SfaZWDWh\e B^0Z\IeJ Wm"4ՖE0!jQTu_% b]_RF]h'T؊r[xpAo-oyuwCCl3:YqEخFP,]k+w_lj,R"Ep$u[w<#YZS&D+f\uZ+B`},"2[eȐ6ΖXv Јz0 Lj|5m6ZJƁFK38mk%BՖ1D^.'֌~Dpwd*ϯ-%T3333j\h|}E'r|~,=}>^MO2,̉2t7aj_QiMFPQEtBZ얣<ժ+\*iU6ka,眫o@ף4|RѴ?$ mnuc[w56M}ˑ(E·fZvR,7"0u=+ L3t].<݉|ˤGvW!6!Gk,Jޟf%9p-zwmŹ%$\(P mݥscjU/P&WCZ%U%]UjC!X™MҵzC*?CIDgtK3ن-rɗӢf+bGO%gy˓y[V3w}SExC4n^] .ry'r&˳pTk٫ʾOa, Rٷw6!֫ țzWƕ <- rWuLc|(Jp$;eg][opr`YG{3zk8{ ^U9v6t>G·,0ΞGXVn뻄 {tI\])sp0g;=$֤Y3 ]Yx'GBJuS5o#j|WTUCW~A6H5=IW e#K,Wʴ[ψv{`6$ J+9cqmJ.*cù΋<ɧoʦGHs`sWa-G$8g(ZNWZK?r*^9&krV]sm9۸+{Kmu34>P|9W-Iv0!c;4) x+ݝk' F8 {@0pۥDa4.!d8@blG,Bf#3 s0*F C0zI Gvc@!C gAxvwe gB=DO;7n<* Pyjx;7kP.zS87ڿѬHkhF5wq>bH(3 K42X D00;2.M O7K;čG8ވ@\JB'@fr̚>8ILs"Jro1ťF\`SSzó XIY/N9vnpi]`S!wv$oq !3b> y~ ) 5 B%N(ye".iP\g]G+v[6L v9!~S'Mm[K gngwo($׬IiI3MCFuIɎ7=1;mH"A;x|4]nD=M-`xO_A2Hͱ DT£e; gij0WʯceLغQ-UZojlKA,H%gg|/|4~$\Eh>`fQIm2U?=~w75mG06uBu=uzO$K%zGw0Q5%`Z}_՛ 'fd, /S?Խ{iZp6HWZe'w0mdvяRM"V؁uC{Dn u +,E^}}t`=ޝGcfhuX }{O$;a=1%{{/%7B9;3 m3L g& V0Dyy~Hoto,2v M^Ӑm9<4bCQ-ERrHQrB`~g;]*g\BHS'CO>G0\^VUYY$kT3R:'v%Tʐ.$UT cިz} <7t_e_baɈ)u^.2$Og/8:*Xay I=}po14&JZ@8:?!7TAN^Oz`L XMSb-)9=:JF\ ĥjRḾUճ| uɁ>AoBȞr w\;'$SOlg=$]2[x:(r;9KerWe5gl{6l5Xlš\>o`޵<\X=$ zY-oe!\XsʠY6gqgw^xRU[w0lѲa! 8̥9:{[>8>ő7RaYϮjF^Oh:% l㲺p6H,}-/n4)#ydX_=t7iNǰJgo[dgWVy~!, ~x4l =H]֢@94'LR>|x~yJ."@2c36veM0)C q("]fM$sۉCAMѠ49 d7OUT.ns":*9C44 h,T+- |?#"\U pd BŭJDNfi:! u%n-բP+?P [T솛G#q-؃B+*.z56%7JOTM4ZP2Y) "F1e =P"γ9,^I9 Id73,C+}jacq@Wega:v)զ£G N8O~J{j!õY>M>A8H<0OϮ 0T oTLlسJN"UySjW;pI 7|$i [;m'b[[ yw ~:p4ls|5`L~oQLi;\u[R27TiOXt.{?I0=xP4`tp_q<ߚUAO>RPkMJwiSFOlUf;zԲiڧwuWuw]ViAPoUPoA&HAee ʞ&({lA|A )(APUPA&HA V v4]i4ApUpA &hHAVv4]iF4"R.m!9X.DjiRN0|z[7ϡyNf% 0:'gy{A@lV]ԁcyxvrݸp~qŤUi︽MY1^\qJzEӪL&VXhC_ڇ?q:*JX&hI&h%=N6j,!H=SS: R7Nst )Qa|^t-+)) PhA"h44_,|+RMGBZK @iOyzdpᬭ$Z^uaQX$ӫ @ %eE~,yJ:R\I Nt4-gPyrd>6VEapXf,\Wu dXViHz5-3bx2֋rg |WF q紀{caY\칗О8Y_#,/8Ae ^! ۵ ۿu:ߒ\y"i0Cxqz_WIi19QRC 3`XhzSY V L'x,k]||az<<I)=&g[A*Jg*`vVr8G^ W:` qLHy Mt2;+b\fcN/@ .A7]K3`=k0T?ubB}k,&g\(OR#0L2I!SqbPG3uV&{!gi2>mJD (BmU_ .;!|O UUdYwXt`A>KE$I0+3ACaN&zE|-LK֨|n%B/Pb ?j{ wIh`>P`1U#wT!(GӜF 'vxd%^S`g ƒ1Q h`/7Zߓ#U<-+eA*"= jh f! Dw:DS5+GkC14.Dk-5O~KMўG{f:)kփYdAq[(oKe"Wh@oO#BT|8HrGX!{U \wad(ΠHF0=u>#;1|,?†|]f O=41;g";jy:I٤=Mݷ]*U6뵉^*äORQ\J_nT)殕*<{MZԴoZ}/y-mo&acSAa"丶x κPg%C4 Dr +<"MeIз p~vZ݈LuRoDjn-0Dm#!Ǯii(!.-V&n&cHambuVF@^EօMImRegXI oM'S3H~U<3!f B9oq5iG:MT$175E6j}n9H(p,uK;ov;15"n5V|()z^mOk@{+T¦Y4+QJ:CBGnťAbb,b, $L3| G'•<yfWH=~><^5ƙͶ {A=]tk>)>Ǜ2e|=;&} ~T}%r#,NQ tw}s}g 4c+DIç0 %_ȣ` B𑧱lV-54K, Y>_/#,!_V S[9<cs)O^^ڻe =YW&@G'[1zeL#V84Nd+&s;uűviq<eCokvSd"4~L*ZGXYSO`Z_%LmEZqֺyrBaT;9STpb&K&M^`j*%?ɝ\:d8C(ϧ?R{9SkˆLo Ay孠FiA(|O<3ƪp<`4I]ncG?+r_  0 J(&YRY DrFU# ҀT afqbG깯,Ǘ|*?wo~vI5|4|+m/cRR>@9~6Uo;_ez6+N!d_&O>YWd\KOpe3|bwv8'1Zqeɑ~Γ2GdfX֝W]3}fL/]=Dqf_4z)D)n;ErO+ľgUzja * Cճt/?8+(>(#Y4ɱZSQ!@}39 #/ٮiaWC6Îǯ||h֍=Ӛ%:%B #=rGwn,t9dG849B[!F;>xp{Oh- q2%ѳ$-r]c$륕csU֬6evI5υ@[{ F;}s3)[ʫdba}IЁ1o:Y -՞Ra(yc^+M#^S )PA#PcDtp`10ZHӁk=bhR1Rg yQueڷ{k0r}N"(yEL+ΛhaUeyv["InL E`7f;8F0)#vӭ8#5ZcbJ+jsXqa$p wyr(@\^'KEX3 1q[ ^5x蘗0ǭQsBC;ι $":Ĭ1 w1Xg=705*o {8oH0ww~tgRG?6)Y hAnt "$]~n8NܧkdqIkT< p F;P&v:j%[B%ct!FUD܁b0pYiQ\pdɼj䖳MX/=kOX."B8X8Mڛ(Xn]|XFer#IluX}=IWi?xlpuUȻE"o?$8ʱQq)" ~z85UĂa$]z};4apBF#,0){tpHiۓUJx\ e OMEH*#dpB+BH[ੱ9!$&g8D(kF(0 `$֌ *Wx[/, UpBQ(#m'.Ib˺pˉA"AsLCZ*csmNh+AFuIl}hv{" :1 i x ̋дs`$;j,\ĤJ1AmFY,J Uh8/#X'  E#c8>p){ϙ(28#{5(xd X,&Yxhcy*|8vx2':OH[xcu (!剥$ B1YHb́|qݞ*$Hq]mF*g!eIj0&BRzHwuOءptÂo‹13Q3KY I u!0 s{weP}teݑ{y]Jl.?\g!8T  `HQL8V'1r!}S%n$CW6&\ 1b.>m~sihZ DJQp52CcY.VTlcVm%cI|G?*@4(0 VACk$*#$@1nU>7bL2+( c27o$8jsU$ӏhj;g~ ar9VH`<8סy?(qo>8x2wnRy ]8:$8{3Y;hj+(10Q}2q#w?UAOUvm <.JK] o ܲ竓zdéiv=NuTԹ%-'ONB f8: ]|1OZb! $QܟYVn.T¡i)1t4&y,vH $KKxMHc%r`!@,j`$Y.}%@g69zp׼w:\8DGcf:yd" yUI.3D|2\Uo:}{ Z%=0l\$c3 Q kBbhV9&B\ jl"b: ,x#:8D/t@7:d[>F-WGKQCK^^Վsf.5vbh_ԡFt; ' +V#)2܍5fG'f[^ N]);(^F OcVc.4xl7M.Cz$8Z'[(9"dvxI=8Z×^↌FA*sj) bEGOΓLUQ-@MA8U^# s;%XI0SQp7[(qcml ^$440"sz,R^tG k(y6FNn鬹|{ҿX;~!1&[,ˑ.\59Ęy/Uњ=9mrϽ\pmW, +qCQ$~~ Nh1EV!D?x ID8YB> N)iNanS\O[lнVߴŲIړ F3 Ԧ4L ƒ e*J*[Lq<ON,3*g:LMexH@l,Iw3"Ěi h}=j:,A%FIP,V|tsL%Gy]\r%xgGW4}<%IpFT)yz&WbBZшRQܣ1^|[&i\lYoLWE9 N0 {Cݠ}u:%84fpS v:#G^Hp4E Q**=8yn8nXS)ş uްwo8̾:07z$\Xk"[6*wE~.S0}7;@mpFy[.$8 b҈q9̉`HD6ذy"!\??SfFko`Uh%0)&LYQ|W >l),/Vd& NF̥paXs\,c[/=m9WF[EfpVg.OCsQ/~կ$38&%->d1uZ ~8 h on6<gpavLp$-K7 c崒_[/KLǩW00җ83t3f%3 ɑPp\:1eQtOdW_B^IϥJ⑓{facy> WM3ܹb: ĕ ={w@ZFd,w=LBxR4o~ODݜAh?_0ec4n}pUC"# c>8*|6R$wvOdhEzhb! h8ak"U m]"L"͖h:^,Z:7`gge6 t(Up (]*`QkvggSϿשⵡpהka4:;l\*qLK),av&fg?!V'7<Gt񲿮X^9eDwtU>PUIOn@Y4>IǓoO˿8d]N&㣣 xX0FDƇ?.|r lp:r?A<&:CE5 #yqo] ChσK|v4fJRڿBL=n?<=fNh`=BV";LS=gqV0J(LvUKT;F}3#BTv.[Y^#O3:OY>'hYUa>cϓv+1RTJҴPVrY1ko|J'`}҅ %)w'?Es۲*/m]Enϴܮ͎D*#.NvJ¼}3z&fA_]I*g ,Mfa1ڛgWb܊+rʱ(,\ A̙Z~z"6>wox \qr -kFA=YFй[q:OozyGEe~1 `ht~kԺ}bF_Dp Cpquef>xfB(.y2͒t&o՝ɛR-MK腇 =C`ܙA <އRcD)<`k (Z玪.bm*ECT(Aݡ VuL86k[ m0^aطeoB@iǬfO@u4_z<>6o"0 $xK  (=̭5ETw,Փ4; .iC@ALjf雼'A0Ɲb yrO N!Yt˓ T.iմC*1Iyå&o13DeV1$;VTnn| Zh~0㐅?dm9QH,wI7,޿PJƅAQ'pKqrʢBW ^4N߳ il R#i*%xlSp0 I3fpKꔻ 9} ϴu /oJG#4ytBji+%j{ޛŋޝ>ֵ~$Qtx}{y"Q/6xAE<=X49KƩI`sw9*蠟Un>8[,P)=Q{NyewZCYt0H6O$uw괏a#S1^4bN@ CgT-n{f ўoQ`fu:()A 88$pA}fH ]u?lw*)6SKipo Z?ʔ|F.ߢϺhY9Y2G'4鸾|b֪:RB(^/fHujPE;%G .&)@i,h`tru+磗@~.7)ȪY@ߟ/΀6\I+͠f a7}0#h]|[lKš/o瓡B;֫{,1^ ZJvŬGgU7;3)C!Uk&)7Zh^m7"lmZNVԃr-à _>B"ٖ|9d4vg+^9r;#}^9~_eǓk^h2GbLt<?r$޿M~{̩l̄ ;^t s1lWR=X Nm_Uyc%mBtg4k¾\d-մT&Ky>vhKעpu_*2zzJ2ZjvX7,O0~ )ZSN[~C RZ= % /9Xw%f86tے6p¬¥܅h\\).б (RA:9fA9c,T)"6^#IJSVT[Vq)G+?18/ٴ t4C,]7'@!:),/Z`7f`Rq|f, &d :Xk{m9iU?v6֌5Ҍ?8:-=ٺC[낕yֳmNGS*17`!+,zӎz {`p Xu UpZW FUV0 aA5At"FILXǛvHǛvX &z*oq*T:LfMӌinD)ryJKL9) 1fp8 x54kM x`D֧8L@`ygR~I}|QOZ[KW๦V̌-m}}\(dfH]?Q~tF# ~xIUV>7ޝOd7chԳb8E;U>9}5_>\N|2~dl ~B.Ofl+5`]X¯HÏל,`*Qz4O*9lxc^*ŪSaݤ2ϗa8$z,fS ^&r1 s/*R߽6jvMN8^g_ogFd>"<31E9^Yx P7fbR ޟ  պ&P5M5J ͶUe4-lfӴz#4dv2MlN5ClȚ)J\Цi:8t v`IrjdwaTݽ3i:H97}3@Yq?Ge;.M'el7|^ l\:'QU;^xfq ,]Ax"5-cUt>a~N'/'- eJ=]^GL^'0FhK5F)-W[i b9/M!R35HʍCiW}3LE~rP=3968hqh%HKFIy؄~$//@?rd3i"@Gë~0:u @+TܬTͩZլeǭ2RhEᕻBhnf"|S0w'sǭ@Ī`SMpZG asm qdLEa"J`ɛDRk.>ͫƠJ[8JE9 03cWvqʞ&ve:<"VJ"13,j5A -#EDY"LHz;'bO?iO2:|+!K˹a>EKFbjP:,}PCD BYT<14aEzvKQ~K8RC]<‡D( D >}!?7@X$ `ʧ0a JKL )L98N5Pʞ]-)<&-PQSu$j̠qPf a.zp/LC@kW#%~w 譄C%̰BᠩXÉ5c+R 06P$!m'Qn.:Qlw#NW wF;* rMnm2Qua0CV輕@3Ni(_pf,E,-}#քۄP&5Gty;DJ( DNiCڠ>b7ć3_us*[\QCS^4|+&sWO\F[ݑ2:}ZGtDx1F*(#}JGjfD1;p~ON (kM:ELp(@ *C\81`faBYf}: Җ ! 1C!6Xl{`, L`Z8%%Tj,Q؄)F`uK%><`+$Q:)i1G1W<2mh0Tf@rZow%YwQ#!w%vD?= m!$0NidR̅ ^ khy%5|w@Du:OK.h%W TĘ1z-#T9 6"> 8Mp( gAED!RgAٻ6dWWAŌgP`>ÊM~ _B9gP$vBCapi* @l9M6g(eHHiR؎$>0okq>"9~q !bTVWL 6=⨛Ђƴ~ZE`rӡMSdp.ar+]fEa\+ #a};J=9?O&=S-2 h4kS|bE2 O!8.ԾO$BrI0jc?4'P#Iդ@ǕZfӤjo.|ϦgtJj]k5]')65u(]F߾xb0>WJ j'TH_o_?k)`r6Jw>oq>v}ʭ3GH0k5ϮQqU 8cLO_!i%Eo m6jQY],wϜ q(/k|MO֖LT\{.h_[ȫGٲưZlLL $L?nO*vMgYARTYQ\oI+z//]FӠ2>??͔w>qNl6~lQUǷ/s1>_|E>h>?/?ȯo8~!Cm#עA{T˞R3Ҥ̑ _/lCYrU(pT"r+jZR=$R{yv_:bsv~i?TC1RH{3/߼f2NF :"_BD֤Y{ꗬ^ۨZrñXYP"?޾aP*u;tVQx֧ȧs18r;>}><]J../>_ޠ澵>Qg=~ WOZ*/{)|gϦ3~,#]d 43$MU) J)nϜ\܎f+ |pޡ 78j۽88Lc\ϻn" ; !.]|zw].:ժpW} ise$ALsB`"^ZO!cFܙ]WU#w~Xlu @ %Y`Y : ,U ,Pn,D;S6-ybId!+#,nqpD(^R4ermRVKDrAs&bWTr-ג5Uo_m<:.mB!9Gw"\)IX my@(]}eZ"\^ʠq.F|lgGir|q4wj 6VT3t{H\-ˈp 8,i!QֶVsrMvY*e0E5;(~86k窋T28Pm$^o7S,M}m|ہ{C&`WHb)w֥uiz׮ - $m:L}B7/BDYqWT_k ItdjmzRZz;b,XQk]*:u8 7Pk7#7)ANVdr2$V3E J9OB;+Sg_;{۞='\Q}UӲYbaņ>?F ))֘N֤k=_;Qma97^KcDBj%qO LɘE.oj{YFwy8)֫;'(!u5Qw^@/qJsBZӴ t+v}v!>՛X>-A2 |Lj:k֫Tj-j>uNv31 ˊ<\l&ljsdj3mDb!ZD'#6x:Z'| ӎYmڲϘmmƬ wDTj vEyΕ-{=}c pkMڪn&A-a@Fj%QX>@ TjȤ =|=@n6W)@ OuMA L ;Pj%T;*DWi3v_2Ӛ,@R?Ukg{yA=V:P!HQVD]T ]%'6*t9`$Ny%U,Eo~ m@(zN[hmzE>BVMfEI.U1y,#=#T" I Kz 74fHh{b ΄׭-i .YKpM|"t).g]X\'u=탛"Et8^-W;⍫* ~1یZ$hw^Pd[\kO#SDx9DxU}Ȇ~i?|O#ihj>|wpFhhO#U/wn4:7nwMú>o~bB-zpe<-sg(J&J!*sTJlr.lCnBMպGґC+c*sm IM fE*,Ԑkǜ~4"%՘N֤$j%}Ï 60k/  حˊtht>رȇ\v $ՖGZQ@`=)2BA/nRQKYz ZHI05Z|i`-#46n0Zøl BU ^g[+(:+[T N#z6ˠA`jfx;9C±w{&IY=rA;SKȠ}7dLJa)DB$rt,i7/$uු,M.%q4H( Ttt#+R.cE XX8krۦYT<~eTKTZHH032A~ H qFHh5哶"je҉N|͔0Y\= AL syMwsy<_2PV]ϹO/i V^a8rP?&%@LHa Lv 6E;:9@Ak;@PZNOvhoAza |fgM{ax0Ψ 8ٷVRWE6'OI~=BL]3 lB|CʧmsO9!P "hЁr>*M WfLFKH)sVi`gu4d=HЪ}o!\B}.aԽ ,do͊ID1Lg䰟5`Xs O[?o\e92R Vꭻ6,^%Rw?Bfq-kV}̽pȉ\96 ؝hsp]9{ݏ/ѨUL~JRXU2ׅyYY֟Vt=bZI#Vλ"5KIw72љ-W93"H8; ۾GN 0!۰ VF N>:]?>\u]?8Ew%6}Ov!9N|޶/7=羫MgS~ȍ qQ3^ߦ?;gyal3 jds;g>|ǭyxGiF%kYӂ ա6ͦ'zK ^|z7Sh.W:F*4]Xb iy<Uʯ#bxߴ5RD񇧣H ΆiG(?ޓݺZ`:f3sFw9~ռ܆tA#'V~wO'MܟWy;u;<']ρ@9*^} B*7=nHýq_ ??]tV y4RY0CF$-LI@CI܌ rw( L9N)Gޝn2ߐ1zLZHwAM'#F!Z+~bl%xB>tה9MօIҧk?Xvt7wkwKQbB3FgS]-Be5AR&-#}jSbRuea3ʞ[!t^^u=mܾ,U۶.r]] 7'v}AmʁqćN{7uR/K议?V~~g[|OR{n|[jk_]߬ V?}7ݗ\6_Y~B"d#t}h;o e ah>cM3ڀ|J]7_|dzWfWzx\+l{gL*H)iԔ_fSmХfN4{<zE{P=8Z9iΫJT1usD^ m1>5 䣏2Ԉ@9U5qzLd8*FʡJ|2 %|Œ<6iS>ӧgx|KZqh6mц;㹏Ϸy<IYVd !:xɶ;jm;@iCMO'D.xQLM)C>zq\2i3q/&Ax`k}К;Z3|sEؘ> oBs6m!"5>MkߧFzXOK"F^z6*Q遭J$+ig^&*_}}0Mzv'#.m()[fm100#FJ7Ɲw7eqh7]!'2;#kУarLrޭ\И p4_^OȂtBV8Ŕr@Z2d& `kL!D S;bH1$u=y պ|v¢:V-APɓr/ݩiӊzjjZqɴ_6xҊΤBTL)#)U;q وe#iqlCcqd=dFڴQ齮?D8Uuag0"Ar;5V$WF}&"rm^9t'6?]wư|UUg7JuJ磃8o=%]fCP5 ]AkWnBp*uɚƖ?fAlH9&k#Fuk#9N";BKEI# #׵u$k#g&}er8iL洑rO8յ#5սV4#UhW`ID&5*X%نdYL eITG1bXL)%`V)tBʒl\vxYƁ߭b.;s9Ye#st.*7-G 1CuC C ܃kZMͮ1o%@)H&C9\=)uCz4ʽz뺽%ޚQ,e9l̢xQ ]ZWW'F Xtyu tMGhp8eH9|qJ{ Fge ܆&>coy1f !UM)v"ݠgim%x#B$8 SYM0Roq^ ꂫ96o^7N 5To2ÛxAj2a"]jj[H:9k0롘 re)\ސ/{׾f3_esl#֎rZ%TIg)Ppe%gK9@_/0A#1B> FF~lL٫8bԝ8Euj/#v~{s{W7q4"CVH6mpYe $4Ѕ $\z~[^)ē)`L 8ŷfiΤƖe؁Vv2XKY4R;z8b^Xݱ\^ֱ $N='?3>H1p@㭯p mjV:8uK y4L'xkB?.RI`;wo#Hgs2_Ǵk= U&{nsV4 j`]xtXb87H12 jjD\Mt,Ȍb q@<*bǤr!֌lBQך֌d=ǿ82#-漤Ƥb@@`V`*^ L`YsAnZK+`1m*-4c+f`ힿj?X;t9CGNSMSCF5w^]:>!@hlC%˩ԑNKƄQكYʴb0h<ܦf/ܗۜC #b'(NL'#sk;Z:ok]k[qpZ  hd)AAR Bcn%Јy`@cH9PkLhk B@H^݅1TYК4R ֖ڶǚ=DvR^imVڰ #•?9 iU@sxרm'Bjvf)J TBm%|Nk_E3FNMOm&-E_7N0hY;tJ^ :ap&`- 8?F5"v"@dѱ 5ݨ֕;@dH9a-꬐GPw5Z$z]RvEVz%X)R0'KaxѺf!68nܼ B^=&B(U8JeP$Ŝ|HR&^Tl dU/d.B?߅.¯WSbv8*>Ϡv*Xפ7Pe:fH[.q|C6SǟaPֵ P hjXt0P̎ 댋(nh #BG Vbfe?-#}R}&%<[x'./k[1YcbeZh#TJ|B#L/7<\Y')CHaJ`Ƅʻ9P&^r[:twIU=ĨM8+j|6O4ѫl\"֨=$jX-ۣ{ɒOM0nS%_Zn뷘^Y قo?biuUh7D](x8M>P6.GgI@s rIǠ9i)VCbFJ 'sVT.Ru |^>Vmܟ#Vc# gpW#\)[DSzic#ƽ#c Sz*ZpɅ F@Pchte++"*6q l=Е`O"̏(D[NS:lkl8M1&xӞdj 5;:eDQ6L|z~ aOO?7^<&|9SiF te 9SR\R1PbLűا8[OQ7UoG"3  :-F6QOA "awNb Oт,cfF6*TlɕJ !6q%! sњz#")F!W!G>=b/3߳$zhlb*DaVb݉.7QN DRg(yk liQar"$8aPLc0ʜ {H1h9ōwhJR WမKZxZc=cq6݊|{ [jIs7[ze;nMsa8Wn, =Gj D\#(`T~F#)e j>;K)֐qgߖ| VLlG!$ÉO.>!k 6Ci3DHyECtu4TuӋ"*B!f"mH\)0ܯ6qO21odl#c+c: +3qaǽw}o\)gC1`2,:$2eӍ~3t>M~~wx2!pԸ&-("sn^3\^y2H+c-Ƃ k1^b<6!٢T}0p7 H rJ%yk:p=]%wR#h]0jC8*I,p1eb"{I>dit8!05ZiS֯sR'hx>/AC.grː\>Y]/v_M~ 8RQr;r؄3yDh8._,^[}p mRxFD~ؙ qS>ͤ;B1ŇE,FQl(X Wn Aӡ‘`.*#艔n-}j[ U5JzR[X[Lr4za.DD<0%{2a*TJ Q#b?rb[p'9Rډޛ4l,t;mz)ϼBS OM*jR5icjP7z^|,  y|4y$wmn}wv|}iki(d_^}tLh}iu[z sbιÔxLL;̸B14uF (4ũr&ŔN+2/V_1!'֪ s=B]> =WG *,,wxVqZ8H\@7U u2>F)bzBy؛FCX|1 +_&J Jfgv_< ;1gOKU Գ~$j؉4LdOao"#ejڃ:WktdwGB +=1l00,Fu+9n2X?~- a2e]x2je Cn}Q/[_ ]}%Rr'TB1,R Kɸ\PW sqKwCӈrAF rxOmP/à+0[tˆgW\oBY]C{ǔ/Ϩy:x:00:`NjCujq t\uW//Rf0cML%2>P7ӂU )fN-ŐLc3GmGٖ~r~Q//r  "",AYwlg1^x_! tj3z} ?Ku)72X "((=q8J^*.<էZY &cs{jԣJV0rY5D MOf6؃rn(t#d`ܻfj-\6߃ӼZBNuȋQVa{6,:{΅ŽdSkT}O!ī436 '9Bcl5[UuKUb*p u/I1V`z* _:}3Aj-dboguc_7 #UZ-D)(ې? ͣEɛ,NĻ啷 uYAͮ&Ek`iʺi0L^y,OK:wBfOlg:vn8PCbυb \"%E^uS7os9d3wwϊ%|׳Ig=n/,2tаLt )G7RP-)wƬo seF&w[N*YQRHv󋕀o*')X[jϕ7"e<'Mb>+J*>_D06Ԫ'5T6X-E<f놹_{a-q8a?f9iγQ3{j zZȸ@wP0)k(%\e EؑڹT`+x&bC\RJ7ƽMJH1 K[A[464n$/> PP7p .3d[4΄̫,IjIEKEʱA25A ;dzr$5)St*1K)2p¥ISG$p, c}mƜP$hs ,YFuI!49X;)Y${ nPe9B hV!JTaRP31l8-3L3MM橐)[тi՚lHme[*dH2G224%Žx2!&(q/`b4fƭ`Ze.E0"tfScN h REߴw[:%a y@< 1e7ށ!8M3(*`_%; HTj=4Deՙ4L ȳA|C)u8 Q;FFIڷ&'~XL, ,u ^KpҔi+qb+'KZ3,M{:CzЖEsi<3I j@e}޼(6RJ W[b#NZfHĤPY>/ &5KP0V a7X~Z?xyNq{L[Oյ2Jb% PWM\ TQBzf33#`M> .HuܵhVE}Ds 4Z;yfc\˺=r7} Wp: <%2`v@rH P.(7FhN8W(+ wzP󀂔Ё$ '' R.҃V;l֫Űl+$O+IDsJ{R:9 HwrmRA'yneè-PPQˢIňTP:HC[qJDL~t,[Pc@sJ6FK1rJ>T \,%->3Ɇ{.fT PQjFe󇝃%aS>xSI)%K$ZC*mP Zx7zq`)1-,a2ɄBN3 _B\E}.ĺčt|g*AkOV (TuJHMГ$@boxx pQi0Q*}Иɇ(UP7]j]X1Dd hVBM*bćPp %[b~>Nvt; Ѷu܁ V?~ ,oҲunxb `Iii `ִ4fi30p  <,P>BX!4Y=RTdo>JX#Ǐ PbIƏ<r `J@ 5\ uU@D0UD;q2L}Jv(`5If~S^TV~;;3?Z^o$f*ɬ_}Nc ]@yX#˱R #O?u-@e)i#=58c[EJ#N-Π FBp7gV@93DU=5n7Ca e0udq'?,G>~X=W?`q4ռ2dz]/.!u=&/O;\{Ӻr,byZ|o\'$gHb%^NفJ2( OQ5>WQWܾO/ڤ;4>0!_ܱ$`{m שnf"&2qoeZo_y<u:Nxhܓ W'짫yzurm'[SԮkS^dd(n``(/ Fϖ"n%5c*w>3;_d0Ey׎!yBҙȜ;䃖&䂏>꒍3.Ztbjz81kOc=--M..[>ySooMg9YOf%>a:}˳MJyofߍ祌ULUwreNWDG`i]BA˭Wf%JUrh'O`j{tu>Li#@re, eOV*G]Xu?=U XRj `O:? XWǝ++_ z`_˙5[YGԖg_؉#VjQZ=TG9$(=CI>ħt0\>]=#( ]z ]-}iP荮6 `1 p?|.tUZ÷u yo8J=p ]U_ztnk@oTNX0>p`*ZNWnt-ҕ!r=u򫦃 h3ϝjĚFW"]mVuXxVrHbIKYŻQt7 Znح%OFKtIDEhWZvwr5ӫt軷y5_ 靯/p緯:\OEә:;0V6t޹8Vur>WF\4ت@T")7~w^fwMt=ƫŴO-vL`d덖~cQ3`fx_ Ñ0EOh=nHqc =/Xucs3jk ٗȓW~{zov6ͲzbqaLsΗQ-z|j=Ezv[n:%x a\Dc,Q4 ^td%}$Cs}xc j[I{'f*ǙSSdnTp%dOZe{Dm]׀}4/z''2.ڞׂ` "j0^M{-?Ї}ny6u3_"f><{?Z_Uw\՛{|?xTd[9gQ{~y;EEN\vu^2_ny2,p?LKz*K>e]e}흁;y]&>Gn-.>^ԙԫ;Q9!]waaw+2s?,N7O8Ox/{pϜv}7ޏp?`s\y|ӊ͛n|g1riu8@|뗭Cݘ?mާ3w]ѵ7vMηoiGfrdsqϋ ِbp.ifLb:9IYu:ifh+|='4_'K}O /g靝U 4ue!3vM\*qG@ r0o#&:g[A7IQwh,XYpg9xT0+@s2'"goH-][||..ݯ**o T+jFSIi$39霜yhPFs>9KL[-9 xs+!B#OR됑xkb^mq>V]jSƻ ^bpxx,^Q3N +Ih$X {f-6&qICېͣNL%.HzN].AJ8kbi>r=o57%IT{~+7K``;䤃Fr qCYnGVe#y` e\Q՘[Ɓd`l-?KZ} LT+I|%l HF p |(˅V(kXȹ$B;CRawĀyeJB-6Y7+ª@j@k[If*Sq*:2SIoT J&2pNIDl.&|kI`s9 eF|; |(SPkLlqZ$8)D*O=ф8#@+:<|\Ctr@qdN3C!#T&R%Zr&f@֎J`D}~:!wiC5q""'5YF% %Iw.Frp/ 9aBoSh/WR1CkdZp(\00 $H\SX.IwDh(,;_ǘNK"07WUŨtK6a{lcB5~DYglL_- GPZHCr<",Ml39i j vĦFԁ 9l& |(sQ,6W$'1Zؐm9̴4Iy2В4,XE8I8AIa|cyf6^nN{ZIHb* <v~\ Н?f<' PF󘛽#n_#.uu9 Y=[dHx_qaJ ٪2g99E^3r8pa]],jOx/m|Z/v{pׂw+a[1d.8ܭ͛Żt]x8ppEonarVqRۛ6y e,ICgA6QCK+u?zCY/(`zqBټc gɜtP1K`!jsI93>.wqZmGFODk 0mjjWF܌Mڻو ]?M`Mv[:o'쿉V~$k(ʨzZ+w)sʦold2KˊqSlb#PܐL[CXN)O5MǗI4"v>pIL+]q)^PHx-qAk&J)&r;>{}ćPi;hRxG}Utr(D^a X;4I HǡBJiy-ޱIOZgºX.EEe閯i6(ƌt|˚)fS 0Ld!1(8I Sr3e͹!E>۔wBrY8v884H*a79Ē%*sISs}Phn{V1R&u(SPV6h3xJnn!^mߺCe#W Ulskw|wDzLfyҜ2&xq ; ̽9|0%c| d6G.p|a "\r˸gȏ/\>ʎ^Cp:M%{?FGPgv~הhZ ׸K9nƴc[ʛ{6Mc,tgheI'"'0cR  ~[UZgnsUQܖ2'i< 3 OzezpB4iof>uu+i³3̵& }j`S-M$^;Z3\,T-.+f㈺TuS#k]&Ņ#c+I&XV|{>9U&ݩSc Is!9<:-/iQJEDb8ҜMPGs;u]Y r ~׽as>xDxU0M\lg6^[ע֌T@b_t>X93_W&e]=9o'G "a*!>'2(DzIϳ=4 ў(,rB z\(ťD̎븠l9)94#pYr6=Svϕzq6v"xeMqq.:54[[(,c&^ $DFfx z툯BKL٪Yg6nLwM о@~yX aYW*Acmk~oðTSw-ke2ںQ#Ow(¶WJ*4dNlYW|2e:o֤7zeGO!CՙWRsvݰUKFx|W7gƕ[P=Uw-~~*~[\TOjAOcxo׻oXwjWk -6Ym^oŗ~ .:W[-)|ѣ09wfbZ-i*2ìV6Ksרk4ہ Xnl)c;p)IrDRC{p(|Ub{_?||GE#":hD =W`[Zq@.q[Ejrw]Bvz״GLe]1ou/"9m ƅWq ۘәٿާ^qdF䌈FK,'Aۜ$.M. G;>C! Uø~fnٽ;K(O& &.пrZ:j8SsI;GRhZA=ѣcUh5kP%؄1TkS,..*TN0 ո-~]VA;L߫q\4h8VFFx[ն?1MP%"ƂPGs-WZUb0r0o<% ay SWP|^`wQޯV86/KaĮT~>ō-|,Rޯ㕕E_k$w[cw֮5kOYV5ʚvJ|ۄNwOyc~jÀl{&o~t~}p&v.Fvs~Z;?pOUŘbΥ=)~|:ʟv-A|So=W'`R@L[Uu F nq`3ZŢBz|XC{S/{/=l˛fXۈGG,5ŏ0P@t_[O6ֶ싷4EDuE=.UXrdY<20utz}D{W.^jĹe'+`W98<:?R^AizǴ 0V0I{ogGUAO+_%%TuYKy0_ZOOݳ8nW}Cv \e}Alȣf%#nQ"99*f^mYŮwe=>Ȭ@EG/#,y^^B{K FeۇvB-zkvncsus,J')5rچtUZD:חa9/ Wfx%`k">fv":b&VM}&_M0Ԫ1UCTuCq :D1k0SyQBE+對9N8Uk%qM|t^6'sncBؠ~U5IM9bBuq%Sji3( \: F$bIְq S!lp Si΋|rۿ盇:]?)p76m@㓫~n}ޟ ŘcH+YPQNk-xZDTC-GږH{J#~=2J3˼̽$w xrru! Kݖ5=7kdnq"9da#IZ}ߚuY|2tey!kHKV0I^+X!^/t2ILuڔE=\0%sdTnwj5ӫt(SX ILyk-"/KiE0]a2M֩6Cl>'#B`9g"P0DjZkԚӲ xֈhJs^f4=8Fs$btJ۬^v__w%|.#vCSNµ`$ȋ̑x`'Zpǘ,\( y«@hS)FOs$׹fImW9ҔdSɂGaXm؃ǒ5p3 abRk0dOua#k&4C@\b̀d%aFquҕEш1rXf)5G'Ch>&#` ֠׮Am |,tD.._w "8X1U<iA#9e|y{rK7^i ahĝS( \8BJCeMlj5Z$%Hn"#gG̐BeZe؎fGі $7*%Dn.E C= ๧5=o'ɯaӔ fށl .ll@c*'C~k@0ź.+^ E}xʒŸB.-@]{ vL(3̲yI^lsp M}k+ j3xx߶nx+ D fD03V0hsѤ\_EJ |!|:SN; ag)E/@juAf#gDeb!Oh Ico2]/O7jL*Yx~~"Dij=8K$0<{|e6~~C;Q=ajJeثR[Ex w M@y=8M;&ؐXAa6\9M L k6cjzs%;@Ga60&RnT9ZAM k5-C9ZmcN`(NAXBZlL?2f(tʮ}Dbu; ZYBNA>>H "C/2rbq߆_۽7wV7;+?8CL dy&` R\OV^P  @l>ۧ=8jw.`;>2V*"q,2{(?;ʾ/Xe&KJҧ&hn_~pq!JqPV:gY)*ĨGXKAOf[x!9=8NR&<ޥ'6*bt s&M/rV^o>g $p3ewG_pNG͇n BGs9@8F0$&(D@ug '\H̄zچj/z-A^^^&'|)6JӰҫ'`1% ǖ"-,7% 1`8v5>'Gx춂{sniL E Qk} rZփ-Qkݦ3^t$-iU~@BӠ)'(5 Yq5AЀsp-cA$d ސWq_~H8߄JŒ~ѹ+!E Ù nlQ%\xB,qgyDXC 5 B )S}XR )en%Au Ց"Pf[l?r%z2"MksmmITN^N,N GH! )iȶJ ͍? ;D5~\;>'QYbP1IE05m|$)&vʡ} -jMӁhCp6jLĊa &: 1p_ 9BsJ 4^ mQXt}_K{1X˴yG@*σ^s fygVsʆo|8jSl ʮxm*T #wwIbL< )J-]`ݟy4Gfv?TMδQXYl2N4KQbHc^NgB<_/$5@LNu|V7LU7Cx3s- E1N`JxX͈e&g|nQ Zj w_">#Q1ڴۯmqG譅ֽ\|tddy >j.'ہ`b68&).\iV,;`7 :ؓDrLzf{](N i81RbXb.pARxBSܣj/]RX3s0OʧPV1+@\  kJ 2>%5%`WmcOɘ%j2b3p2S3+dA%'8cF}3XNVѹؽ9䤖ēAr2^JלrRN~by \ppH2!`]k YAYW =v68vXA@αP_ _~Fig7k+kV6&8S˟b*MȃHKEo\ap"Uα dVnLqppQLE`mӜ5pob 2Fˊ8 ` KK eQ/*95j?WR:~SdAIF]͚iԵGЫ8bp;,X @3)iH81bVb#BDL:|::1DŽb&ʝz;dca6ehxr86.fpԯq~7e&4e蝛^>q"\珊_7u)6W&ߢ%U;tocc!)llyYw-sAט'BlC4M w_G܏#k>G.SqǸuO6GŨ)'Brnvw.Wf[[[.<,,Ղb*f QB/~h"9O Wayc]^ŗ%Y蔳'-.6^SmkJa~~iQ\}(KppV(Rr\/wnZM{ݍ\F̈t-gz $L2eJU<^+yo *k )C{$`v_I~ʱG~KUgp4=`5ʼn 9/h |B_lcӘ]6nTЧ_|C="6f9x}<tĆmQk[Wnعmڒ1nh >yttWW6f9yZFU㎗tڐmQj$lxm!:ADRs@9Mզ2PAXfxW8Rt{k$ʈHKB'upq"9da#IY֫NrP|G= F{ |,uĴge"fou)2kTGˠ_XW?hIx>5l19ӑ )<'Nq!Gd{pxeV\,Hwftܖkjux=>VҟAW;] JEp耦%A'ibJdKҮZ$$%K 4Ko\\Mg-e5G`ߣ,~#KyVP\0&.jtBi [ڎ߸pʂR^,p"b:mfF2AC=cfz,zBYF{9+X`QBj*{`k߹9ZU`gvm?JǺn鱰{w-w6,q[E|?mh8k6it. ]_PuTa1HWdԜ[:To7v~ݝw^(̋+W55J1u0PM3osbsZa3?w]s;T4Kal\/wayZu;)iU1?{WHP/6bгbi ɴU% _ReIi%2e6iLF<FYJlFwa]!](v9j":]*1ztyug8~͛N"g\I&.}"M fz{{'O`-6G?lGnt-/w& ӧ7gϖ9f8J]E?lneUYm>&ycD_tfaxr?.?6^Z6w9)<* ㏨cnd/o/G"Q'xj'Ud$0ϛ&aa@/T,CMyDNȰ, ܂Z7.-Kaj074oG !n,E%5q@=4^ A}<ǛBPBŐn,+qߒĞ\%p_Z\uS.bT7Ha)j|̱~)l:7.AKay*<F Ă_ .o;HA JzQԌ'&ᘡ(F `!}h%}bex~2 'VXxޡz|f7l!QPzy?A)nt'vЃwСtEJxƫJ$YӶȹTsNp A) hh8"e<=m{hÆ_G!%$ڿ8KCpG=X|e 81ŕ8 X crJ9@P?')\.NrbZшl }q[ZNCjdjtE1}Щ5!gE:Oexޥ(%;OnCV9M"Ka* ҍ)5H ut1D>D;}ThǽjQ8q>zX( IəEfhH62 ՃkiQݛ0W؉CJ|LJL9־,lzCͨ bN5LWsraCUefԦ6;7x> \yUZ_+7/)!in8z9Vt\Z~;/eoώ)߳5xSw !siخb {OOjuh/vg?B}դ6{(A݌2+xiOO bԞKtl^$f(y(TQ3^Td7a.10]e4rHGq>%hS9|YJv'cJם=)wwƩӺŗ7lEpl:,&T"EԐ*J4JǸ-Q0oБtrP@}ɷH'5HUaߔEq4ii9^Z֙LVt+femo*ߐ}y0՛#6|*C{}q'p"$)c MrU"E nA}!?%fTˆj8}Ww8}(숬l5Jg/43.2J*o\<;xx[bd&0%i1M|lx6zJX@Ȩ`Cٞ@oz!7kX+ ѺFepk%caXiĨ .DX8jܵʿXXEr49J=4*Qvա'wh>8ǝZ0WðVddOX` Z 5XHÉzqQ +sQv0,x,q}6k560* _‰*nߒ7Է;wӱZQ(Ip޷((XШ Q.n]G@?  ]_( -L2Fg~^|Qh@\Y:(wiZ A{VAjʬD.16%\0ϐ$(MPpW$s,Q0Equ!!O\bFu)lj iਇFep::l-%Gnqjz d=4*S@;5ʑdx ׸q/iPr{gIMe.L7 áqqBa3ʹQ-[绌QFep(c97K Zb*R׫u&^kg2 цu.D:%qq#H0k??Lyԁn$WF@*/ ?qe*82#&4J-TDʰEN<52zBШ \ڈ3IT 8XdT4Qlp F,vD9t4ue4 h[c>vwyH1uq|iaqz vMfiHȓ*% jpR*R|/m^{hT,'aB5Ɔasq27\E (uFHOas|~3ygclL)רm\ĈzΖ)wV 2lu^#¾<~>nqXjaR\=[:B]׾ȌD`6Ud:y#Pۆ=7RTHQp\gh71$C28֜b{F3:UquA ċ*oEΦ(ƻ߿7s;BPoNJ4F 7MT<78`{C.8o/MB+#Q {[$+ Pp05!|*2C -ILkgq}0\\+qTsK28>OI0=4#1>^ ;*\#Feph_u.!U-e 'zTØ:x6^bC#dv4uhҜ#4! 'Z8{`u\L z ޞay-j/YSA2m}njDHa`3ZW|Z&fPШ WÔէ:b&p*VhrH!e싀^@ZCecD*cn9ӒwD]cצx_|Xt.=ϩq]vŸm}ݡp,K0T37ɷEb TatTM0U>)sv>U" E`iͮWޭ _31xw hdp46z_ mV7~r7L?\\3=97& s=Cr JrȾyk;e%l>~ ?_,O9RYk1$4HqhHF8iiA -J8I4N%OH4 #cNN!^znb1|OCP$ka,?ѨR>j/ őFs)>7v<3/y&[K}s76Sw>/BWhٸqXRuE+i 5l-V+Vil!l2uy_,[.(ܹ h1;dVd/%+_SnK߹̝m|#֞y6 q[.j23JURu}mыn18(:ԌOOK#5c<#\R&)DQOSke :5=X# uȔ'vxkg[b!0~s95X1t Dz\g.e: tXU.X!${ZL2֡$"jJX̢ҵrZiKﹴIkeW,;yB"ghݓUq h/+vJ{td0?%(SF;;{T+r%q!.UiI!+,SVDbcKϸ:woNug]Ĕ 80?pmzYq{dG=p["+, sS"k|p.My^O}3~nF0jk6nF쓿Ai'j`=ЎN,6L W&'ǢϿX8r1Ȁųt,h;|`]Z4]km3w-ϖ>+|Ԥ5ɌQ1*}^`"[ [/3&[tԧUoE_~iug\ZΘaڲ=j D#=n:=Di_|vXviu+h8%c:!,Aԧc?0T=t1=tƏ8" $+UyS:`Rf2e|of} l8fjlcE>[ɖ3*c@@Elg鸼/ESw}9#'yMX4T(ccAwvUM]-u7Uز,b"9Hs&lfpƁG:gK)l) v/sS4 W((@FF_L͙Rdp熓(w1on7Vka=(99CZ/Չ}L~Ga:~bsэ)c(ژ){q`Pp\Kሺ9(ii{e28A2\L⸋md9ΜaΜ02͔IB.`8R3FQ3֏vgᅵ;glզ"vu?5w!PU'Ա{v|SQcӝ-!7|M/w6+~͵aY'vUG|w { pR^ˑ9ؖxm;g;?E ́G#"&0kB_Tv: Nl͎)Yc._;]sk/G:9mD'9jhz-v5'i"&|f(Q"{ iGĝ- ZCTB5Q=)5I֌1sXel;s9s8LAK(Gx|S=!#P, o)c\sq̃*at)ʩzI;,?F-U -[bGt-9Z8an|/9sHl86muXfܐCc$YGu"!!*cG7 8%~DfDa_Ren}H}y0:2^"AGOv9~ vdQqE%Fq%;{$;d^|7dqM>Ypl2bu.QYK`Xu`&oY^*r,VRI\U~O=c>Zɠ(˜^ `\f,ᗗ_ g]+0l~;O3+/|#k<68ˏ8@H61+` Z`h'BЀEH8$Q(Xʈa_f6omR:^ڔpl}a4$4] `k`?vL&B&?X;L2h]K5ap$ӊ%eyJ U" RjXƑY3&S)DsFP"u֎5Ҏl HB@0 jCnA#GԆZ;*패Ddhc9}ʌ~Ln^߾,ӹn/NI@2%+g3o~ !|&I(E0R21;q,MGa8Rb53$5ձhS(t|||2BpaiSڈ KjZ[Hl*.l ]0@I C0'O ǮnJ=vxTM32KR:8{M(莲wwHwϩk."ےPZ@eWydK~I:ɯ Yۜg;6/LA_3zM"sCͫU=/pZ'M,Kf][\!-((ˏ^|4Ѫ\Kf;@lޫӡy,G^`I(s R\M*y7gϾNfErnbkLZө;?"UC1u6Wߐ YXoUٻ6$ܦ~?w{``wc!q.3M*$GW=|h$QزDy6X"G5uWW}Տj" <G˹ 9^=8Y Lc!eGF60|[ABdi)Pm^DB#+P<3xPœʬڊ_"k)JM5(KUiI䈕x5]:[P"aw_m6嫳Bj}GW®,qv(s]Jsiخ+oGrՋgkbn:M`Gd/KǏo*),L]uϞFdzLIY <=m0I(#wdoUV썻m)]e-շ讔kSLߦӓw/?ZWd@$4pڔ8qvt,4·e1j$*W4"@^uwZ|.,|h&0m`BHs5Ә|)X&.Wb:D<VrEoR%t!y$SD@kcyi$u[ةI5!oelx$^%e<"]hS'dKR-r ETВ6&˔*@clRqeQCLҤɩK-yҦ 76Nqӿ~peM%'2ՃM{K|rpvi][ o >j>9f,xϥ3Iq8~nqu jˇ~nm—ЄӦg7<_@;U4 >z{~iot9n\?ƭhq<84yeG'd:,l^ [Zl+y=o,+66> Ŷ>vj91?olq8'#?z:cK<d-:G^6DL,IE);c:OZ#Q!RiylcYx6fI:,~ Soh2Àxr:oqp!}a.SL}:k'n!Ǫ|e_C#7~ ή_͛Ӥ9tsgxZQCuW͗#HrkmˆYbY~pɕe暍5wk㍐f(ͅ5Vv8/f1;]?G?_Z1yj>:*ׇQ@]|Y$D|Ď`AY G+Ks)Qp+<ܨme,e-hANjh8$Tb#y5s5=]&]\Ў-ߓF ;:7do{==fp뒳D|:4Ghix>) zwHka '=6 gP:ϩSQQM]6Yiؠ+atL`RcX3c:%"xbAsX =3v|ZI1 '˟yjѹ<ުsxzD8yPa7;糺Bz;AKvvS=,3n]Ғ}BKB*~}y]H^L*dK6ߵ$?)]tOv DO (!&(ŨP\0*65|g4YNzչnI':%ߢR/|)<{!Χۍt+N>ni 1&ׅ7y˿=>2:ﮇ[e0v4!f~cƣw_k=j ci:FjݺWζXyC{q>y\AoZ׊7ohYM}H \𞸠VU?`c`cgZ՘_6?l6`#zDJA=6`c?T~J}~W+jJřثZxojE4}׊Z2׊kE4*Z}׊kE"Z}VD_+׊kE<!9; ?x7O?|NrR4?,7ld1#)6 d FRY5vj0LF {f~>j,aH1jr9hg[WF9up䨘2רQvX肥C]oZMV7PW-Bvt:lCi"{R-_+؛ڴY e4eMu1x{Z _ |0VqFs 15 ^/]~w͠q#*־ @</??j،J,#?VTpiP|pWt[9qO[@5-9d yt_9dX+1j?#F})ۖ:|oNĨZNꂏV&=4Zw`5ƔKR-SB˭1eqΞM?Msi;tX{&_~/)i8۴_(^ 8/  '.XI =uGd f0M_ 48? &֑$UcBGnk9vd=s1a}]~1_:d,SGl0JqeQCy@Фc%urgKY*0WJ+%+!VB8W=ژfbo6fu9ۗY[aԬopc*Jy]X^c T7aIVt;.,7m-pY^* {Ra]zziy:JN+ޱN}Y8jᬮ⇳>t?|A˻wtu}o"<&U rqeg~ߏNNj Z|]*ų'ev-qre&[doQR|mEtj;N:gUo틴UP7?f&k+wqlKr!A|H9ISB  {\eJrL3Ѕ?"Lr)Ғ gxq~.bn16:FWCW~ϩ/DOx>Itu/<-~hْ<2<+/tjwud?s?~8:>޼^ḚÖej]ҙܒᛯ H+p뿠uTg7wg{oM0휷ŷa CW|a|uǕ5_{w OmT񳔾@g; >2j7fk^mfG(^MGQ] UM!/kϫe"3p@q߳o'_gc2q b.GMXJPfkkmN7S/Ub UgKUw~?]E.@.#~zywx?>lqtWp){32-QঠJΚT|q:+gIdRtL׃[ ɔ\HnUK%h ZBm37)ڬlFP85tZgpGDk3sQ5-6.Mܻml'ʒ&dK*0Bsw@>R;&Ylnc0j!SCS㎢QN!yD!8<@#.D$q[ Muw:d %SR(U- рOBr>ϛ4gUA2țk#8sԬf[ ރF%k u3:9&F7'ߓֈscO昑((E?cA&΄|`5 j !BR]oZǛyD@K$XZFR-+Ɛwr26z=򰡥27g ,(R>I 0jA 1 ўF >B.聺0Q9^&THdFG>rN jSNcG3PQCm>+h-i~G>83 AL7XuCE bȋucmP6Zw jnV6`Ņ%dCwdus0^=VFW,*6 V4ozTF)ߟC5hS5B-E9􆻲 (lj); oE>U#\ "U+wIQ22)M輕$xIiNsly#Xm#d*zSJi/j3ki@m 6YoyP BHP%D&T+c\5:co-$$X,xϚ`&䎊CC\jSS5qC':sy޿L5 <l24~ :zojMp%Jid!^x@,NRl mk87n+_PCPqj 13: *EKR|`:ABb|PQtl]fNU"1A db5WtU/<-dgF\# _n]łTM%^1P,4wU)*K;h \`z7w ^ת`X*uwm6fB&a1=Cw /M TxU2GH}1[ZmU&1:Σ&i.Y. Ao%CJP$JETdڇ`)C#`ҥi0Gt4/1 JWLd:[nmG᭐C@8]]UBNu~T}:MOUSL l;V+/$ `M0Kvߝnc Ȼ:Y.RurS} ]{ #AˈzuC.M߃b>MzC*jKB*q`(#(vo)ڽ,=kpⶦ3:Z.fX;6@D]C:x XbU30ce@0 9b1di1`EH&XkʛvLpyP/FF̆ M F9,`=KF$=ki0jy>N’_C)cf @o"p \t*M .Fgr\UmKcbH0r:@R+>Cl.Xvk5#[ U)h].?#t=A{G-F(G% ՠL|Ջ_d~x^`^tc֬h%Z1X{sG1 V+CW 7,`d:tb ]=CJ$uŀrFBWS/ <'_ ]=>T{ pPOd~pZZgҧâp BWziAtŀr pRK+F)BWχ2gAtŀ_ ]ՋQWQڋ BWa 1xRy`@c/N^9yˆ?=~_لŻڻv5,NкUb7'GAz=c;]Ώ6V_~wz~._|/VN.|:uI^cm@{ՋLGwL]q?`3f;]Sp+"xu4TUZWz)]WEWF%߶owfaYR=&'9p1-@<=imf߶sO%]^r}/_V?ر/V_ 5U;;wqXuӛ9Kw}u|t~:~I/=T2]; 3ҦBEΗ?\leC)OixB}q)EG3;kF;_Xg5]3=Znz#*\//޼J}b{Ds ?%92m$d?=ys2y 4C )RoiMLv;Cu6A a2 _SǍ]ٲ}pCj',RrUxbݩ#dab8do7B|;YT0Q?fxcy?o<6S6YXX;+$e3Fs>2}Lxu̾ge+XeL۶6\uCf9%D(s!$BWJA%&%H\J0j8+N JIEWHRz+5y_q+=P )UzҸ2ӳ:7gb6?_bay7jz gU7+Q,1-?\r>~۶W.nҌ/u1[.0=nn[h8`}įz'VU{/NvU{j~3U=li.kl>A控 7_?UlǪ;q|<q *Y)_\DpSx p%'a7LBJ\л!k>f][յpuOhmG0JQ*.@W.jsatvx+ֈu2骗}z_f?En+6~knx<.ZY\ Z`luJorUFM#S75E_lyV *3*c(P31RUJ*[Vv4|;0+Br!d>Z'RJ!]I+5ֹuJI +6άTtBǮ+lL+-皐82B\Ǩ i]WZ1|92J(F)wʒ2Bcҥ`ueҊVΊUdt]FxUu堟RFW9T;CEWH}rʳƺk)'+“ҙ c/`CJ!|!Ӗ_ v]Opx!ChU,wt哮v =7\E*XNFW h )]Uu%Qk q+5Į+T6骇Kg dtSjbR*tC])+8td{/]ORUu6;:B\ h+T2骇2IIIW *Zţ2B']PW}c K1qbh28v'xAP5)F fbR#?h"qKNhFňKz(3iFsqJkYLh'BJ0}ԕwN*J`u\ȬuDZ->Jvl ׈kCHWJP^. ieOR }ԕNHWuWk䮀e+i2pt[wu `+Un;Һ QF]]]Cυ8zr`]!TtJŮ+:骇$(zGEW@+]WH)誏];]!.'+U,v]!VIW=ԕ´2%Y P7|>3*~*?gFnaqOene^GQr_W`c#[W?_^݌t~R8]j(^ZE1nTTb:ʳs_YiKSQ lbxc8w7㗻q>ze\UUB>i^~ו.QhCPGh;T5r(~Fɝyh?X=&V=#|:p6m_R,O?Un/[>vWo+詫^E6͓uZy7ۥu߳w؆]z o X8lH*^Sm`ҎMaѦ۩c;&w@ d9V|F,D] O'?pCc-c\x;r[Q%sf՘`; ?ͶDM{\31[׵/Ï˻U-/j=0f<Éi6^N?°o ]7z'l_Ul>i7_ګ-5&kKvYe7|irE~\fͻh!{.oB-!Dz{qr'Z_e1,iX1*oWPhoC4vC׾9fq.6j]|".ڵfvxzx[Vs{ZCdf e'9O^KxD65*L K]JմTPv^A\)BSHK0}Linr2B\GfU9u$]PWVq.-!]!#vU&v]!Iꥮj&?guraVl<L< *fqnff͑)ϪB,T|.<`y q7x|dJIJt#|l~XOl\O.?tlq7ٓv}=i.jQ̧()"8<ӄf4Ni&$`W)*3uѿ&J3hf4'3A\hVEx )uqF㹻ߧ u\ iE􋇐R|ѕjz5dBo;u8 tx(wiSe(ttv =ZKFHWl4]!TtĮ+4%]PW02 -]!Tt%$SƮ+:骇Rv )*Z΢ "eCIWѕ^8NHW,#+UL i:}*RJ+n#+WҪ'HiTUueҊR PdtޓѕQ_TW@)d1E {:+L i72|EW6t]!dTtFƮ+t"骇F{={_]!dt̺+՜Ǯ+iѕnz=du;?9= t*w`gIW=ԕrJ(JA pvTt&}ԕvH2B\Ǩ h->wB&]PWf ~]!sdt ">w&]PW̑akRu'9=jQ[Teש<+!(Ⱦq6M rGKG*ZGĥ4h=Rrf4=8Q ;7 בYu"%`L{~t}׺jk z[0Z06m22IWÏI̥ +zPw (ElO㓮EWayAfrKEWHdBJ#+#.n[W`+e p- rRJjB/՛bZ֋y] mUW#9ko;Vs)+9e59|CciQVbpW>ӫ˿o՟:,{Wgض|~:Tw'|y+{! 曉+~P2+1/?{66 Ŵ~0zG2u^-O1VzP٠.[ZLӋ3zO ;х4^ƅƦbdyGn\؏(ͤұm=N-{W6/E]ՠUov l}eO|dr  }'L"w/Q/36on&j$Е1DBsR֋ǯ߬]].Wնv(Ӎw/;eUKUaUE\)l.M4iմAU{2kCjw lY:K}K-!H6^\Vp[J7|>C=D W?Wɾ}mAmz5X|,s-j<ʍ\ T,Ϋ0j^uT>NkyXZ׎}5pcܨr4չWJ[Bjlp4 _ "f< *!%tqwӍ(!pjP@1 (~Fɝyh?X=:V1殲ȗvg_R,O?Un/[>ou7Pf/+dK׋MM}uFIl30~ `9|k<),JwCoWW˜,{ۡB0+?o?Up3dܬ盁WX>]+-u7yܶV}E{#I_vqIyvϮm흇Aԩ"i$Y67x(I,Ve΁aD*"/{ P+o("x'#2ʸ]GO+S.OKJicupinr堋Fi^VbG|,?;Gv:)hUp@6ZR*qĐ"h!1 1+~U mRo}QV*ը?$lKp2aVHHk|B<H|PXx$ @#d}FI-SΉԚ:Yzḵ.͜G]ZCq]x<&;3if3љб* %A^ʆD%:{1?\iItZItSe/[SzOunmn\"Tq(/K-~͢Q-v_]:yk;n4;Fא6R KMQ+,5J dPgUMwL˰)fuM7{0^eoY h4T:Q9}9`C$r}6fbTwǪ> +SVOjs~ߎ"P<қzc/[%URKhJ)OYc*e)0^{yDb>Ȑ~;eZh,0z?xh8nnqGzЗ[Y-\-,Cz~0>m 3_w. zVk]`4R.(/x+yETS-&Ö¥J4 ]J1zcA 2l)얛1xa#V eJuG y{cPKCot‰g-QI..C۝ߧQ0,}xسnj-]9[j#F 8^&ɡ -,Q_dXq\<.n g u%7Y4>]q`Hq?.D^{ɼa1^4Z4co X4QB߇sZ$5ِD-:(;Mg|օ&O eiv-Ucp"tYF3fZ9(}:c,8Y7^J27'X*z7#r-j1"ܚ,7ƬɊ wuhwMVLnW&GPպ8rsةizT>7\ c[E([ 噵 Ѵ^,bFYhr*ʿ5Džhlk]c aG P Xw2DߴP_1J|[Q%ZXu/j k"ӕ ,u|X?BDӅ21 T]a*%n#s,30nDJqbx ?3h-o>MBkՏW <&A!P.cmǣuȾfAbdDu;! $Q(h(f1"9Q ׃Pla38w(YQˮAnEo&7`_ 慳9jWV&؁9ަ!q ռ2փI+ݔ nI2Gi\ Y:jZwU 6~6=F vDs  RԷM -R`)]$Ѭp -ddM`igCI2,>PA@>>||`|aQ~4z f{ͶbkT?<ǣ?w?ߌEdd7q~|^`4&HA NJ1"k*V],0 *A:je|6ώf}hR+¤H.8)zqV)^:R:2I擥,%Z Ak0X,5ݎ᥏I^JUfenB>ҡ xۗ^TY>AzXݥg3ּMNB ARל?1dS^! M,L_ӿ5WBlxt\q B>Eze-skiG9+jiRD8V"QPz9lߠ YYqB̯B^bg.HJʿ]-oi$kT;nd/E=1`z%yAo&/W-Aiwh{YAءvOWᵖ!15D:Z{3H F{dU .J]kL pn4T3?9"7]֋;7Z^.߻bK0$i4 k{ˎ THゲ1D>0 3i=w>mKc뉡Nɨh;ܷ3vJDo&.n_Ra\xz2.V @C*0kۻ* .CYoaƁ9 Ҍ} ,FDuYV5Au *$i=V@[FY(ErgJfסphoZSEhizK) JG^i0tsX hzZ"mtP՚<?.l>evVBHF8OX4[.!3[~KWk?4#% itpz||9 :xV{.mdKe'R2~&MlBP\@;bVwG*0p>U7D]r g3! ey0Į=ؗUOC5^v<j6u}b\q!h0Nޕ41X&0 w^cb0PW u !qz|9gϕjl:=~IǙUAd)cN/SJ\+ni>Ҩfx.I U/p?<.m-F ,MGnw gW hPrd>:Bn/l؛PWPo(܅*g3 7)g(=y$v5卅=[)YeRf^s8wz>vw_UOуK`ā9 B9a7t5/)s]*tO-ߧ"l W T7i~' tc`/ u3Q۲g~&7 Ofɤ5CL]Y*>풰>oM| yoVIZ⚏kk[Y6Ty?T`ǥ|z0@|iO%  M9S 3NT/gv|]"e©/wA >%Dݍrl񗗅 !uC֎nnTo8EBm ""5V,4<(7+lxqd]5EFC4!dK˥S"dO٤,AzRf1Vbec?qvP3]rKPw]rV$T ~FG<$jKԿH3cs!$P&I8PDn>_w\&:sEO(m<^ wI u3n&1$f3 ;RacAEqDsӈ'IAaDrԚ;NԈ;N~S!p3I[.}Z| uǧ44;>'GXȣ# a4y$ILԯh,j*9gP}$7In:'"xVJ5602|lMc+ [ە0D9j$̺ P7&`Clh63% b#G)aDRc<2)0O(?G#E M'j$aN~>wPeѫuPg uPg ,; U_LN"\de",MPN0Q2H(,玦5⎦ JŭUP? uP? ;"ȓH1 U, 12N  ,$Q,DY: J6(Qrfkh:Q#h:JQHlخ͚QPɲCIr"SΔC3^#[E;HNl.R/7a5ـraGx^ o! h逬U Ww#"!qPKeui݂>/~C=rK}&R/ٵW:X|f꧁E0ΐI1zȂ/}Տwy4^dtrҙnZ_ TtiVwjwe$GqfErkHiޤ7DEfJG?RU~ c9!QPFuGYcH&k "eNdŦL_ݭF[B] ~c:mǩo_K&R|i4|WJQ{D=tM86k^jB8J w4C=df55QR㒮1NIhoC¦e#bfh>MK6liDKbV(jkmt!Zӫ:D q -fՄ8wWC}w5wWC}w54﮶Bӛ)7@,xJyTP%[J V ^)?"F@}DB '?HqY;ֶK,qB#} s$H@iFR0 Oʹ(x¼}C1W-ÆoַE++>,T#s|Xϡ)& yy$be3E%\a^Mr~?hV8EY`lM`ޜ(ĺg^C{!w8\Y0/@!p;˝H%j}NW})0y/qY,okV0v*biwxټN̗wLE L_dK+%TC\g󧁾v+ [໩]G:)? }FtF8@ФPP "߁_Ow@$ o*2G5K@xX|u t+2綅s*۲WqpX {u<>R4 }rAa[g "[9h^ĢʽiS_%%+DȌ]}S/gu9LQe caR[2 Qa~ 1xeDo嫽$K~*BwUՒT$LX5&E a )HME&J pTVFvr BpiY 4qI;r$CE*8gbS8Q=EN0lRH"^ՑUmUʘc&(DQ91 ̆^DLq'0TH$kWZd!ju%B41s(ȍِH98{4$LrKaZC4Ԍ'`4󧛟ܠO3WiRV,wtOMT_}`'@h~R 8}qo.8^1o,p~˛K)DSr.f{Շ,o'}swJs}OĿ} +sqsbJ~S=c䍴I x=Y$M7lǭ] 2H rӹb-7lJZsQ<#6\nH*3rCʑrwI)Z~얍EJVw8Y̐lT\Hk+D',&2j\=\q8eBahՂkEW`Ok?uЕua[\q04y2䊐4Azgccxpu==8c 4Q|pAZ3t H`0FJpt17!05b@)љ\v{G*(l>_*ĬW_w4 EY ;KKb)n!ero@`xB @xl u w\0DŃ$,6H  pFEx3ɇNA8[M_ -jkX>U{el$:=7wlRx`C8o)gu%71_ʈstM@N*8Жml"sy+RtѬnR|53Y<ĝ0B:KFn Fyt%PX9Mˍ8ۃMa(id岥h<]3k pdNXMe0SCٛq'́xb~o1P Ya7Ni푝{Ye@qb53 PaRBJ)ؠxB*RfShW)vz&B+;g\ySMG$@_W׵,oa]X?g{ajL2WO%"rM6М{qP6 br{ *y3!&U; 17P;dze4AKK<Mvg1iAzo4HNqR^&*1^{dBu :qK,,#%PA RsɗXAđXkxKΤᆙ伄Xڀ> Fb;3Ȫ(B𖛐uҚɲs%ӢVSi ݝvTᢨL)nɎwewy=!7׿]sRyyT?x}h{]\"R=yVB˯'yqNW'qvt%[m\o? OsOo~yv3d Y`I7/i)#OQj} XosqC*uE~i }^¥ɿ΃WKLV9ZRs7'[@.-t$!G0e'-Lu0ՅLHR =,u#/mwox{Og܂ۥb*1n1&ڪ3lj5֯fVN߽[l{dG _2A&%~ܸ|ބ_/.f/9SOeLz8?ϋ;%-厸4\S2lxz.mR0;dXںVxӂ@g+9 tKmYxސv#Ų|h j؈>~[C'd8~yq%.}Xk"?uFU, Ps #p 1Fi1 0Ӷ}1h\LBy۽֫8hFL'YvȆHvӼ(CD$u2D]pWlJ}(>}1L]k}P8i1k}z8*VMW Č76@})8~}/{qq QF8DbݣaOXڹi]"m{Ðot13oy}ǞCSLV٫ Ge5ȏ vLй=*ӟfzVm=rii{x]c'9$1lQATԌoO>Dמ;^.vB\:=}S(J ]w7/侤gu~oR׳Fi`/aSo O7:!XS_Uj1q½pALрۨ/R(Y)]$eSiyd]_X<ӯ̳Wd)bs:T(r%iy Z<)=@[*~)Esj4`Kɯݧۏu^+Ll>͗}NZd{D,`W\&Ё$7׳[NK|1 QtWGeHbnY[)فI.f/5^N}vlH;0s3Yi3wM.iBğι ٙԭjĒ.uF [b S)#:7E7>S<rGQ53|GP@cL. +5f5 ev+F!3wXa^I~5r;sFhFQ|L>y)`!)]MM_<[14˙ r@N]Q1UZ*Gqޛ*“I5jMpQ 1"7GЏZ=pւe-|\bd7wEB%$fs. aG G<12\h񗬰ԀQ45%zyR((}|q"B5a4*ki= ٗ1\hYd4ѐ4yWٶx9Y'9lGv_yXqY`3}]<0/8s$ çd3E?eCqVv8sw#2;{%pʉyl$?$.Te EEvV oY\']Pw|`=IH>.8 s~(/Da~\~"*VgQ$Y:yIP%WMP%ە?WH YFDO68G:6!S !F46ڬL>U!8!L{erF{bBi_B 2_Z3H#Lt3ج^yxTNL7np2Ά_[SLsH?CT0Is֯< &>ve/>^k{(,P]ű_ L;J]iI ޙid8y/BkU1s$I,M-^zj݊\§xi q$o5:[B岣.#JiA]z (a%T}V֗u$:o\BӻHl. &K{K_R .Ԓ3!OMdH]+ $'Ty(;$?g0Ԣ}>C%YdO-DžB% lʫgӱN#%y^EO2KxmVHKpHҲbf /~/L]\ ksc܄r6@_Loc&ImULeaW_ i]R*$F2eq8cEsIs%eREOG-_kA=#b@j΍HqSSr0WBRYU2d'NTD侊Y>MOMpRj~"~29Kun8dzQL?'bR-TQOϯoi_s >ݢFynaDPTqZ6F7i %Ycy3~">korh_tݷZC$.?tű.ktWs l\=ia@'f[u]\=-tbM>IJ3Z`\UZ׈cjAV`q 5/isXۢ6kwbfJ!3%OG}O`,V pfԲg1FXg$CEܴ%hZoa>vnJ)w f@RpR5uX] ˚9g.K .֖˞[,wCZv1viO`&)suCPjpI)U4RW*J[6۴ zAHѝ3""e=+ 7ǟ;zLAf.;4Z\fT< Ӟ5 UJGsUF3JݷuL4ֹ5 1PKմYO'nlݔ?De[Ej\YA ^kSHQMi lZvk]"L :U:;mMX]Rj?M]se zļ|9HFwn0_ ?>4UY7V4B?42ZNׂ-a۵X(@hYtc'4%S@Ƹ@pd)+zd+`b'mp[H,ԉTf!!/6Wl#l?{zָׁ|%iݿO$ok8]mifcĭB%R`%(;?vV:aA"{|cg##ι{6^=k|DL;Y3 y rWIК*C%BZH**Kf˪FԃsuZp<~kfotO0 ґPVz4Df0ƺՠ9o,*"5E:]ӛn}:;*VvM;am )xI'>mh6UkKU씛lsq0_Zbih;iȓ`蜕/-<^vt4ݚh+j}TjS-tg_FWBM)0 ]_kKh_?ƣ?}:rʄ.;'qtSfD_󮳼<^'>w{X05YT} ~ٿ\OWc:vn|1j0Vvu\-vׇBd&ڃ,KL0VGL ~F?#S̋ wk)LDd4 Vy9XSRtxW(B̉]Q.)-'[=O!6] |pQHw+qV` I$k#x|J=]+i {:fZGKli0YEȺWsCjX ;^*Zjd0U>蚍 и0aJ_6 RL9~עL›LWsTZo|>˞KT E=m~u/9aTe+,$ cenT;e %1חne &16QH},ʷE11!݌ۿ$Rv# =D?ěr\4fU/4YaB6e+Tjrz]DP:`+cDҴI#LqSkLmUl` /"i)qm<z3Y' zHӏ?~hym_k0c$9p$&X;p0]d"Npat&v6EPV4*iY# Cl(9Qc̯[4B~}vW3"JY=_~O5Ĝ{gJԇ :SNOgΔ;BHr(\qLYMzH6SiM,Oc+ԥ+cYxLj9PT, if3$8hi?ʾRx$LC9Uq4X><vgEJ?` yˆ2ʤށ@k 6;QUDW5JwAb'4{J_a55X3"AZ w$'yñOG5Wm9+ Eߵx;V1k^ոH] U,RW7Ҽgę˗oMLiG 74("]Y°l{xȒƦ;/Oq<|˱/zuQrո1+D=0VJQGiػ!JJ/A탅bs>gAg' hucPuڅ ',Sjr8 x'>e&K`9Og%ʙm h\EVa0v#lj#LNn8C+*RIH^ڲ+}LB&*I .ϲ.bX[Lc=#{)RpP/иT`xg ctўIf}dKhcG~gүôveOӴN5S0= J"XN1O-&`>%7K+a.'Aܡy uqm]|LO |0eKB^;ֿplt0:%:F*pAD$*"ڏs1Jߧ>7Ct14+Qv4fg$䚻luzP5vV0I Zx 'J8yKZ[\L0'7HrQVLmpkH(k!kLm\Ra s$=E 3Ө F{黩5Mgnj|YP HJUjQRȓ҉s2L Bgђ-:j7@Oޔa('&`nh9G;8x}{Ȼ&ҮD"t sO Yl g[:{eZ2SDa r2,&$R J&;4=URybD~^^F)@.&/jަ ޹y۬bM;xk1fL+Zx1E]]i[orD e:3!'Z#m}H KBa,5NCƬZIZƒ/81%4n\e^"tDAmZ- KG2/+jk{%"Q49_IyJKX*Q4Z19# %: hZi 9G]H2孀Fޔb'Gi'31y\AhrJ`SR8,hc6Onތ36kH%1TJw^sr`s4,_k~Jm0$+Jemo]L ӣtӺ-DP{+p5-ҸcQw*l4}Z[[LF/zhlԜݩk'A hu_`EsQq#Z)o0ulcATƂ짩d}k554i2&.X!S^KMJTd h:5YȊwh?hDCK2hkv6*oo'nO-.U-%L%) Pd4S\13L2n |"8Fz0וǐIE[~7\OoIF)Ÿ\)VW+@HT&">ƼM4wX4Ɉ0DS78VPOd С&} 7~ reqLALuhpQ*O |FkdpðAǸA| &RW8b*1!/2u s&F,F;_3&sc$>r_gv/^# h˱vJt.~{(5^FK AIwhVj-_ΦsQqJe_e:JrSk0r<e+3貄Sx;z"9Ϗ%<)4ah<&c;&˕X1vh=?,deqli3̅X6VA~[5?#v?~/Q+ٓxZEzny6.X/YbX&SRqq޼t U  =w4\gHl8MFY25۔K{v)i.'wڴ#wN&1UZ`d\o*$%c^POe G|f[

peC J`52r|{gW<&3ˮiQjj/(QsI2‰ۅz@>P~GMu8&ݿ|E}]=Hkda磭 $-уs:" h\]a\4;W~ހ*k{n+  (=')*X*% UtBq' KOae zwGƗ c<+|j,VCm91g,QяQ:~&(νal@ 83j:G m93~;]Neq0,\/#mI/ Ao h^C13$Ο1o&|1 t?-?Etƃ}:&_;Ͽ!bWF^׷Nky];u_;[{ڱLA݁J]P-I =pdmaW$'c%4F%& f@9jesly(ZDI-jsFib| T짯NHrND#"4  G^s0zw ʞ+u%3$Z ! 7R˯}c: J7.%8 UQ|>Hjc?>( 3'*(=;X?h}N}r/}Zl{R]|bcOnف :d+6u:ܭӪ ^7+e% _\qK{Fo娙:L[):S*c[$i`7-SdTq_*8̉kO8nr-)I TjըnZn('>.{_}$u2LXPgIGXB#r [\B,J>DZJSN Ǔ=[MfK)S)<~=Ŭdu1s7 eϖ/|U ۭ*HWAE^Tne$ ׀ƑPl8SF'aWo)&CT~Ş<ϕS7 99p &|޳9R*UW$:̝*0F]sBuo/v]=m*R qW?N$$PThE)w/|?O?bs $QKqV^":Bs|B0R*|Z# aB{NiW1G!, ףּ cv^"k@z1QE 2R-Zj~;4(j DvF #Ez =Qq{)hfߞtl kbNJP 6 ؼlu* WyJIH=*1Tok<`Bz jkq0t: 7WE3I1qb]#!% g#ʉ#%' \YaѐQyB͎EG"\5Yl~'E֚,B@,G KT1SjUCIq,>#Fs4j2"=pw-8KYfo􅒔U·t.5rLsx_O5^*H2hiYE %.֦*1 ~$#UqI`EJrj6|):ԧ $5o/:Wk>W_ 嗿L2@+Ct#t])%K*XK-l'djUHʺҏNsHX%<pTu9@ɰB y;  S{^+ʄݐ(b!UlV_C>[F,eWQ*w-ܽ? L M$+juA}#yJbڪЇ:hpL k4oqT=HE)μQJHo>Z[cy{Du | xS[ǻS`-8@|0e3zy& "Q,S.k$H25MٌԡQ C^sϷ$0q%ݦ4*v&O[$sN߈:2i>wyU^Hs)i<-jFo$oҫ+RiRE!! Y>6o[Mp "ˁ8# >شn˫u%אv*XݼI\-_Tޤ:Py:Qf }|[[E6g;4UWSjrAQv\W& tBhj3Ge֣>;omu}$ L4]Ʉ|uMs\umΐ|l2Lb`#åLlpY۩>}W]h&o UW>R՜R).**w:v!o 6%)lrQMU֖{6,H}JfMfLrt kl$9 _ vfA_QhmtۯP0S3Q)um<6ڢA41~۸6@ #dI4 Ez}i{vGɼ963B &%Jع̢s /WhO+mOꯋB/.u" K%߃Ei4;fȻ2O[%8}JZV- B Sװ`RFbdX ߶P5XQf-Q"Me DZJSN ܜ؎c)e*eۯ;c::wJ1Fy5a V#qe[!r?\<{x6!Zry4Nr 'xg+-5l`,mA4j wYdLac2 3y&8Ze-El3|:Kou2d1F*0Z6['Q&/U Z0le>"m]<\QW%U _"Hz$uU@UEX!Q{ֹ{R4H@]c+yVZa@jTRF/c2.`Js9Xrv%jL#q { $ Qˈ^i`+ 8K;vJl׌a4P }}I=h=ܙ6 s0tiV<=xMɡA]_Ze3%(PPcQ,oT~@5?8 {5&eL Ι{\5E´]|CX> IeGvE֩~ʼ>{cw<^8]:;u7\6yNҡSH|XAKMJX/ 1T2DweDxDTc2ps篔޾=6r+)c#vSaFKuh'UIfEH|SU>:6¨B6q̢Z&,SR{lFE@*]/a;a0L&ӱI 7|OXSBYg:+7 zvT(y݊^u+tVIUសԍh;r(,ˢ FđkCbjj2d3hE7Bc+%6Is:_: Fpy(x̳*"=pT rמēi\YLqTضxw%iES^Jʄj MFЏFp˷2>ΰ S+s@ SKU;4PXg FXw&tO }e+X%K2~. ȽW|1'F,IÈ_lo. m3}tS4%*kFd.+opũ`IF9JjA@8JrlUwZQP~YW!zH%W<>?qחflKkmLz MB| 8>rFb8$i;6 /D]m ȶsFG/-}-L[۵dE{̽~=̶M8d8E"NJY4c-"\#2-1-SJ_ʴl^znسbU!hϔgT+A >5V+  н u/='x'vS^.qH]Jm8S`cl@Ʃ|%B_U6=c\ 󦳉!"]TLWGso4n|+Vx΍0>BS&يu)[gܙ+@ Hif `mn+nmpA.|Q^BAlAv>o[ڒ1_Oe`J%pڜHPK$OrRsW[Zib >EBKi/`X0O{dYN,dGfO׿# G\rP"Ve[{6/yesd#8zDl=4}6MkK}wy}v4E-@DsYBMc($&dqK+>l#0R*MR*\cNOƹmdU:Vh4>- R$4V@fDRgQ6+Ai=rFr)-55>%wŹSXeƴiP$AI64ACbb/737[z`xɕ0՟W{cċC6'}vx>.J 텒*#9 5hsI:]+~iR1t2ȐR3/;: N<x1d"yR8GpCڸF(ݲ-_W ;{-"iry.AMCvC tҡz(:"D ܗRn=k]?XI49KcKq(FOөy&Ϙ4lJ,(m0Ӥ ؔq&9QۃiI2hn ׬/e $¹f;X$Heuh$5Ɔ|*+oy؈CuQѬͼzoܺmp>t/հ$%{Э5|fj-W.SpG/1Kڙ^KvSudeQRƘpèK&=Α`LN3R""k:iLJz|;M6ӽ9z0SB9qL#VzH(nA$NbƁ]5ac&3zk~zuNaEIXA@rd^tJ"8vP%"[+9jJEVu!K!6-YcEU% ;uV ¶SS?|,G^XUy·Ѧ/kZUez-[}X7z)F^.muŸnfU+v0# f07q\yҀ#pnx|y !p?X hQ;{m ;)j'/ڬyiS6.ڶ+j&)zQ- qKJjEZ)(qt:@)5ymLlͰ$tt$@ΡuQ *eXNVS!wrqͧչ>d[K~{w|_$=fFK "ĩ4H#Ig1m>qV>Y..,gmYrJK\udZD+([|/߁T!ؒl8_8D,M"iH_JU}il%V{HcSdV_m3ߥs{@]4< X VO揅[y{ﶆn=E buQ^9s9(5߁[4UW| آJ3"PD`A4l+cz{.l&r _c_aЬR-z{+W6X.m}HgK|օPa5ca[v3VhnOE0 p繇Le"y#W|x{Fmvl}yZdUWTp/@S\a q 4W8fֶ;)+Tr. S`@Njď,ʣyAm9(kcs썳/AcrTQ$ 5"kG/y$@/\O8*0()Fk:A8O'9WD^zJk:nZ3Qy~[D+e-%uov@údոRɾ jO=Ƣ*}wa(ppɼ{;]VcmEUֵ,  8fXi0TG9wvXԉ$2P_4/' f9}mvu&,=ma$8ř@/L 83 MO`+!c82|v#~N ּY /x0}Z]#LVe6 ƨal\>fF86A֘?x᡼k>(?;Dxر[ 彴kC-~úoѺW-`sqtBtV8' m{Zݔ7XgvuYuIt(8zZxNbcƜi>#uw8w9FpI;88I8s~Fl8|^H9oL="Sw#i:)F|H>*Au*O>*T ~Xlyco҅X J)K)[=0֘e¾-#`)_+sӕ>M E3lRB.8s^B󲳜Wc*( xenYpVcP?TuN 1%BSod~'h)ݎ#iPѮYR۬vP4juNLWmehP+dQZ92)6xݴ$P2G乓˄bV3dAȼD6FOy5?K>vQ &\(+o`~BcR458:9k!CnvL]MQB-#p2^kc,7"nҧӀsD~M4Hd1t8ј?LqjeIPGB՜r'Z.53 ):Hӈou)T`5w՜H pYs2 k4O+ TIiLDk1ɐcpG(2F~9Mycl_%sE4Y=uxAF.3DB49:V}KQB) "rCHD8Flzʙfgh'ƽɈy" o0YAܥ$M05t*L /0:EppM YZSznBλًZԚ֛(N^DԤmŖ}u6٣h/|;Q9J4W&0W͸ $ܦr*YL T(Bub+VU[96@Sԩ٬(b+Sl(&dÿ@- ܘa`{.ǥmǀ'W*@3`FTR1 >7&Ug62cE鶳GlwF ؎x9-hpx \:|9h٪/10 9.q*B4IxFj?3S9}m xvj*քL96btaJbߒGÍXx& '4>Ӗod, S1= ū'(M"āڭӚ>sNHڐI.ܩbr;r¯9ǜI؆A(َ[jr|*ZRU2c{($BlƖ$cu\mr}sٜ\IF/!ߠ0Jr1u%B;bM '_+rP؛#lƘIy67×&_䋽$ܱob0o?čԴi=gfxDn:(#Mcpah˯n׆EzͿdHsb, FQBX&b00#ܙ=--POk)sğ?;jM9[y 0Ga, VcdκatޠDYpJhg gBOK@2-c|?BK`XD(O4߽zTjKfB)j,PОVj.= t@znw6̲N}%~0{&3"ɳ ڪlm×l٩ vYɇjVzudV@AfCW!=oJhEzVg;;ϏSvP<Ҁ.{g0sI;mF#=y~i1!x'm#IW><"/~X{ `{hyJ"-k4H%HWXm[UQ2Pj /2sSD. UQ)dٸI֥}kuquKcdo_hvMB{2L{pߢMio;uqqug_b_7KK[q3M}whZAwB*kUX &mQBsxB JQf+]ƾHfvfw`t|yz,x$gk(f( F1) 7{~D8E8<>kDYNYho=ydHצDu(˹)yKi"aO@õS()0]]O;@OCSں&|R4Wj3*)P6ꗱirn| T@i(̙+ Qkn@#ّHFmݑ &I gh@9.\vi))_˜X!DŽ2qCРJgA&r:eXᣳG+7y$z\P9c]i޴pj2Fd<-n4W\ 7_u~I2m$o>xF4z=~*e7\MӧqЧxs|Sی"Y f%,jdPP"S~g{_]vUj﬈"32B!#SJ!D6/j KV1QqG.yP&k`}*}xu>!tЁKYCEo2tcs5 ]GS]baJ#E'[`,cMnFwx2 ZOWݨjv0𯀢iY5$ep򷼋m,//Ի2$FHXBPelda|qݗvIɣ/ =o>&;t K0d/7-tw CRJR/$jFƂ,v\ab2rec,M`Ւ |'1)} I a)R¥Z뚹 in)V/,cqA UAZIԚj)hڋX͎6Ūik4~6DшV-1 ")IAiE!P%Z\o7a7}.Գٞ}~Upr=O0k3xU􃲊X!A9ZK^AʠQuI8^Ard|+-%5/Z+bRzF/7fSINnc2(d![`,,,-r~UflŒuv60x lۢVݷ<߷hw9L-#Z1X4ZZIƟƱ5E.7AWd).ZOF6-mO*+ D ɂeJB28yY__GvmbWkN_[t}Zʹ`RS7 a: h;w>;r5ݴ0eQ04u,Cv2Q-8ÒV႒!&Z'wC_q!OlE*2JM[ohpl6Q5=UɾMtZ$@Y!H{Wa*A Pq.-0L"57% մfwLD^<:2hDV/3MDTEFB+Yu1f&dQ1߼#nK*Ͼ!,hC9%:HY҈)B I+qA*#It?7/7B+eԑl8}DdIv0'Y*:9e"s:x3Oi{3Dc$ʋnz6`;P&Zxj3lD#&R(ZXXԲ,Ԟ66+ild/ҫ=G|-:ki"<ҾrF-zZ*R*( OO1d8}09fN萠'&TC;(CG"ʠ2hVkiX]126ú;GVd-Z4> n}O4R5$(jѝexAZ # NE$ʜQ-REJ5׸3 $^!W ւ^nmWiW@/2WfIᱣӿYm>ퟗRb58.߃S{jD}ϴ80^|>U3򔌍p6[̝ hТׄZvNبSK/mUr+%Ui%TZK~ܩҲXa/[EwxR%*rp)ɶ-AߩWIe yMK:{iSLBV q\6 BL1-5zS|+I _P FE]|HMAe}ڠ 4jUFq ͧ0#~@Ϊav\_t(XT?K:1*/f:HŅcG]N!_\̅jj9zt:ūKQ}_k:`p1=]ϋfF.~M~CJ MJSbp>Z)ϔyt)D(CyHv@^Ձg} O'K&h2s t)}"܁t MYjU֔x!LYϴMp84^ؘhQ*?K%TA2QIYh-=KdYfa2"n4R,]rֱHcҞ b$FHX*dM(]>9%d2a$"R(,K"~Xt ':C'nSRmr17 Q8dGZ:aA-%XḐ Φ<ӌǃb#iwW !&Yq05  6`yWyx;_5b #)kKdBT0V(9K3&b&wD%`R=WP z ̳ V%bzO;Jpg1a{pjMaBKvE--"vӚb n5HGu~1ø8Y@uGp|[!ǵ|<5L&2( i2hׇRWIw\`;++lkDPNѾZAÒp* %UU O*xFn@ O;D3V0Rok׬&PTd# X`hY{+Ur>2oICfL+ɬ/9ۼ_I"əH+PB!x@#&L,$a?i A h!y a$!@))azƕ_9p[Up`}X`;O C8$M-nufKbh07w;XUXYxO%L(i][Rv@rWj y~`QDNo \□mI {IN??_yq!w~^>r#&ɲ<̘LŞs: Kg9gͣ7v` Sz:η8uاY&%vϬts|zT`ЪmE(`br.ۍhT`9D11OYf l\6ϽY]p(XRnUOQZjk{ { V5oiQka+'Eh|*Q]mrƒp' ([PoqӁ8.n{R1 R)*gPmFY\_^XCXg\'n.{TᣣMVυqN9eZ,ۥW7m 1sm65̸z} lDoљQo3r@Y; y<vf73Չ04ڡVFD[WMȶb5/YWN!cՙ}t7I/v8}nY[{IЍ|5viѼ\ωz#n3ֽxBh{{E_B-|%I$tK;P< 9rN|A)Mik:pF[cߤn,:۔Hd03ϼ<#C|ԁNuMLu +Xa㐝*_a R2jjhWgMc6]|Loɔcڕ;iD7,s)%DiHw_|+b*cd Q&)-$ycOd7WC7 f洈rޠ(3g+ GN*i)1}uo)77j >ݹ1+l~ Zgh$S4]&Ga² ؼA?URϸ'.GL-?N34r2-f 'fN|r:zތXV'0ϼftfg,fC:Es&u( 'olpSXi9kcӦyOc*mS6攍> ")~f5Pt:xnG\:elgB@ióķx8NuM`֦ .^7w-|+: S[a<2}'pt_\N1Kn3F8xþLO>. B B%\rl1%#yVYC4 Qlj!Sh*MBO%:y$PkI@ONo1_c{ y s#W@A`Κ}h@OlhV,6_nƘѡkqh5µ#b) pBhd?SJH[{d@hܛ::0Hm^OmdRm_R=sir1A`bh}Ohra\0K["-@"ǷN9kU4Z@6g:i^ݹBk$Q0>t9tJJ[ׄ@>J~b?d%/})cN6=zECu2V:l+IflOپex"zZtY*GBqNX;G}}~w`>oSIujbѹDecN[BU1.y{R9]]i x>P#`䝸LE铷M-(싐BڊP%'RuFQ믣i{S^m1D-S}W1 gPХtӾJi})K a(وb:ޏ+hȔ{ě_{&N7?B@p6#&/zNCն*v)DLb%chm@Og.&z=aJ,%]9 sk2:>H!SRxeñ #0J~z~r_uaH#̡:˘!؂\\&cc  zbUreD+`|Fŷokm9핁1Rh~=cе%̕{?^zLS. ]h.֞H7ӂT{GGEv-_>]V"V [fzGdXV j~@*U&!g'6p;fcyE{lvzvzR-%hEfоkiԔ>ݐ,;DdۻBKK4eexؤ-w5m͉no$M+~}]v:=\D's}sWoM5<i +U#\R*UȪċ<oYi~9; A\]z^yxop2dk٨e۬'L*7§R$LjEh2&CX[k!kJ:\ٱi;O3?[ёU?ҍR}-֑cm@^|xhsW' 1 573=$hڼG= jMp>!Lw6 Zdz?̙ϼ 1HlH{6dB12L0&4lTP.9dwgШlc,G 6n0fㅩ!;/׮;=9SٞVQDVQ9ګ(Y+%*踘%0|nvAz/ 1dİD| ::wøu7z<(F9źb1:Flt}h8K[:F{pl g x8l5 9FN쁼Ț=8+e{QcRu1d4{io9 X=ކՎJ^;t8'6IsWӱ.KYg'/ܫ x{v>Jӌ{J5Bzsw+Ǜe̦Qb숾uJ)YŷZ>x1M(U|%4(4F7e6ph` Dj=V9;VMY1Lev:d5WiyHƃ5]Z LQÇ2XXrGR#痎b%CEY9|Z]/ϫ0/g qs`_;kS7xWߩg We.7yIvLJZ6+.М^>8Cݨ,?3npWKu!LƂG1c0=3q/wPmx 9)hN zAI~t?[e7uȷtmAZhm-.t_;yw+ b%aQJ XJ*AOZ&-kp~RoC(F;뱉:FWѻUlA0ʶUkM>0h*p1dX8 =bYVZcJAH0wgxl} }N2~ sjD#!2\G+D9-'7mS@+Vd`դAn}|m6* @S]3QגQWTNQе{͔M}f@N hB_8 mfe4SlCiNi6Q˴I=3XaRX2iIv( L 농1e&%Xgh{Vgh{Vnϊʎ0ltgB1u`\FƖR͈aRH N7: Ʊg裊$WCJFxAӊm2A*pcIH:/;Yj GܤK: goC] hQ+@`' 6ko,Ԓ2,Č F3KUJNVіFhz⿿5Uߟf_Oy rBԛ1Fa2K,3B&’rĂ@u6P`;?IJVMʱueTN+WFj#O\ Vƴ=EF@Ԋ"_9M-'dm%P?QM%V@&X䒲kade}l9q./m֜}lGJAfA#B m@uɪnr> kr䘹JV ALd֎5(G9ZٶEvm6sjRIxYE}]vZ_ \ԉ &g*'>LB[O1hgT6h+})W36f'&wLh &L*Rvzv\WQ5, \4qKj32VEa /{Z ˲uPEieb]`ƙJh fʼnH1n|M1p5U+$:OmZ|H4Fa%R#Xxvѭ[{PtmvHc|ze㳳kϹ#QZNj]U4`RF:}Ҟ@Y=W!sb9Z&ڠb` )\S Ȓu\jA1ɪ'3K` r FlՉ)5yS[ 'dIZY15iŖ<%R^,hGp]"qFq9®J8/ȣgښda@@_գR;١Xב((:(.&HXTf&P1 *S" 2drm=Ҥ`FMOuHJ-QhBB HX|,?[*%hUDl]pWVD-r7FcҤl/d^v8-/ ly~Z͓n+qB+͢S[\ 583S7e+tJZ ?N#Lk+ 6WVx O=Ե5jQ{QȈWlyƐ&;{Tߢf?\;`ۏ{'Ry'nouG(u]<~vb!*Q$dŘEo DM-a{&'([ޚE1*"'=)Pw=R;HyA(ejt^ EO(gAڟ+9 ɧ6.ϻZVvr>16Ł:x1k1?aӅTHYgkjڪVir$h:\vw%7>SMxݣܒKe.LkNn=9WbOiH;) .Z~l!]Cj * \=#dFUP6oU+Ygw@2Krpx1Zn]ym 1w> nF+lsBdE޳?jK\*1%hZQduTƬ noRF+q} 2L: F$g'\:{T7o"? 9?Ϛ{EQ?!H^}yEvЭwrٹ_V Zxh#?Q{ .gywG۽ͤ\o!َ˂}?Dݓw[|RjY׻ws/^kv9M6!'Nۧv[^ q=P|{vb c3AcXXO^ֿ]5 sn=1j0]+:wѼM2ZM^[yj5fU5xj`ϵeK>:< ʵJջ`肂}o{A[ ;Nܠ~qe%N8+}7:yl^ ɧitrv]ETw|p`qթʦK!{ى mPbȹbfJNFFH,)ȡFX1(ZoуQq=ړn:\1o(R i7VH5a~O;2K1෱O !OEQ<1ټZUT7#6TUŽ^vhzҾe~3A{-X;( l+!v!@<#qWCoDѐp ܋,2ܓ)p  ird gJ#SrRз[N.P{Pײ∱C4bךJ*5 p6qxέCDDQXUgu  WNUTɅPRŚj)rdž#|OXΆb5\*r.ʄ\i]M0b,T5X Y)Pk\]v)D1@mJez;nї{Z,v5lV (0(-aZMMKݷK/9pAgj!O%02Ini1NI^,vY(:jDܢsQC(2-« u:lщY#ahTPWm9@>Ō VɧIG%Sݢ3p$'+ cVl&/6(r* l>:čR-L4h(kB:P :$_\PtA!AVE[$""P^w"mzv :lضUe_vP'R1'O]?jL>ZV"/T 4E̢<<_7 'Y GK1EJ(ol{42mc)1AjBrL\@*zʞkKmSso6nJ6 K˜pu1\wB_}}ꃅea.tlۏ*] 4鸎~cԾwizmnPw&jloώ>{!CX9X2KqsXNrpo kGYN{Qj~_@_z;bύz0A^k<;r<3..mӵnbh Ic%y},:ch欕F0 7@C.d+:bD&(*4١Awi ¤QM ɥMhʆqgEi!&(ƒiVDzb gaDwy<ȱCo9ޑG^Pd$9#ǃ͂;uw(/D{k1*Y[7>R74}c^+^ڸcv 5_C6l릾=$(b1hJɫT,^FO'pj t:%dUsUkX5긏9lF|vZS!'DlHG: M6a;JtnKˋwsȶ\bm XPw?/:\yGǁ{: F R6l}JНpهslq>ox1ݚ(D\8.V[K`uN"y%~8aWՎwImKh֎ܴd΋OjO"sY20$`\ڵb C5KՁؘFpVu%ғuI1֚C,']:͜h $faz\Ipot5򄆪XF$>Q1L)EX6ʉБ#BlΙ pSSJ%Fn k6 A@?r<%jYknmI[&3C*QX{/=kmjb/lo8pOݞ4!9#W=@}9P-sq\=y~7퀆ZZ710Ơ~dN Zt{7 T]q!aڸt4.};9ǥ?vȤwh׏jTrBqGva/RrZS)XHHH?I%mY&XLo8INQΖҘW7&$hʏD5e5ǘ=q}1sEv2+ȒJ3hQ '4ٚ!6gR,H,}Y9AZĊKu*EUhȡгFePKq6lzᯩbAK13pv\00zЍ' 'LzD/&:$AzJ~כ iw ѡ26참,byum6T@\@C yԋ3(g{LE$k]")r})( 3bۥuBʖ΄a,~P)}h_coSpɕXلs+ rRaX {KkS: ?iokc]lb77^7qz?K?/Z𪏔j*}oyEXhVU}(/ƨ[F؞YɡFɉM.[1iJp>QZ 76ΈdsGSrՔ$aH ђ.:Zh)ADŅVmVE]rn]([h'& +9qB+ۤjQf!+1@|hr1\z-^wby g]Ց5c9E&Ǽ#e0b&5{ xcjaJ(BGXaz3nUSD Ɲqua9(ׁ$ WP}U7V- nmBI`oGWHq/U<>&U8RWoz'a/d0D00vb3!uc;*0uo8E眞170aÔQ}/ڄyof. ]zgٔ /Yr.p-)y<)Kw:]yʁZb?ĮDk/k AcY}T'H^c/^QtsJdqv@8g$Z-Pia1*iZsଶ㚻w^kG'?^vɺn{l"'!WjX߼}3>wPs8&i}r fхňR,A@wh[8RZ͑> M=ݳ!% YGdM-X4.OǮp^ȉ㽴v"s'~ b$wh.*_>ϣO\X'jHJF~&zE_]t|8~ʹk?:wK l݃ؐq:n;^i_`?ꌸ8# CS܃'MdiB;:ōlmld5v({gdk~>VkYr͵<[smV/Ti vj3.嘂C[y.{Cpy񙚣c%!,k?v8m;X=06֣z-$G6Zp^[,n\doRg&NG*[F*ox1 ;yY"kGo+`wt\'p& Xg:Աw N9FN/EMx3.j7WQz.$Ql-jt\oY{@SDG ?N0,?uELX 2s;Y3CbJQ#@*&j>MK%ore2Nt [+FF,yKa 6bV%ϖEUp%rK,[Wh19%wm}mr=^ŀV7G, KW ^ jL7|Ӷ FHH3ܒ7lyEqBlΨn=lk"5=3TRz:}Fq_$sBsɜ)v56Lh?ױsX$wFMoA+}O~&ugj;t)2!jN1z>TޣA O\`^x<=Gݛ/M4[g//=䣷x$r!}k~=mxfe2~yVߟ^^wmmIzv]]U-OL}2N`R,˶$'[M()cK|H,G<_WWWOWWm{CyuM7M/yD?=}{{*D5~;c7O'Qb8v>L";o^-Q Q^ r X\8;Oso[uewErIлŒI[9lU5ݳۉ~sٽ;mWH_AkdF|@woWfeƗ`&+@MKR(5-sy(sM1npy7=mFoQT,.)1 öjd[5\ ||deRn^oscŋBl1u@omքK܆Ⱦ.;T,Z2 !BdA$ mCdo})C gu-/@3|?~~NMj`]WnR!/?v_nF,+/M\rPv-(ЫrQqwOIߡŽ@=.1 (!Q貫P,sc"ݶ? =zϨw>޽nB(,T!Cm>pM(5*$ !=!Dڸ(}a|EY~@ MQN7<śl#*2a7pm+ 8r\6ࠨއ1sv‰l x2&玂^n" ,uBT0J{V ]U]n!JPklL)5˖Kz)2/@֖P%UPzRk`.;f7*mea:]RBkYBkpsu`UzNj"E@Ðxƒޒ/3yhtjΑgϜڢdOgUwwN۝k=ja[2(7Lڇ{tn|%3َkN𹮎^ M=(3 9Kx ҙ3Z2r 8R BƫMchjT q̉琜 Ҥ{tuZP&}{uIJ!,=FP0gQ9l.I+:ʊ gf|Z0kAw|0o &}س˗q} <B>kd[)Y0w~['p}N:;|/ ?Y6Y+RεtMחalMQI[Lūj}| (5vh}Ƭ>{Kb{;y(m\1f_kuܠ?$SZ-Wj:BCxk_ܖbs&7i[&S&P l9ErO \c!*/ƆHt$NxˠaE6 \ eF3JR ReAw֦޳+ {cBAaD6kߟ<;Ǥduj j~ cSRZkB\\Nuz%4HY@ d%iYA=j9~e^lQ;kxX.h\>⇖JZAjea9ab?8},aqY"uO$aɛQ0XY^~ Ke_+pudRPQɥJBXaT)Q㗋Spݝ-&oS){QwFOQ`ꒊV~U$XNU*CTu33Ѓ6lRU>د+mԇLj |yq&%`Yz\GFC]JcWYi4N)'瑪bFeZiVu;a`Nl886"-ErfFik356ud6Kk8w5;_BpAŲoڬ]A.&nJ%4|4 Hk-%t95ADpVn֝n4d|J,Q\}JɺL_uQ: >`K LݗjƩ b)Nr\mmn}DQJFᡲ JJ$0yP rxѬr:W+rbt[y1 ^kd'CوlՂ\biF 7k*c=0aAcFt@^bTI.) Fg])BHPjO!lPF\dy="[_#W/x41&^B7^|灕W8ޜIxCv[Nj-_E)+ ha[_F z t1Y?[]xx!#G1^bBGe<!((̩{|Y=v9vc 隓3 6iRrEISZ%[!Pܳ*V16e>C^IOFx\!#6TjˆH[èl걷\[CT&"!&QT~|0ŵq;D=1/Uoݠojl+$*MuހUAmخ WlUPW*,ح(M!Vw\/ERS ؅C騺!OX<w՚aqkY3Z[8ЊN7 + 3#&1u; WrM9\Rڳ լ^8X[ig1L;K!h䂌4X:1/x<4FvRʪbJp\2 rV`I_n$׍+bIİ:6RXdxTY0K3ЬZ0]V+%h({pH P :Ô)!DrXDlE:{gnY$> Vv<]' u2Y]xjOh{wjZ]-sͪN!NhOotnO ly_1UFS?/PR‹] ~ ꆗ`hÃ:_BrJDe&\)0P/q\r( Sd*pplCAp׻`jyLVKd@wWSku)I`1g%RsObcnԲr4$^V-`!S)hIQhuT} YRa(1d[-AEEaB3Zv\xA×bG=5]o՜@5VY}M}չUjk Qk%w!Yסmީ&$s=4} n}'sw1/1/1/1/o}ZҢp !{!R\IKH אk.n}AZ~1uz}ԵkoM/x6hď+_ex濫/xZytkl-BwV0,W<!–Pu}PUW8.hPd z{0Wt0I}$!U\‡~hŋQF멯(8^=b}ѫ_e!kYM{l] b,ٳ3=;v3֓*4ǕO3ݘRELNHlho-b &m|1[׊訯Fe3,O=i>fέv9,ǃFn/.bGTfzfѷJ;/UV(GI9=4\Oze9pcR*xq5DK'國s62[Mf|/Nz,5ohTG IJ}O6ò}{FTIVq=sb:P)_\̜bo\+Xc{@+I(Ic v4D$( ZhZKJLb r 9&q%\B!$.yR_|<,SEfDٻ6$tHjAċ͇;5~DeIGQCR҈3g8$$0bq4쮪_u=2;HaLuDfmIWdЕdvLMI=[nE:/\Z{b퉗|}6uNri"|g&+XOrCMV>z+%扗#Q@tl5#SSPdh ?5 cm)`+9 7Xk ߥ촺u`i矔aJUner/'jJO{:@YDY|(?E\1]Jȯ8g;'"'ˇEw1[ŧ~ ai!r`_`ߪY室'lR" )Wr6Jzo_WQzhO_PFGQ|,KT* /MK 1i ug _Z :S`k0Q^#|0*΅ChPa/d)!gTH )iiJⵇu8Ccէ(6Nz*A~j!rrNsT<74*x=>7}wGt@l\+irMǡ:錖I'ʝMOO2^T#\e 4!JT#u[Le~F/*% iSqyC~.$^|\!>8:hYww>Z|~V .vs!@ؘ=ܡ ^3 rI;.w쐡6k4#tjR#ty,d뚊+ʎ*O+*Q;_?}d3ٵHhgQBI2AIAq +y]O?fy7sg6Qֱ)~}+o L1(-gȔ5љ H '1aδj9Ŗ3n+}isًU&Lo3Bۏ׬X9mٕEm>v&·#td?U]JbFȡ6_>2rr&A}Zta#;̈6)>=NNl+8F"ƸBWr;&qUڻcx>ȜD\wqk1hQm ܅&2GxGqDaƤ%KUZ5/iyeLSưv~#*,β}_2]LDX(PRQ\[+{o3<-QĚVailO[û wϾ;ff516C66- T-\wV+Y@60!2Iv!sp;t9-T j#ۋ<#ۋR_>gK*;Ѯ9i}V`j9gIObrEХ@0J&]!7{=FWA"#ji ]o44&4g!ۨ9OO#WwvVoyg/aCt 6?O,%OK:1pYyJdg֣K$9[V*JN܎CZ .N[SA4D,iqQNDB8í,:h!ڡ 8'G`XܳAdoQsIUGW@YF8ƈ%xKj7 (- F t&RjH9b`|NDXTzG9";2imU3i ʌ}Y bB]h5l,֗$ĤDjȪ'}I Ih !Izbrme_8sR/UI ts|Ҕ$MJIiMT0Rm3J {u[ZG'[[1X <ܣq6An4wu|t١p<&@-xd1:~v?9T5*{H "pS;!xnzu:s=ߖ޾2F2=hwכx.Ai°UbI"+FI mTBo.-F:`Jlf'9ZLO_$iM#@p7H )I`Ϸ:-Ơ=zECê`[q !b+zʷWj|@/rȊ9`. @&II3!t^E[yLrtXAQ%y! 9R"%X2u2 6PתL^f c.^Y )*2Jxua88~/nztvZ->ކeAҌ2lY &f"M&T~I< iG/I-G5\H%0c{2VF5?dsiǀpio'8 Rmðu11=F(y(+PX:4\J5 !sC1B1F(V>`JR7V|  yRcs8"e,*MLZLlD#hoLZY/'N'x,7o6<~/Y]ޘ;gX_ސ6߀4( $Yy줐aṚ ?|EETu~- '֟{-ΩߢFx!(5CI;4=I:K>dK{# m|=H7I0\" cqzqB=sW0>"K+6t@,*`ᙞGPɞI׸qW4 l F)EIo0_}4lFf7"^3[{ݿz`͵)\ݙNuh|4g?ZN6ݤoPsq59MZxe Y5H3l1)%:xq0J+ Zx * gWWdo~ {n ^Zp F;ۣx.ͨҩ]”Uz.G>ŧvξb[dk%#V'RV3mwi:J#W[4[3/U =c2IFp S_H|Wmb*"fսyFt`& i &_(gk2gKZfʴ=lzlII\`a;s%t\z<.ԯxwNOBP6Hl 1+a n:wGs&?_Ǘb'̭O EњF\!N@tcHX(@|,DN?֛GGjNOuzuYen# #ݛ:%Ǽ)DΦN?{%'^o*{$D쨸+4YzE&[βfi7֞Gbθ.`9unHqJ.NXv !h]$r%y5j#N[gq(R mEPPɍ4ȈqIB@C?xH l:MDW0sWD"ߪ+r|eܤP;k4x괿cLml1P^<4VmELH ?[dS */X@ 81vp)x[SzKՎYAsrYhC( nڲORjIΌTj+ ^E.+ΰ0ƻn8A)<Fi 9)#O!)TH%%=0"s(bv YeS ;~9T䭢hfN'cg p+, C92 s"YpVO_`oFWg -c`\QqA2JS!$དྷ µvs\Fk(Lz\ϩ W;sh5q9SXFzZJ byqFv2n\jr, ӎg1h2"OgW] /'M0 M0 7=X5#7-lJ.+d!f;:ء΋qPQT?j_?~ԼzO:t(.Ňg.&}w<۩܀nt6Tvk 6|3Ddж_ܔ%8\ܵdLQD>Yȝ]t]a'B0r9΁5F9O(8wE(M):%]Mn\̫IXKԚVT/!_nh4Y7kwڗ{3E̞)bL̞YV\ 1NےY:+cA-fA+P KP/$`n݅uV M"ci'BH+GEn Tg..~G]=X*،Pvsd !Ua(Ph&H hT)5E,{3*ΔVI̵+qpCzjy H\"iBA(xhTKAsTˉΐgl8TrSpZ F)4 Ja])tw_d怋l/*L2@RR!@4B6*q DÅ{l1Y :a}3DrM<CXNQcAGt'XYΦcoiM`ier|=dyuQ;}P,/嫍뾴WQ334dSj7q$׸gtp$߿*pUš:u;v N:{b$MĒL:#Pvܢq$B_kwC;]a渑R;HO2Hfι~dRQ|RSX~6 |D3 ZA`vJ:i'q qhcQwr8ݡ+L SW/5ɒ_ZUVF I!_&v!RyL|T- PzMrџtQ?^gRe -㕤Soi2ƭeiGyne{a!6\nnN[0I vدf{|W?sⰀ)v~[yݟ3p.x}T K$e8*~?]B48<8{z% ;?myNa$A(/vἢE]M̌];DHQK1t&ERѦn+)}dAƒhK,K'՞4VȋCV"UY"a#+!Ɓ:|Ie(DP<&f8">Q0@X<2-<4+z>:f)uU%9Es"jqsmhI ۫G칽%wjrǘK?/wW~R dcx/J;7W(JGV{`e}ڑ\N]Ϻ- ;t<:x5bpޏJ3z$+^M"H"yboF;Fʌ--qsYPۤO%%Uhk)UhkP{g(Jf- +PҡP/B)e"BZ$qmz !dTi/p5I0ƗIvڣPmUl/cı̴GazT_[L67NA+$nQχ u<طx~[J_Ј Ga`/Vs.?^FYZL21.7s90M$X ĨUIO}qx/QdF c7}q+p]:)#w Ԙ=AOT5 :=I"El:]25`~[j.5B3x5];G (;gGo{g^bs7x8 ?F;=I")RR$01X>Q.z$6g~FcWчhMWQUdum%cWcFWC)}`?/.3e "u Mo*u)߯8AeqY{˱Aa4h`{{3*0M>.pYbI,V1,1yJ.RN%cjV<9G^ ol;P"`XG_S~oZ$R7YEٻ%U#y>מ6l'0x(H2_wjkIFPdr~P)o Sց~rֻpWT<.ĩgn)gLp l0)GQ#2F0i?&l=}tSX}R-m<}gvua6U̕j0,H|=ΐ]Ix)Wa]fM%ϭVOmZ={Id)2.[nv~߾ne꧿eg?-3s,FDv#v Fz8v̑1q0\VjK\l B\} QZ1&0ZN0vFVMzOHX)f>k0ɵ̙Ŧya^nG˶tzYM*^(VnKw;.;K!DL>W!YHxbur?c)h+1 KP9*V퀊U:*%Ch'0 a5{)#x,b%ϼflVS={;q%fIYw'#4_JaA%I!cL#D#yc)"VpO*fT9rp=e[-MrIlc~sZ_mu3^g-7ϳT 9 usP=vk+(@ Zlx9̈́ C0L7PdghD*u_JHI+P׀ פۉ R7a/ 1 0s!#1< 0yD o#q}ђbh"G'8 Pqp_E>%~^#&!<TPS,*?j\f`3RJXfKH=ט c?C"|6DcIK<؞@qkJJme F>-V<}2#B(庽*NsSܟI*AiQj#ʃ$ i !UH#5q xmXAց9@׸X~Y93t k)A+-xeg3P~g-_kt/$3} (,[w OLח/yӮ6z޿ 2#3fpAo&OP0C P <;ăd:5O.AO7Co:e0L3 *;5ҷ[z͓VMgY gɸL 礂|19n\!>'MXˑIXIͣ0v?ƭLn-Wz%_]9Չ6c?ܷVDyss jz}3JgbhjPiA3wăZ;o0D)lM.zmIMr~br ߩ> 0J-KǸF%ǂXj*=\4D3ε$k>VBHP΂I Ĝ!wb@v5j@Xj' ڬS{zPN䆗#$+9Ib.dVk$KЎClt1G#MB5;0ƹwWapu+㲺D&zRyGq,|O '$TK`$=ۜJP(‘lY@*1Ikh| J裐fN0C"bHeGqDBix\+L=/u} XGcٚ .4 XH)J9n,4p2vv@9TTWq!@)/g2n>Mh't!٥qhc|7BvReT֝dМ AyeV.,e681:-g~Z1 B%˕tյeT;Pw f0BQr¨FĴiN[ R(2ၙ`8- BuW;Z2:o{(w@5LyZ(}'+H.wl:xp Iå< KY |^{IG;O; oPLN}iT|>i|m>qAWyUu.o괗mCs{%^iƼ=ob(]:t5EuDv|F*,iF b7,P|ϗ\#Ƙg$2^#0o߱&(u^E+ECCW^t(jmL{XKSz$.MiDT* &n%8 `P ۹"!$Y2@xy)BcigђΨZg3v$eܴ>G&9NqUe0]7g<]]~qyqj@Z˝[fi!?zAOsxAԁ j46pJ2֬A2@~W9X6؟0|8+$;$KR,7Tv'uQ咦qkrY2k3MTaД5x?FKnvP|=jXJKʞq35un%t64[ep2Ko+fUQSC1%W (ҟTEJ-u뺳!*#W%IuS\1* >=\ͫ"Vw.l<-X`݃m@C3A΄f!NjŹb4elAZGj+*+֪BRިBrX; AD,!ђ5p]AzZ)8v5< Skk |I|}=VX jaJ)"go(3 uU\{3{Z}Yp;̾v(>x*2l?A)9#TYȦ ;Y[~Kad>WS DDΟ&WyM;^9#^o._n&_ýwμq2[^x |iLO$u.7%/[φƾt~=7>ώ~۠[sʺb4ҏ4:OX oR~7OԚ '@WQpLw⹶Jt]~l;#Xjܮt΃Uv"N{Kz(P="~~z `il^Sfeq6jb.ӹg9}f{%f::da*^V,˼]:?#O t6)2X~x_ؐzɇh6obOH0V qQF#(&ط*h!Q5PO0S+a%]-o`=Vd*_I͊ 8'l ,esoR;۳5e ӳUZ=&yYu ³.Օ7 Lqlc?V&Lۜo߾=Bߞ&A,P,0Gi )t@J=J`.Z\%SYbIy97pS X2f ` E-. /j5 sTj ŏ(JpIN7Lrd˳:C/QtQT`Q9Bўig'tzgF{LOT"P^b `aiaX}t"'Tp?hupOH|Wnu覍}>; 04DD0OHؔ̕Pjg{=?y =6wne]1ʳN_ZV4{sjl@C.a4c.Qgda̰VϠuRGjMgt;Ȯw>]РnUL %z'6>:'>v.HR/; b t[ۻqGTs#Xe򠻰ƴ%Qstp A 0 i s mŘ2UDFsƿp[O`b F|%6"<0\šf\@p$TK4 醗ت]]uAUz-ԉZ1pH82!?V>E#A1B##}E V!pDE','_ vzyQtYj|NID}21KLur@ "V^?XX2دJa+rQX-doWLˋ7{S3wkd!jwWH׋?kz5ʼnngcLM'ٻ6ndWXz9g32Uanmɱ+T#qM I9N?RL 3eV9E 1 ݍDsfl27R>2C9Dɢ`OtsAo 1p];w2;t X}xa}fÛq;>-BBJux UKOŦ`TjiV?=JPvvOզtT4Ŋbb̂HdBe V;#7H>ۭ" XUt+T;89Oe@9"'Qʧ#dUU[uH'$( ,ViPT+i”H_LP1vq& ЯW,hӂx,|U|o I_}ϬyIs$D,'ybLWbf27 ~m`bGHT=Gê e{ZVcnRĘ1jq"6f~59AOu(%cƨ<,3/oFm.ͫ0w~e(sE( 0okoҌfݩ 82qOBH|2~D_d&q٦) j_Q%[YIJm^[6’iφC৳,weL!)x LWRe;ޱו;A)꯵}Ģ=./yzxg㫓t.4wur{ !QNtB:m)ҊK/.iK2N-0Y;#|UƃncGYsim?wcTSVmG9oc$? ]ۧKD5%!E^B5ޜ(k{?p=j9;`&8"+/㖂Zk͢Vҕ][ejp?LoJhfu8ab ' fÉ ׊U#'/r;S ;`u x23cVq.` 6l>k8ImgєeV)bׄ}?AyJ S!)r1ɹȝ\xB"NQ.y "bʣk1#ϗ\yvfMثrBiN09! yI:., Wy!۰` AЈJv s+jOʪ"E%a$3js',>517yʍ9[L%:EoE۽fb }ɕWǹV \B2rkӰL V\mkľF "8iP[5u4 0`ǘYb@*AN鼕쥐+hFJuWU&o]attN]"U ^ Da[s1F_hK'g_n (Ѣ9SnLc(,*3ZdK ҉q'M`g% 57!px^DT)& BS֒σ;2 -"+=v΀::/1Hq֎\n_眑~އ ,ov(炙( f//rp*@jcbP b5|K$(:}{p򯘒Kʅb ~1| zk*_ 7b0Yq:dm\(i1nmKn6G`gZd"a?ߵ#wLZ0t)D(qR0v"omG*T @Ze24_BjYU"C$AlzQkh3exSZh[TY ԩjlh5_o,hW 8Pq#: NzՔI?/}^A(%N[-bM*B(zK* ڳ7f[Qu~gg;RRӕ; .ޣEĴ?Dj:>fr1II[K2-u ŘmX<_ĸbEN,swY^DcC$G;jyV)vD4!+ _9%x\;dӮ(l-OҋZx{+L7,fF-&azt'-M$5<8JpAi?loj2N<h+|Z$4㎠BιNdV2;'IʾxYb*s ܇ޛ`¢'wNt❭s`I֑xa~zaiy:kg': A&!ZN"#+GkXDg!:#sx[ rΓwcFJ_Vd8GRty΋'pu#s.%$$HKB[lHJH»ҏO#DېOe&l4,Î#m~;k>X J;䍐' O4_ 5b$l F_7Qd@DV ̠+4(Mw HH?D_d^3{u0u3N,93ֱ.YΥ$Tid-8㵶$Jdхq(oBts})+W7)b8=y}};*xӀ˷y^p~;/?Xեl%uh_¹au%#<1q7Q}vO#H`1qy7*YCgZhEyYV34B` pD*Y3omW; 0һM(#Jʾ!p7UmPyN<Sls(׻Bȭϱ($ r%+|XD[dL+MϔYUNX_4q2D9 *Dj yi5pT{Fa-FJ1 1Mڀ 4$?1 rb#@֟Gdj/$gBWwnF B($8 Dx{13~B3F:\ b5wl -HRw:1Z q3AUڠ`4WYSw]=qHNk5z-qv2g!SypsTg&8^&[lja1#o= x*8FAec%2W.dw-)m牚tΫa( tSVW>,Q7H ~a$A]B쪈*uJ "47=@1È%!aLCHGE։*V-u0Zh]u]7Z"(vLb=ʅ6(L. UH9ʽv CrD9wBE}Q"/?XS%=vV T-5sp 7! t7?d0 w@-h lSK15.L4)[׈WḋwQ 9 an^O3Qk @U3QǟOs 喪xd/Z6WĿ+iIЄ u̫V6"F1ޭ^%:P,u,e]vBo״F_4)} "$]Cv!eDhSҠs`sɀOPBk+~B' oaoV/|m%eO#3dG{9x7zwzݠ?'3z'I#JOMH?Gg֡hV]`+o* w.nNmw)nƗȕ \\[Xqp8NmXGr1OhL'&&t̪ a_?:S=ݻO0=e_G WpuK8/༁D]X:KKJy iE;2W0|%g+ŮCts ~SD4NL^"_GU1#*\l7$>"KnjОx%b*4qgz:_2,ҷ6Ex3}˒CI$[MR/V i[a[&yXեo /g]WK\f]g'q[r?zֱQԢu腿l5'*>9~ъx?ݔnToZ-~})"1`mK<j?/@O\\~{%-rZiQu_ c|Ҍ:ރdMD^~I_pʡ]/1 k+);=U k{/UdFMto,fTTow|rh ^Ԕg66Ah`r+8O۶-|V@O.sCX^ܪ,lȾ"#0=m#Yabw?c ojWv_ɕ>WvȑD\-Gs-kop:ŴRjnu A϶2Pl#è+"8 ,jRXzh< W ̼T!חv2{%&%xSB! z^gl|yM|MJ]wuTWW5SYXEn\y'b/)EzK]];h[{+Z\\fu#f`oz1[uaE]W=Qn.~t!}Sߚ 4?ς~뙳L&#]g6DD檙"l<>ڪt+Ҫ{fN0+ \:ŋ'mJz|RcwOϲ heЪ7 ?ec溏Gzٵ wk4EOu^P,Z8״):Bї_mxGU/\3 g7-@2|~y/>/qQA{t\[):/MQy([fpJI)L6EOQy= 'JoyGN~ыZamZ~{yuq_:/uR:/uv]f s)I39R f,'0DXFpҎ_\K3t#};j@&;!*g\sm&_TD%Yi$k 39dL 5C PL& Q:]'fLtoz0'Sϐ]e+L͙- zQX  htQx“A%8H. k`FmXAG*OO<-)ybq ]0Yc}لY6䠘,H}#~c&9>GƬn>szks\4 OMˏ_\Dt +Xr-pZ0!S3y$d]6ZPQRnXiw1z&.w_k&RSco4{yr|6#B~,g2-̤L$I+zf8EP Fz{zF bB2aFaM9ЃTRy(JOѷh@YV/ s1)B"N/aєm4dnM{r,w_4W#A Pry~~Q \i RT;?Ovs>w-"uz^GTsr%_W#5?QC9!td3Aedۥwyyqԝ11g\ۿ,3K6GΕէO޿~itQ0ny&:SZZ5P4,:3!ZEd@z9e. l:X=nQ4%|v[Cjb`x n>EݨU%m!$8Z% TϘ3`]g`Ϊ9Fbk+b{NS=6m/WwdmHbˢ#oܡ(hQv &,|sO 3p\1OՈ9 yq&azoL{n}cO95BauzQYFyZ dQЮ:.[̭ƲSJ⣛5aSRjqjzp3azn |5-@syȳc?s ekp q=<; ح5 Q d%L0(?ˣ"xA0HX,w&&`G^2!j1+7;ȁ1YoEb%}:rUs&`“Ik Azt0h[w0[a@.GCc6Ru'ZO9W*x,(YBWRĀL!pCR ``r>*5&n߾ hÕٱ f 0n- g BnVhng`Wa" tM0@SnaoԒrgZͥQ_LsTjvor_ #;hk> v}S3VVi&ѩ-gd\y`9in.& YW{vLC)+JB5h%&Zf߽}hKw|:A@C:ҕjJ)^fV\zQŻ]޾mϣ=xWG&`%59J_@{([eLsOa(Ldؓc K L{GީCv=PJبgGɘv.~ljn&򨝅B{Lrg Ab=$|L ]]3Xf$F8p!PCЖi~K- 3b v3A)=rYÃ+pN 1Т\݃$;T ԗxENm+9@򛢽bɁ$7tԚlAo.ב^rט(vvM}y>ghʅ֔?|vwuK!̪ڢZ0yVҝ6\S. Nv[Bfo}fs>_Ǩ Vw)Orsu=zVۛ2>".p I/Ivwtrqss7g_iݼ8z/"""t]UIﯯȟR8!x9UΊ[FJ_K%;'M?֋kbV7/?/姒Y<}}~cd̉!]޾<]%Awn__1Ч_,. W)ܞlj8}Q3w $ ~ksoY ZAJ;]jpwM^Ѹ`qD<2%=\ ,1Jz>Gk#ǘ CM*%nQ~jW5^="po0BՇ4є}Ȇ>CJA53"\̀WKrh'D C[{}LkqhRe˳nb =e4sNWTE2P^/Eh-g)l6)b–1z#*2hF匂2/sG1Q;'/K8@jݫQXTK ƍ.`!RU[')JE(GC(zBnov7eV;ŸSza>yP=zGv)αo!` u@|E4 ϸ'ZH.:ԍiV}SJ6OqW5oLl9_X8bm( yۋt3N{Жg Ω|w,y|Uo5o՟ӓH l6$I<'v2AB2%őbS\LoG{$KPa@h,asT@z޴xX()k9 RUp' Rd`r $W|Alꆡ|kv1:~j;Gmgk 9NvƻA 1tcJV>&օ^{&|pWN9MFue&Q&iP<فɹmlJaYd@$cɓe6N{UF+E!jņa; :u.P|Uҫ{fIUo#r6| qĐkֺ:zeQ;H F3]z.P/=1T^ Wgz:)U ? W7X~CT컇Ac9@=i28b*өjൕYd Q&13!wIˤD/jJR ?@6ߓ%h[C:Q{Tkeޏj T$w6en >?{Ǎ_hlp%Xovl t#2;v[z-MÖ#̰UzJJ)cE\?TN,Zpռ) 2`x[AhocBJ~Hۚ?6Qf\ݼSC0]7gb^\lp(۹"wuMI>oh޳+甐uWu\x#hTRJ+\+HN D+FߊU)E CBEM*-Hrau''gR "sJ uԢo<Fa-N穌 ~JLҒq!S)1ȸ Br|"-]z7j0 M)| GEA!Dw1`{k-=fO BPF-1wZlc<P 3FD.>Ob\qx30&Ǭ'z޻:JҀT$p{.N3tzE{8wbvmE ,ENBYsWN kD$A ! |,rQIB울fN_%2Zx Z?;(]<L;k45e5*3ql)A,%( yx a=/Cl|ٯ X,,&Ђ7Υr$6w:C)\"/v1Lu?L[]7>5{\#&oqLbjE'9jkSFD!.y.sP{kʮIc'9&r.R=ENv1hI?=p-55Zr7Cq;=$<<6@$ɈZFWJASkoh;1DSS=A=8O/#U>?E%a cWƫU 7 ç1ĮSzaovňO7$*`Ս9rوk5neP5^>Z* n \Ñk4bM$15"{$Ո^z墴pG?q3zp*^DbLW"ȥ?t+fㅍFz1x&Vo,wCҗl9"4^Q"7C^ax(:gDIj\ò̑69f|QV\2S(ddUT]W(Z]:tyMSܸu +%gj[(~`CD4+KU","Hce)V@8/9w ((]mWGw_ g™JXH~꾵ft6<4M\I/PJb>oh$Ҋ%1>o^a:C=9i0 *['·oV/,k0Vޕh$9e/զzy^[ U]y$I"%į"#cH"$ I# IRJEA7 1aT"&aVӈ!OSF`9&x3&@1 @>xc?""c]8AF?jcL4џPZɐ$NZI$@Hh_4)C9;A{`Oؗt`fǁhh&4L4C0mewmn5 ;hDH[`wLKVI* oDj"\D&OH1ФP-Q~r䛏&2>WuͬZ-fEd`]4!fCǬZ2eo0%?ugҕX7ջ7SS}QAI="0#L^{Gq([-+E.x>f DKck>CwOD'H\y(`AJpwg@n,En`H薺D5z]p#$kkQm׋>V񓹭yn}߭7q!ɮ1*g&!sԵ΋esT5,rq3ֹc۫ڥ(!ۋz|D-4gŋb ,$lE]LWdTKv]K-. C&dJ-ʀ<>nc(U#KӍND5Fv]~];/讋.vS?n]삨%6,MInFNHH2t:wps=780Ĉ!{\v[ CVɻ3JhPv vף5Sx.-åb" K qks\(f*#g( «fk', 'M411f#>\p VP%YiG%VZ*,^DJ!7X-c`/'X]}! (`\sS_RW\7_xޗZͯ﷭;Goh/Q |IKV>~?\r˥a{6cQMYqMfkjxԭ`zg+^K2zu ,{!@Wgn[zy3CE"]|Yn0-vpUM#Ϯ^kvΨN{ X<~'y)#x kl?ճ;w^,MkKl[~V!)g.xDQ8]_wj>E9^n#,E_g/y٨cr rio]OFrDne;(75R頵A8>Lku4ZW7Suc:~XJ,w?^K٫{[d)\C.71dD{;@-W7;8sY7˳Ob1/.S_8{š]l7  z0 M)TރsyA ţn*W2ǔQ^~[&f  =~,eP`-p=%wFfT)OP9g\p b>g?sBR=溸|b g\k*x…XHo=Zx;;qdMlh)pD .1Yy?8R,yR~0(NZ{_<w'k99r8(qG;9MHu+NABJi﵄48ffSL u*zlF~^1޵6m$ۿ@{UU%vTaR Ř"Ro@RC"@1pl } zzd ۥ jDm=oN/UA]aM"]Z;XyݨJޅ(FQ%TQJk/7*ҕHmp1pO958ae3Ê^fn_Q1!f#+KwNb H)f'gHXV3)BBeH^(bGM_ amY!$BS@S>*$B4v*.p^rss67iͅRV-|ByƧ_WQB-+^Qdܞ/gXLؿ?%!3Mt?#wefeE6?1KiA2hۮ7w.=INUD+ k8l%w$H2䱲~)P4%(ZJ:[Ȝ([a $U IKaaa+Yɯ 4GgR,&3 UOh%Y.n",B"3K컎1Մ~.L5_݇*-EҝRق/KTAI#ץNc94p $\/IQceV4JųG lܱ8Æ ;s/߆;9} ytOk^VCG=wttTSTg7_H2]nilJ.hJ%ټɡ?~bd$y>F^}T)t?o%^](^^l8 T a ٨5S΍rBggV'EΔ]3]!q"Jw *k._V E- 3N25+=f [xS6Scd з)#BRTgٻB*eٴ4j@"6Q}|3[zj@ğA&"$BRĶ||El[s_ C[Cra_F- ,VKa|6ݭLb!j#fu:6 c[b)|9 c *7HevcA/S7$r-b pAӄl*ճ*=dwnΡwfXI![5:Şzݒح,~o@|d6b7"ٮq+;_^覞moeꁪdA@T/ 7HJ+ƪ[[\.K,[G=ZQA㶵FuhS/p`1%z[!l:'xgޚ5By3U$j+5a6~ݡ h-ȦWjVP\XJb[@((yy H% 틁/(0_D}}ׇ>. (sXf).PyԈ\ D ==I"V[eP"#ԋ&幵gv~ ˌ!%.Ydw,J X?c#K2vZ5-B} r9p2NxKu+^C% EsRX%A.L;OjY$-v ,%c{T`fc;(kHamSt@aF^,PЌ.]m_5L؞IYI& Ud^tXBV9;:$椈椲)eTƒf!QL{Cr~jw:3z401%~gQY=/^X [eȭJCʵ(X$ X(HL;hr(x9ZKcJH@P31F :`2AJȜs⡘s˵OS hD bH kfii%qUbm4FW}t8ڟO!q1c!D|ux 5ġfQ3YHPL AsL_,<66DV[f8*6HXa-\T!*!&I xF!80J JXEBv;V&j#ߝx+n9QyaQ` c#ƍ?mq\Z_ܲU/+bJGncӏڃ IP0D29S'EfUk S Ś0 iv8Vh+6Zfy2&_!UCibѾnj*WxJm~~&! !{Zl RyuYy*-e`*vXuP7SG*rR/zC֡{1߫WXJtjlf|~N 2ZL.=SHᨼG=C(FozӪ]S#rTϓ+ЍU_. jS`fޘBxKƱ{Cނ f0tr7.9iC@<.,vfU~Ǵ -#4ЏIǨGߕP;L1ེ[xdLLmtϵGgȞRs 'V$T|n|E ISIIf}Jba9slF17`Fh0QDF\" +Mq" t`BJ-uP%\ ,0~,`lj(MD Fy(A 3h hj9 62#MĩQ hJ̞ i6z`IѻU,&iz퓯׍V>5[{:~^AGfdri5~+<Ndf0MqОdGf󯘚T0-`pzv4[Ǯ(Н&|b::Q>wMh!)K3kyQtQjbD6%OW׆@a2I*(5 2! UYX|CbWH,e&s2<.leF<&=J${Qx;^_.LΓ:N;_~:*=r8h|C 1! )Lcbd >`ʱ@HQUS8a$*Y ko-|0%$nVA-*5B *`|7JnvTk=pHOpf27.~Gv<^ݙ/N#bر~.{êCag0RJF_WAզ BqS' ᩮ:]"!/Jz{;񿓿ߎ.΃nl7_ԺEGo'k=՟jj*<: A|nzv(][0Nk3QrB]8:\*sx 7\Mw7#{!$.03臩w'މEv¯4Ȇ0:е{SJY orWY]&q7 yQv `!ME- BQ1 .#n:%qɖwEuj;{{1Wm| bCsF0O&w@^7ҽKX^AA4ܷܯȁ,lrcPZwg0.QwJawJJiYd|3#rK൵jxqvt4t`<󓺺7Wg'InԦ_xûӛ_޾?x{&e5mMtF;w'C6=X~0$!X&jՏiv|nBD]3PX;76n7J^2.9q*o`p׋qܫ3;nwlm9$47"|p}[<}3ݪa1vKvp N6cݾF6]M֌1{]?jYkHh5i-eF_ꅟhL: LgJAQ:zD$R80%4,g:tQ AohӋ{ 0Fk4G 7 %!&,ŰIliQXUXJY\q'IfenL5e)SMjTS2Քi)NO[Zf^H,@e+W5 iXMjVӰ4aiX&&"q&=}0{xvHQ%Gz~NB|SwrϻI]GNNADfs~M TW]et+e5k˜&=\K&1$ZIx)f5]fH;B:59uwR4GS0D2&%# _|\r W+g4ģ{%-[K ujI zszSe>Y彳,M;v '#s}¹X&`D41NAR}C]rۊ"IBA/rRBe7L5~H:YL>},MNE}MVfwNgDcHz!` \uUW"_~7X=Q,& l,q(qLOlABBhK -EF&(FE̋D:V-N4wW_Di\3s`D9zp_[E ~s13d3n|}}A(19h$'6VDk>V>V>V>V`/9=4m"ETS6r&<)-#k;RVO*G -Ɋ1Kн[u8LN/<(NNn+̤jNƓoBwx*=G{lO ﳚxoT`֞[ӖW!:INu0҈~f|oKGѬb~P/0W"3:c V}ӻ'<̓e/^~BƑ6c;ݒTQ$Z{X gRE{%5vR̲DY]gxHں'n~ ˩edUA!h[>kM;![< K9MঐTaF'o NO'7з%~.o=MOozzӛ==Mvz,I&4I=MiROzӤ&44`՗&K~n,B"t(XyU N&҄=K\}i Rs:PD*ZBi2KfTfǚ{[LTPKUHg`kmJ#2U09$킧8NVu2c]Uo%(4pojRèGI&1w:>H%.o_ J5Yku5oO) k\֑EFr6~94 bd+C$Ǭ]~ Ʊ[J%0_ڤIq{E\̣3'#58gE #NJ~T.(CEj%>xxV|WM@&/"!.4砐6.:6Z&qILk%byrAӎV8Q qů~΀1k&3#N3Ֆ8 =#ml "t<4qpskRIeD 9*ϼ ZT(Oï+P_"7w ";'`kT NFjIlʭj.ae,)VN{Sit#1MW}cFsY3Ņ8r8K16x%pAl^,+O_YGQWHQ̵## +eU*bLIo=ז``a,`MQU+V*;\L63$>T>!K1–QRܠׂr78<8Lfwqq..d~iաD=D ~^79{`A6Hp???geJP,fZ1aQ[<΢927%<2@?mxZ8|*H+k]nM1:Mۨc.GrAmu;|*D$#c]nM1:Mۨc.\Gjr6xU)4 W&:Őͮuv5Š4mu&Tqq6xU)4 WF:ǹcݸViC *mL/(SJ^@?6Zvl,cv:>u֬LgYJ_H7l_RlO!ه٭94_O½iׯc۶4/aˍ~DOm1neA2N%3E_n mηitT޴8­W"XtlHJ:AbPEuQgѧ$|NZպBCp)UN;id87">`K*Iz&3?z|b7MAό¬ E聪c@U@=cAȀ⥪[@Tc"qCӏMT'Y,t1AUqM[XrHyfjvC=f t?RHAg}ckHcLjI`(FG&'79lwD)z]vCC*hwbK\KWsXe%#;X*kVZҾ|KuYR,Qaϋ5e@L򴼨BKƋ5i~}ȲOtwۺK)a$&{X^aEXK7=Q2~ƣ(w,LqET =RI-̠Ksr|E6^R('`Mg0ĘjIU7̕qa<",T4o,WGQ<7ƵZl-qz-*Z!ɤĵ֢,Cju2pF o5:A$f;@A !;y0uMZ/tײ:ǢltLtC X ^f7cB4`)jCw~ J ā"Jy(_^tEӠKZux|=B5GJc@p;aA>]O/|EO ;+I $b4JPF¡Tf,k9%,lF/.z-5XZ^ná<%[1A^o@*_ìԳ2t#gRōi8G@/L2w:̮:sLWG234v5Js:jD/HYc%4V}*O hKc.G]Ylŋh͡M}%2CyH,?ae;_ Y{c(+8rep30z8s6wj45r\V]d vg}J%ƨ7gyGHd~߆=b1a+F.ӱ{If1L㣌B,C]G{eIn$1SGR rEүwC[`%V+]P?e>u7LyyYU>K9 EF\`SڈQPBmƪLXڨkAzHXYqGYۯ52/Pf1;b.&FX?Q W]|2?^iv.dz/Q`SזbQ^Ëcc 8%A.N&ncG!R7ĜbҎ\T5XHalD~Q"X`NhvFJSBQE'vtE% +'I;V{]c* uu}3 p xZ\{'lRx®jMBaN>;yfVl}뀀 ABRHуz}V i=|p<+'YGc0,$̼hn/mJ}!; w_stsɠQ&W}VwRLn3% R^)LCIC%tbXi5+ʍza.ɋ Hp1rrh !ATut !>H>=@yWFOC;8]ғ~R0{jt1'// qonˢ[B4%4`<YJS]хJѦ;%;v!n?G4n*pj(~b73Y ĩP' 2j@ɥ+=:$d!wB΀$w )1iGO\ $}6))Ćho]y8Mk2ܙ?IAW(?b_~ ɲfHi޽0CS°=AG―}P ωFzr@ĺȞe:ud4AAKIuPk|':A&i $t5YBO^}3"|pIňC8)~_l$!` yt,Jꖥ}pa=$8gPx޵}Y{낁$u{wu@&6qru0HpOfkj ě;T9ZHw^m[/NrkP3M[hl(|^=5EA)q_ָxA1YehRݻ)by72D@H # 9y[Vhj@S?t."#4du)\WK][hUh+f&`66_Z S<}/ٗ ]GxW^-=N,s9Oȋg>RLN1?JKg9: V:O*)gWKMrZgZ6D_j *ø^%A ZQM$%dN<N +E C8:˧Hq]UѠh$Y~v_q$;G'[Ɩyw[% CslwAۈ@vԓQމ=|vQF>wGAkԴL:B՝u>7hA:v+L=.9NAjc~5۷oƄaKAď[Ɍw a1㺶O/CI]X9]BScieN lv v` =AdN 5J*>bj&{ hAfOGz4yO#\p!5_G?a{dI򍆿׌Z" 0ӕދx{; ݒ)]G%$}- &Aw- wM6@c[H!gdףSPPaN[$vǺ& EMO!ϵ M* v kd2ފ=jg;P#iR=['OqswCL1jKˡ`?AK[;T!PS qSjÍG_$BF.H1՘\ٟ+Eh f2Ncվp~ܒc`(D?>$|J~ adv-b47s6FyjK fՓn\D/=zƔOk31?B rPu:I(;Lx@qD٩!ԐғxZy3WQo Lcoܖn6R])a*uAƮby(EUWa>Nl$%=} -V>Փ}"JXvG8uԺ؇$Un:0C3\_-*Yڝ^Cw4Ǖ>n LnQ |"H#w\HQԺ+B΄Z.zywy.k:ȿU;kA"$s,$LrA)G2321+]>lmE*7V{U)8FY%5Jכ| )FQlR kܾqV]KASѳiE`$Sf"V]yg޴P)! x ސz)0b>G&:(P mWyfw )N)HMiJB& )Zf2ʕ`Jw>ؚؓ+Y7F=N,%"gS0Mp%A%@)FZq2P*\h3 A}R=ZLM9"Dtyk̞t?*!yK4'`vlis@r*eN#垶AĽIG/2QT"k`<ʡFsNIEzDaqZ\QVZuѸNG4[$-$'lN'*EKT qkz mLrnЦs8Xqcy0*xjA(QBk^=:(PKk;Ku]@ߟǩ@KP*feBRP8_|kEW.pjj<)zp A(;yΨvZ$Ҟ3;DDlbApȿ/pa Dg%Z4ZqFLQF$sECpڠ((քH (R=F)v7b=kRE Wj$;C`Zr<bMGT"/ul+Q,Z%ٝ،`S@-yy 6 r̫y ee tlGO!b)VgH v顰Zjr4QV&$WVC$RBpge 1+gv]CIrI\p0AQ/ur")eY'r I& 0!TVL+`Hެ_\9ܟ] z7T|0ky_:y1Wz0(MdVd4d\9̏kL:Y(D,\oMuI"R3pٿ7_*a*W@b̧ʱp2rexk79-g螽Ǯ$&Zc8!Y(oYQkPtt cBVy1A͘s"}ձ:'g ΑPP`.nEYtVfG2Vz9hG1ήY|H2?)96"BGS Q kpT\LUm{+N"{"?\B5_8t.g6I0JfW})Njo:Jz' Q S<' 9BRt&aQA3-LpC vakk֯Nt[4 l*pEGyO!eJ(Tfmfb]Oh>R kRDR:)51FH@!CfT2zJg =nkhUE[WnnyK9-Jw{CˉͿ7U땫л7o81च`>@k wsa¤[͗|g/[ g7 b)]&gw.ɸNb.~_zr+ kЪ^xWp=f 4ajRwIat`Cߖ*Xŕf]kN!ⴒiQ @DX缿Ȫ. $gTR $$Dd<<aP#lx=@g;ߗ eL|q?=spS_~ p}wrY폣e"Ulรܼjm OϚ5@J*z f'gM\K:ғ b@ D F.%I['ٷUɮFAVr)i]sLeT3h&9@"|[ЉN#C'?Z4!e|tF:Cs@.ε;"]WX(gq \SR01i fr_${u뇉^=T.+!A[vؚؓ&H+ecVn)Ltp|A?CBrzkt#ض$1'uzooE?j#k! PofcH;Ze[}TlOʑ-LDsShCO#"* ^;tjtiJ NoWԩ8L;0%cW=x%@,j3`s/DKb7{:^7!/f"y#T(FId,V}ӻ[Hg;Г@f"O9iޛdh-ћl%]$?}+)G.mVq*.xB9K s)+*e>ӈ)f&T"dR6w9o(i.S"ƠbRT Ǔ0)Y$1è?fCĤ^U1a%&99Mr HB&Nz&GL+&TȜs+@r07v`s` ""Nӌ'& IL*?J}ei0u'1^謄s Ä {HQ"qWHu!E:K N)r#[^C}_ ɞ9]}${ݯ^C)# Zr">X-6UX-)Ec0$(JuAN{*E֏_EH3 q &C,qJ(-|!ɢwJx"@(ԃFHx ^bGDÅ SIvTi@(u ݷDgiK ouO !QeQK@*œ'̴ͪl FW{pE8ڗ`.whJ;i RJ (2 M򥮾 H]6Jﺶx,/JI+snw-~8F(``<H4Ji&!0$ՄabXP\R؃ ;u" o7q5HGHzؒ,bl2CAR a27D f=8 `b[QoE8d Ki$H9.壐+<b) ; 2`-v\&h7F}궽"8CvCRT? a@:\gqV @ὥZg=$aI{|[-0QAhx1+Fi*H2C$V#@vP!#@a#s0:6nJ0NPjsC\yV!cbyq!.A;FۧfmM}^4aSEp]UA|r0r~E\ 5Eۚ5&, P(ϲ5kLת5Bv d8Wy܍zx_ Ćk^gY?=?4$d8]?=\j9Q7T Gg}g^X]u{E#gP׹?PYh(FR ߿mj4\E}oϝ7ûIVO.UHQNE@hlR|ezJI<ڜUPx)j5^`3wkNY|\ r3ff~]~Te,d׬1eW6Ӽ!bOCI@*YνS O &84=/ǯ1|8( MN6EOEoc *A0艺qB/G"zE ͏"AƠG>#;Iߜ,zrx3)->8]}-iJ}ԿqNnHMZW?(]~iz 77\MYHUκ`0~[dj"{:5Ļ ²B8䙳hOUrFz3v+A딮Dx 4Dnŀj.8䙳hOU\qk7U4v+A딮DyкrԷv+4Wu!ϜExmJkq\J:\n %̜}+Dq‚z-%.h$=6(:"&G`>) g gp UjJMpqmsT NlBYO{[lLXgU ~5Kqwd4OlEiЫ @ݷse^_6gYj{Un^6j-U T (F.u[5h1S+Qy)c+˸=td kv-/^=c3px2*08`YAjBqM;G0#œqmK۵-'k ')TJq2 MOS;QF?Nfߜ$)hn4ys&pHQɌ3QmpN`)v gS1(YcbD<*Ft'oǎsf^a׺-$gB$ Zc*ƌ9LDAHL!2Ju@9\aсGN vMx;T,graܧMJÄ)Ihyz濬B*})^{yc'ћ՛ _2^/ߜDyO}=roNx;Ct s~3I93W4|ࡽ%΂ӳ7'(Mxd]C;otR2\E%mV($IRHpuZ{ yT`1Esc f;Qk8 -R@*hʽ>:H`""-Y DYߵƈXWMpƠ ~+iYH18c ~Q JZ$AU4幀+(Yn©Ƹ'{8L),ǴR G*Yzg`cv؇8L{D3*3QGGG8Ks(QAS X'%Ő2B AQXR!v -eȻq?{8Uܜ|đK*|MgeI|eM4[Qfp E.ɒ7ͫDD}Jɨ1APc.ߧ0YHA8@A'Cn,6:(X~,˲ohcx }H)Im.H20HB$RJ$$J 9t"%yNDJ*}>tccS1Ryu(~WK!9ߍ(,4H"GE6Kt%s6JeT֪;h&2@1ƨ[x0ɱ݁4vJm"Edaz0Ep*N*ϳLvu%e ~~UueSO^Ji.~ PH~z;wr_&l2~?<W?/etF_0+3Q@(iz#B"4Y@-+rYuXB\bwrEO,r@LC=?NJr sheCZ `zc~`_~pQ:^HA뱎\= o^hɤGxA%.s¥4 К92έt1H8 5Urs\ݟ4$*ʚbP$D,& ESo,3UllқݮH=<^|]-??AypZ{76ˏoN^b/ү}כK\ϻ{7tw |ޥC缡 RR|^:C缡w ϻj7NJ v Q 8ESJ>_G&Cȁ/dzSOR)m9*{KkI %{> )f*Z֯$[b,ϳ&[ ]'Pҟ)?|ֹbUH5lPiWVFfpM #)-$pϘ!^/^ྚCךY;Jp(~8_jY(xDsim\ AS1R"Ԁ(*dp (TF}DlzIAv k0*h=f:5aHQͣv/4@cJj%=ZnIedT:= 2 %`~= !5dn'u~`R3ϵaΕfQkI5nuX%s9^:4t6[ܤ4H!QPiWW<2w͍% w4T|nR[~OwQu#Ri )iH"3|ZkӍF7Aٝ=0 GJVV8dsk'^Is8o{Nh,F&{J!nT);D.(w}"ōJYB̉3(pdnygv6k 89މom0Шщ]0 D`9e *:fYb$ s\Q<Ct6Ӧm]hZˆ:CɆ;thĉ@)5:KmKAjPZP^Nnhz C:<"b(5Qsa[:o ]m-tZ9]kOk܋qTXřKe(% qf(?]ENv|u@!t8sa^ 0J%%O##iQRNq@:%"LpT(Eh^s墒F "V IЯEH'M 8 IVdmXh=;:f.ỳ Ԝ1GwvUCz J4 uRӚ!$p;qZ9i 'BCjFUC*nƮ`v";έ]PԌ7 ALŷL߯_GW|z ηwl(Q _Z~9&Wn"JvsaNT^5|.CF>ޏrm%LT=y'.P+&bD\ v6z92 5k)2\WJ .~2r67FH!7Ai ,SnVTȵm%]]x.ܮ#OoѼOc|J;P`As-3|cyjR˱c:2HFK~E~1q[90෯Μ/a|_9qpq53i4Y과pr(Hئ,#HYu@2GuΣOfaitc?+et0:iXqIſ9j ڢ˵e?UW6FZCf|[ù߻;&ͣ d>oH]X_[g5b'u?l8_ öv(g[;[Y,qע{gXo0( XP灞)y.tѬ0f_'7?y}vaI>?گ~_G3|{W_)FW)h%X&%^_ :we|9V!>ػ˒W[q֐*rw5KX-z3x3īM3҃~4z-8k|Q=ͫϼa//8P5 Ri`@Ā<6>gpٮ;;;q~<\??Kw#e[*e|l/N6\- %ī~F|lKmw_ȩ0Jsx&b.Heg8\np;\=僃Դ$LW;՗W:Hyܐl:$MV~6,'/bi/j&T5!ՁI fnĹq9 /"tءq *7OTYhal64 d͟"ߛzRlÓޞC7d"}k Gޝ`0bEhL/?_d'ecn!u{KOA0鼛FTzyhz{;.,e.z߹ t%;T4׼vJ'+a|Ŝ&BZ j;)K;lЕJ^C\ݖ# tAIv;tn$^{1b -bUE ^]+ɢjՁ|.c-5ISIUiNjKqlz :/Z5`ćȆH7#2%qBbt(qV dPD'F9#c kUj-Ҷ iU]Zֵ= +-y!HUVתhn r9>p PXjF0ﲡ6m-vSH/;)Ft(\]<ගn _ [ݽBPf9>^Lh,ruv'bC>+s< ^O/XeqI*zd+* :`iNBp$N -8g@dZN7 ]R1L6`ކ|&EOIҿ̞F]9U95$Zd69~*\WF(! zZ0`#z߅^m~j0A0'v?pDkyURa <y>L C\ x0w'Q=絼up?˱smzmm]2R`lpKyJRJ+#C|jƬَƖw ;jVu uko}3bo27?B磗·7ommA)7?W6JjжWP.i6lޮLXE=)ML뾱 @QCQcIk6@<b4sVQ[O=  N8)@ 1rab.ovSzo8 .mYoFKl.98VJ*BE^[ua:z[z=LoeɝG;;.vZSTEM`_* 96αxH2:¹iIh BGEDr[-eB8椴R j3PH}B2Uv* OvTrqY~ H qOteC=P?OT?\lJ|=<3qDV3էce0-%ZUXU";YK3hlnVig9pʎ~~:WBtn0!9"fuW>)iDQrȝ;˧OxvdǏ_ 1}=,/~2EĎxx5/:%8ŔZ8 8;БaF6I/&a6w $FwIC!jfn{}8>rţH$i+S0L 92M4ZĚ$SJ€wT8#K OHDv3y<Ԓo\VJ=.PZ26ԌR)P2x QJID`9>T0.Y}|HƬ7'T > L qnjSU6S /4=>@箭umYn AbKݽ0Q e@|KUZeQ{ 9|9|㋧2׸=\_ F8{uC!_jp>ƿ8stN*8EWXeyY5w_Ձ6"2t V{^7c4*ƈD%0 YebMQF~ZI  5e&G=NzfH 0YfO 9h'Vy4$N|")TY@h(m)-AFrk i~*S%VTҨ.Dv"}#R xWԼЂS 0cYFAa0l+~.(m;#=$'O+i?KYcZ5M|ktijZ(M"INHAV8"a~ZI΂HVp=)-'%+_+nhjF/-.$$OVB* (Mc |twA\ihjo:KGroK7cF:'Z(0a<ڎ/%6Y[RP˸MĒ?'1%?j魠LƇz!,+ᑘ8g$s!F IĽZ0o:Jc^dk8RSZ BP9GHw&`-KOC.?{ǍJC/ l[_ i d@`H1rKݒ TVꋺH%M\"܈d#AlN$;?y;- #[ȱFll 2"8Ӆ8[qAySRWx 5j͙Vg+X8hՎ )w`&P"˨e͕ik8V;,KNlC դdwv&-|.Gx :EmQ%nJU*;N%ZΙV`X\XĤsςus(qQ#N5 om) Fry* 0J6@WzΨI-wk|^.ϛt ?n=GhVXaMS֔8=獯 HvGsJ;7y,GMG,a^ %1xbZa_DC@(tahyU.\( @ʧ(#8{lV|rR @ Cr)rʡJeFh+56>*RMq9 e yA}))Qs0b+KQKvVcL5fRAZs͸DUB/+e 6vƵSyRE`/wQfi<~!a}ϳ iJ%. O6ˌZ P0 ܞRX;Z͈BQᗿ30=(,A#M&xGG6?J(LKq' bv4Q_| ;9+Kniʐ–emJ8xf*dVE5<-Y rZ5á|“!y KqWj(8_Jhc)޸x "e-`Ynj]a 9k ts }{#ELUm/">Q-(%zz=uFH]osyyiۜf? M.[U6vph}W?>: Vjk=0nQ 9AG1pm:Taj ℑ!8i9 68;N j[d4t>!#ҲH q;m#~v{baPWN&(D)> ݚ)\Zn͔bk4QQk ۯN9 .#^E5$dbmz MMOisPH>2Nbՙ,u" x\ b0R'45C/F"Yr50ܺTz #xUnZrj- Iϧ~x꫹mZMl~^[GxݣXdYo!So~QhrEf~() B<4K Y~|%L/Z:2;5L&xu*no-{o5RUve EQQR9~\ dcW#V~+$:^>jI)tU`7FM¨Vv꽷fPc3v}92/572QCE궋OG{-yl,n`,־VWIؒ0{E賣^IBp+]תVk@?1Cq:Id{4x,BՕy 3) QCu?]Cɋ^4.둓nFMZrfrFF|IdΎFuIh癎%|͂K& Q6bIx%= 둥!J_8ޓ_~|DU1_Qv락 _펯%?N~zh$৯B>)man2aDAR]o0|y|{ۼ_]oeqt(tCߡA e*qpuxzpwvX|J:J3aJa<}>5baScS/Gi=;goCc=9x~;K` pxVyTʫϿc3G(|,{{>=4{ ɱ-Gywv?Ě#,?%IkFbm C)HpQONa3*U$Skixzy`oI\.P mu 맏o(4:)Eӎ)ɦ]I( 8WJ>~r=IdesrH"9eo.PdIp 7 wJOO]Y%8uAgrlhYexXd))]5mRb$jBWFүD؂*,A cϱ{HVm%LT .8^^(Ydv1`v)ڊ@FZ#7bFBSi繰baDGQ51DNr)riQ`O m3>?h}5 +S\ Ij#ԃkJC] V4 WzY Q*l: qR,1!.î˖v0uR* sT^ <J)JWsìS*rJ^= .;Z͈B еnk=C <? P !XK<+K-' S8HB`/Ng nDegP5ysP}X;m܇QCAzdaPP fJ"66É'!6چW97P!tk*B6Tmj6lmC%z9kk `%Bq`%)Wt& f K E6TZm5 kj|~C-چZʳ j6$s 9qK$I+JxlpU(!K`wS$ u`A x <5^pЀ*9=l-pan[a]Vt+Fkp0j>0#Oa:G6r/MW%$ب>ju9}RZX?Zڹrڨ:nYT7`Uo'Cȇխ}hRڥ}x*cnmg{sw]Ζ7mD^,c-IP5]UoWz_q_oFSxZ{/`k~~~̝/xx?N?Mő3 dVvcWjP N7hNn?ibHșhLI1zovSHc:A F6T޴[xv!!g.2Q->VOܞvüPbDtBF ]87PŐ3 ]8V/vӵC1hF+p TnVMn$E4@BdѲ=F8[(MD'mۀPIךH-nP^9!g5)'+@#"rO&V]#k9s%p'YYˑ+v c9G"&`dM0rd-G֢&"kB#kgYOfMKrY$frz55ΑYD;jIdZԕZdM 8Yˑx+ATSOf FͱʑY:geYQkDsd-G֢~"kQEZ̕GE`ԒYˑ+AE┒Yˑ+S)YHpBsd-Gb%Zd F{#kgYdJN.&$"w̑+A5),p^߽qp 0wkVjݵl|7K%HwemItemzH\DxlȲFuU5.$OoV` @C*ˬ<^No{, ŏI`5L~1L(Vygf>!Lٔ! gNN=1ƒT'I`b'QAmt7C#`wf4nIx4ov BብK :NJW(pBxVx+X< ,QhQTc 3J1 fQr B+mpISLJ5`.8& aiD!5}I%zZȜ%X"mOȦT(0E<:lI5?L&Al,U!su*-Qa2MJA=b$Q -pg !WPLJ%38JbR,5 `AB$$[-?#фSO Y|,dɷ}os?G67-$c`e\L&E1SD;ohc>^/Xvp{ˇwPZ`8'``  *T[,8K0rt"̶PZBڥų$*5JyHl(0ѥ 6%` GeOf@ +rny'i[/|s o:#OR";?Dio2-Gy޼8D]ft5}gABLb߅]I-D'lf#\¸MVP [M8a[pF-'Ũ +<أ<߽{t q/r"3g㾱a&@ALH"`V&b? SSLC`(hbyOa,1J-|g׉#Q45CȚaE9E&h&*d L1ie6ߣނ<Nlq(?x+\|jYFz&|-sk'FT^>Wv37v4\`0.O`(_=W_y'k'>l$~b2./:AD|_{3.`<Ctqpc&w.FF `cyhlX) eF?t=Qf&`bѮmPB4'l3kaw!YLwdV'XvHX[r>9Ksqs; 7'E'Z)Sn u$⺾E"hX,h:c];Ozˊ׎&%!RUKwNFT//8%[;DRJ ]NaFef"jty3]2ֲwESL S+#N1\#d;Dkb'bK/ۑoP Be v[ᇕw|4BZ:ÄbϜ XpP|e"발 萉%Jp\XåL y6/6rNpGCf`N kV#hg$ޚpa* KzG7`<2W3Kcl Ң˅97ae3A*iʄQ#¡1;iE%Ў%5Q PTe+ܣe1˜8"x\uX>33zNo./.فzg.ppB_ȱ_'(6@HZ=R ?{3y7<ܝ5O~{RRKL * D}npu}~.a{ӟ7KINu {WfݸwDרT,Qr+LNcWه+GPQxb98X2FgN,ofrOeYZa485o.B\QH"wժP?b}tB6L /]H.)z7DR%{]иԊؐ~?_3?XFخ$Cs?30|Nf}qP,wϯdrh}+5k6 Oq<~)(Kԓbs\$㢷?~Uif8,;_fY9s4A.[\"ȜRK(M~-x-K4f6$mҹa&lf6~2+ňR>FB֪Kt)-dsS:S?O!|޴w/3>@NFYlMמ6nZLPr-`TRۯ-ʡ} ɽ4;J1 YlS" MO4W&(ynN N1^U-v_ W5'y'g uOPZIT-;iMkiRͩU \~h,3EXfJɯ?.m R$YЯ{z S/OD#[)I`ak7HK"yRiH(-Ԏǥĥ'mGYQ5孻Q<.$ZRRtMº\[FпGr+3s~]C(,@v]T;]'KAj^zulNzHDy"V2XJA-m9r PXB-uѯ`._UL6*vczT[֥Avv뷯ҝ?Y5?ZZ_w%SZ>G,[ E.-񎃕#eݰ1}ISs,YSf)u̦V&KeISXzYAE8jZLAIW*Ђc)giB,{I[ǂ`C a|BD#45)JkY ZE꽪#RPΫru .TW'P&,^E-k9tw`ngr7u4Vt-e%0Do7Ǻ`,CDl)[%#yVI!%vgLa j#%eε Dm0XVL#vLQQr!Mq)1TYcZ`U5ka*XfQ!ZN$GX+4ks!RVF,lh7 u0Hm5Q5ֺ rp \՚J-$!bjS [H 6j95XТ Q\o@q_fгBv1U$FPe6w?st"1hVAlؚa;~f/rgQ.۽A /{7'4V0Zb:[駭L'ٗs6MK`[Q~_ǐ-ӢͲXLAe膂4^Ts1u--S !,oSMMC`IDjו돡YJUy6~\`Ūl1!cL0dIq B vyDY[xu3HCJeRib8KqJZvAAtZQ  .\~p'Ov ],(1CJ:&< mIeONE;,NU@][>sƊ/%($=}<4kPl"x7{6Q.&= pjLվmj7"U>m\Xn}RXX}xj":, SY7 !ӷn<$B_+aoa30#_|.sk,1QR*LY]z&:B~F}z. fr[8%"qLğyS"Ct 㓁n7u˻ 0v M{8eꚖJ b/ڲC5n_Cu[=j;ڻ- jVxU4nlT]zWZϾFZܒ{zy߅-7~,*M .WmTL7^oՉn] ZR~2͂>у'dlm=+9cb\V9%[qWEDuE4UYɋ@5HV5Me^Un4QH#E0>:KV3vjM9f6Ag5g[rh*zG yB1TyQ2&N-xijogXCƱ=kg#Bn8ZȇVrn:JT3|v*EIm^kLJ*:},ѥM6P;̤<|IuźT(4 RC'ZDe^oN"Ax2kPbN(㼩ͅ&9)>EZ~Yz XuttFYIEK273R)xM5 P a,NU#E^K 4X޶,>X'f`39 fW@/^1IӦHo"K8h5QM!@ƘzB^tz-Cv_- ݷ RX߄MW(no­]&H(Ipjm(:adknck ]bfùZgW w$k ',gpChc^]1v͘l'Q,'}^%(txeiurNbJ3h_H!l6S/^ge¥A{. +vIo؇'@@?=`ZXa_uO:&, MLN(v#%j. cѦ5TbI"Q˃0 ګmz0:CXD'UOcE?K/Rzv]ΚfO..wAĝ5\tA.|+篿OXܥm( "#ٻŎX ;ʌ6q{1FܕL׹+= ۝>!&ź1hY-f̊:aT .+!f!Ԧ#"^` fS?2켹JsrƯJi;P7[F;Zog FKqIϓ*'U>OY֮$Re EHF"jȊlt]4V)V4 hPɗC6 ?#v"{-a-*)i W<>8jT{ +"Z)Q \lQzdx'Y$8BcLQILb!YAQ*6%qw^ɔ#jO/$JMRqJg,i fZ@k/YMKE{96yLR2f]/3Hem d#!L"X YhYC)X"qѸޘٗ7u¡tr !˖QX9X(|r|"LX0 vG)IW]D ¿:־ق]"J*Rn%P)- <;XZ|߬K1`̆5<8JIZjUcCM3O&GR\VRy&)FפGL]7$Y%7 Y> ZYf;CY/u`mWO+90u}&IcZN]GT笑e`DzbnL|rs%PiC&x0Xo>D&2-ͥnQjǷq*=9/[Ikm^5l:z3-"ӗ'i_GpH/uK壕anhd?y{=Ȩ:OfpceLϼXӗ2>|Md;xz~5h=Mg79JǺUj{%YZ3,Ӛx8$sM_~^Avd)~KqkpNv~SD߹'z;+_{jɓgp3Mah:6Mŭ3 qQ-}l 6cau+e NTjj9,uk:xu-hh|f<{YsVW>5.}sfk-ϿF)Szn]Zf.n/oqw*c v,_㙳Gԟ]ˣoLsVw?Ԃ\n;Ր2 ˻! ˎAW*M7S:īyMX3ǺuERhGZ8r(fkodhj'_7sbΙ7`ּ0`_Wԧp ܔW}jm]B9HiY̊\S3,dsIƌqݷu`~%_]~<6I&9"gz/&_سy[=u>t>"BF]bS~~l2Z!\i{''F})_?C{3 &\8(`]dT6j94m9ij9@TǾ WޓA^"18:+Wm/ ii7gˣ)LbN-2iQQM]qpN޽9ON|*-#jU({UYbqv9xrUMX)Mb[d)[(KY9H@R($ G-"]#n_ !0DP 1DHN[@fЦxQWGQQH@I^n(Rs00fDV5ZHcZILA 5')f\vŊDTB5ߍ0R+Zʱb]3x7[Ks#JR*Nؠap;mL jObESΊNt~lEkCo@}`3HgFdMy[ 40c}EDZa-a2FB[#t ʟ-@/:ii|+\vܓvEG}^ MqE CB5vM J8e_.QZ~8RrǤ3Snf6U t,SG9˵%2%.^W䏧Wד/s})'NNGoʳ~7];vKHz,oå?Qml#!'Z.=x{Ӄgߞ)yzpӗ[Ӄgd(=Gi,҃e; cE Wbu-cwC^HauaKB!A}qD#PV ѡMȞ&zjMozB`<_L? 3v^-ܫ ٻLow+Ѓk]_QKifvfCG+KorRGrGtTn!UcPX+ 8zwU6uʂ4ñTM滫)#oceR9Jv]7f^u>PMx]|}3/8M#s7Ʉ8X vL|Y5cP30}?tc9_#|5gcIN7ZYN{łk[?=ya?Ý uIȿn/Sʒ'AVzaMVo&t%!!T6&~ȚV.c]!jmgZ3R`'9htJH||},ˬdI Ydr :]K-ZI|%t1 ~ԫk!4`8S.p'Z ;tmiM>_c~扄OƍD!ޝcQ3̡Hm`0;-2Ln$&|KB,jq^y0tO@Xp l':K6J4) $!U"m*MT>:QEHhlv ?{A@;]=eQkf TG{ڀ BHIѺ*BFc.20Xųy&M U PėE1!-Ύc:2~k|<T=ko#9rŘ/i,ioKw@KIxg{mf!%nɒ[푰G`X $IDi!"~~=O/k޳>}kEO.r@$É1m#mӣp6zj[ЅR0cDV^UpLjN{WQ5xݐ#@#g@];(訊A\~fmD#X|#n,SkACWDÛvEf8ĢiN#d.9{^Z:yJUЕF16X~Q',2'-Pj:L1~K(*,cߪ<o^"U<";%u8海CVƈcjV\x((89C?ʌ=r@)H&qWxY֚ƙRѯٴJ؈+"m In a]YdBXUG T%ژţvߍօN])E9EJGLHe%V#?jszҳ|Rj|Ouxj*>T-IH8UOIz@6}O: F\N),<ef{=10S 輢0b2E,~X"Eb66-0Oi @(rHSy`΄҅6Kߛ˕(312|5'GTϕ"w=3["))Fa>8lC£i85X,z)Ym&g757nЪLI2 dZǝs>63PFkٖ2inwjMdLϔks4 [&ӚEJ;SY4Y7Bqe!ϝ)#EDg$ѤN#e<3G3)Yk)3GH1#+]^;R N{Mc~F3_ao8v㾟kr_B d<πn5:!UQͪL0r 'jv<휅v8C1DN4 :ѤMҜ'3I1cAo7[B+][sx1+,䔉>Ȥ6^(9^<#Dy/8MJӺ*3Ii%aXdQhJnK d$ŏ ZoD  5dPw5nDJBr%oDω)#(;;IƤ< ĺ_E 1{i;ch|Q\H:s2F@EO.Ss^5L\Gw]fg1C0Ĩ2;Z6l9dpQP\IgYgq˽%cRz5jgmhX-7 8q" ~A aͨ  mP?.RUp~g0g+Tx] ڲƫ~m?#c}ܕO..SEH~6hMaO4F/>A0(9rC@$SwV^p{,ԈGbvǮŧevZl](2jҭ1P(m׶k q*"/ hvB\dmc|@PLYmuE-GkYtlB!2J @vB\BE qqe3&F .y-4afKq"4F~fAtǀebZN(lS}P8OSe2/rQH3$OF7[/L.bhkſq! 3t΋ݍuD󜦢U`/#=\VaS7?Zc=Z\7!G΢<8߃vخ!-3r%׏>vutTE"p?"*֟MgZ~,xzk>xw;ϗK' d!ΗkZpjaT45yKQ1}޸A81GnoӦ<^גD! ETۖ!:$QU1 H!z=]̔Ā*"Sqk6I( 㐝  ڼ|hIƵtzBG!i,2>}.===l{[I=J `'j{ 7aެz[= IjTM*%)1F9iK?ZoSKWjE[4R*IÅPB3e`DiI̱ۛҸ^"Rn(LJlYoŤpb*¨/7JY,`c5Fi,:%{]-E|]J VۺxRAq:RrnS&2 pB(X>?ƂA^3!uY3wRXt;,R:@b =9Q MG2 y,JMh^A}pkѩ=JX95LdkDPF֘/RfQ@xM\< 6ZVnI*SK><zj,E}MfeIt}KLˊGTPwWOH ]%Tm7YՂʓ.W-k<""=.7\v4,:N>uId:)5ϓRO}&-ǔ}<<w b5-buYT!Za&2u̴nbx]C-Kۢtڛ<`ekP~5,#m% $scRd}VIq֡DO Y,Ys&^C~׸X/Uþl:Y5vܕs!ܻycϱm^!ʨo%{| A4L8ڞۇRTACۈx( P+;K7GJ@L(vu8Gd HaX :P,hBsޞAT2ڀjmwFqka?^ѽ<ڭ gr"I;In{UǯGj]|*,po4@ȖMT-S){> Wx1t8He7Iܕ+q'풳{>.~PF-﷽˿,Zk=(NrR߷fq8 BDɯxD,**4B[+ز"IγH"šlYVx؎0ifTAq L4ąhT]bmU ϝT;qx;6cnE1JO._6 \Cz]6cc_oS.xZ&8u wc'"9?]g즄0u_$OI)pMt;$i ,:eI?0}7=nhWdoӕī;n$E^I=S4ǭfHN*sN4wR1l).G%h&){!R}ɑx@0 y c12]8H.Ĝa8h#4M1Wĺ"R1oҏaI\S!P 4Y\ rפR+.YHc8A~i~8_IA8FƙSs1NlMID5se~'{y@a<8}x~L?4l~Ӣi?Yq>D*=l?th.ҡH}hOMUۭ 99+]Y Q&3?_-2k=|E]>|>tŋ$4Ao9譣u`L C`L-n21ـV(ko  g:H+Q 1Pc^XNӌpp 6#y #术!/ɭR>jd"RI5?9+}HJf]^̚E(y*.Ȁuu-ȝxSzm$0}ͻwy_JthX'}S~V\ ?F');$hHxzA׉'9m*c#70 5۰tްD# Gd&hrƤd F-Goʯwo}#O>E/ݾւ{mje8,Zc%# cK g`htYeO.Tt?YLΑض1=1G$̹(f^7 Td] 'R*Ě@sҁn0HQWXc2ĔvE:L#u2m:eVC`LAkK#8/®`6]"`7˻3wIoGwmH_Q)NJƕޗ\^!);1K3$eW"junt7ܰ[$@|R=*qQi4 :FNGɨ RVP-$F2}P]*qQr* //DCQ1ʱ#1^(T3+zF!PWJK.Fx{'ooskPKb:6\Z5QYbmӯIA~\h_ң ^&h4y]mv˘b*:_'RE.}W&7q{7%MڢoF3ømMt=a[ p[dH}CShnm\ַl)}3hGv}3C=n}C%{7!n~*hۦo(Kܼx+}C= 7uM:\LZ.[D\̬1uc ѢjWQ3/ƒ^>~9 &5+q Q,v= 3qd:}ϣ=SSCKpnf!&u -Hco&34){S7[۠|IiZ͂vV4[AtS@>\6}W-`d x j *_ L0S05R-s|L'*_@9^ G'Mx T~^_ǾnTJ{DZ?.҅$=lނt_Ju}psI%yZA3\Kn fƓN) zL1mMsou?27 Apo6h"HٟF܉YWj׫f$"Ⱦ1m8s>J֢2V;TE' ;Dћwg!K(P}LjU']Xt~:A3g'\j_w'-Dܪ LH?{ۡ_ޖXAu.?yVN4h4BZLm#oWFPށ&6ڃԿ\Іҋ4uXՠJ_*694sֽ,AN{^2RN{Af/E;I;f2ALcqyoz5\i}mO NR‡5J3kf(ͬQsf5cZ*UӪBCr)4!+c%U+tp&(fQϲ+}QQx+7ij7 |r.'5+Rq!,lSxm'c+[О!,FxV;V h}x ,.`;۞Aat Nfn< S|f,muKlٯw: }]: e#Wn{f{JG!{AjBbIem4^܀(IP2)gCBj=?kQ;$Rzb1M%`xGnFa i(㝅[/_*2_%օP`DlJ Id@N jOrJ l$ޚύ $j*aι^^@Y~5lƨxiy~>cEae"aκ?qCZ.,>Nf牴 *~vH<<ĻխW)7q1)˿\STtU'yge湛:0uJn(yp7jzp+nVUWRP *6:ZQѱX%"iqVs—ʜjHPke.xʺƘ(*7E7AV .jW!|+$X Dt-BO +#HeӭOD@~BXJG74*s"Z 2@^t4N؂!i'q]Jٳfȏj$~Og/u"x_c6n(|K?JY*pT**-( "$SsP%AJϫz RHc˕B ~EKpcc,s#t/#uujSqc:YS͎#[LW_[ AKlp­F*V1Ajh:.0jd:g{qe?B!ޗ%Ӝ|O0JʪJma@g5^SZC(_i=C6R ݏ=a%x?ن$} !Hˤºu^ؚo7!M5?ܤSGcaS$Z.EҲM[_&z ΅5snjƗɤQ'b.N4%ƥ"jN[ża0DAw {WAzyɩ PWY;8Gjլq+ ZFhw@l'М!CMd 'Qic3 iDs8AN lǛc(Zd?IdIz;^&x pOT3z>XbUqOm_ex(Jޗ _փٛ"u#=5!^k;^WCɎo]a'?4W=_IFhhT_XKd)]'2Iri5e-g~duPRW(j W+WQWJ V*U.6zfZQf8P#1R )>:j8(cwV^13@`"~j uG4'Kړe֒Tv'kjZ㲮guf5C},Uckt:Ȣ"FJedDeB1kAY J?j _jňZf5ˤXBk6U|tiP5qRX^cw)8 `jňZf:҆I_5=+EE*mE^/RNid!Zp 1pB!碑S^3ga׋J+I3i"ҷY룵w{|ŇE?8!<`>ڡT58bxK1hrȕQMr)Kvxu~YnsjB}2zEcVQ)Vq2g'b a4!bS +hj-"5}L8Sd_]\eݛNY5gVspFwnʼn2@NK7>ɝ9`й' ՙ (@tǐBJ9rL#d{ya-A)"w6eyZ"SB UlAU!JUDU 5Q4 푧!6*'"{ } ā'p+EDCDEiD]WXiCJPdz5FYiB0HҜCE"JU p59 *B/DW~=J_ 3ڣN({)EAR͙QY,r^ۤTd`OرT4 )ѣ탢M֩]o\wsI,ң2hsgrkՏӵmcm:aJg=z"S~<cK{f6d]F=#dթGh`cL"%Aڼ?f:H;R+ W֕eK~FrEe :1otkI(E4(qW>YK1ܧ,d r#^xƭfI3bPT19Ÿ65 2Iɧ'rʟx7,,<ijAy&Gu`IZׁNI݁F[,m9VU2vA:]C2T7 5S@;WsT5nUl耪IO#o܍vܯ7SY)cj  ݘ(={{3N*R(ﴎ Zj+F'}YACE(&[_~Ùyݻ Nblv%}f)w̉fauQP % s43JB~WI+\ c?0e"OKX23a"?"F Pxe4@A|n +ƅt弈Q6v@`q[vk9F@;↮ BlT:r"&֝]a3߉:K N掛b.}lP]#\}a3|8y-XMYM5i А5z- 0{@t_--}p쿄mNlbIKF= Gdu=%gHk6׫r+!EҰVroѽ #뷟>"]^mݨ^]uj3C3{|xoAI$$./_|=(FcT7]sF INJjC3U0K4|pvTKhْzq?m\o.WWnfvLu"3.|%m+f g֧8hğXMؿbS{4KPm>CR ~n)2C?ҕy2d8>w6&yzI swG*c #䔧=Mc V7vSޕ/l|nP;xuޭM뵟{_.{puvz{^rhIGS6/7NF%3ר -s l$aJJ|T#pFha}ѻwr;4~~ɣZi ymzNf!~U5[k %xwydW,BKc4cKxT?yMSvF6y@dYypVZ|}UdGzW8٣C{8>F6Oك3Ffޣ:- =кaEM๴SE'QkJk+HrF5YV'rfR?L}@r.*'Go uSUZ84?'VaԌ^tw }W7TXX֠ I 4Rp+ (9r!:Y|(1@Z@F5jKGg%MB$~fR?Kڽl)nz̜Q*R?K^i hЛٹB%r[Ugɜ~= ,ұvט5iۤZ;T5Vk=pD(Hlb ( Z]{%X;斏YAt @c@բ^8} S&ֺ`mʰFGKDED:sEv3jj]C)&m5xE*Y(ԗ0gWΧkpM֝%,j9+M̋kjGԒ+TaijxțѴ@PGkOz^`$. ([{h#zIF=FmK$Ynwn{^l-`༊"ǙLrvg|EUWY4_5Evj DOWTU!TD1(X8wZ"?j~\r](wg?ߩHP >,s-- s͆{򽥙ԷIk/};wt]|՛zy+mvon:ǟ][iY+o4=Y.f/{G,:{nYa MhRll: e..\QW.mP! Y u@ٱm?`Z1z髱A6jM\ޯNb7&,<~C> 5sa 3y˹+Ŏc{l2K_+V+nJѠݴuɼS{9r<Jnu!̶9zn*=7ns֘᷍nZrw+8DK[. K-!k䢼 *ÊPB:ߛLd{^FSW_gO~a # /'GZPUX0vWi˚jYƠ~&4\@edBxv;#!jENf*H. ,"02ůw7 _y_~ǘL-b-"@lٌx쿄,Ay4;@/v ӈ'H_.orS֋>=BɄ{X'*qKUXFiʷJ J.'^ti8)63[23%Bj3A?Gf])Xw8y@.c^o;!,t'o2^i+H?{Wȍ/N"e8 6_0n2rf[%YldA2nV=*X|j:)vnZ2I'y1ҕS6GDr@ F$8E 0TʙNMۙd)#\HYE־}?2YMh(7yyipޜMʍj* D֍Mu A T*ª*z-xĘ*>i*^}_޲B6ls$Vŷ gID_EU@?#H-g([iFa]*S:fu)' R -@dUkTVkzkmjѱ~f5QF NF+CbVU&xi))h]ʘLQWXU5~lh ϝxV|ˬhtTx߳zpSB=\XRX0fR,T+pDK;z_JXXŪ[M ߉F34eê"nĮ24˘Æ8VP9XF5vލI4bF 9*oѨyZt\vC$R>|*W ney# HjH@;GGHgK/gY~zr7c6NZq KF]I:.}lt)xm.9/^e~./:3vj5 ?HK掖 EWs47g$nQu3yu2s.tҼk% scf$ǫv=x.g]XPf!ŊWsΗ su5Gxf|K{Y#ǞYuS_]ȺSg׬&(+|B\ɕ#<%풿g_7K\ppXܰZAڹ#YGRY}/'׼({ /wE\Ɗٗ4 _?~dj.vKҲ`1[]?<)`oTkмXf,SiHa]ϪcfUo|`/Gd~$Cѷ[Ow@2,H'Z z咐!!Q5س_YG @!}fGf!]ޝ\BE!ǻ5膖 @ aYOOa /G)(lԟ Q8g ? cB\fX=rIB"|J8f};ݶFMah5$OHܠJ 2_=JlM5쫭>m{M?0Ba,92{.0,Qc'Blk|AiEIfH:ޕzgYyh^W>4:}J7=/!quR{C6G75wAzx=l>2ÐNhzY5YvH*V8H8/zHd4–$Joծ> DL5f!E~X=[{ӑ^:m}$eDvVg1Enyw j:˗σ~x H';d'Bԍ#G[>yIsˆOmlФ5fd S؛Igy.Nϵp›s)КQhZdKYSrנ GmZj䵫m"V !b]JTID.9} XyA9! *X{?u2$(Թ]&T KH9.XʑcIژÜ]r1ϒm̗c7MBhjꝻI(sik ۘi4{u%EY'sRۘ1 .m̟fڍ4NHۮ픅$˒6h{ZU)L9g:$5sx #C[rI;&UY `u׵y!bG&H9 c}Trun* 3}-~_,.g\}!\,>kD=D_{ke!M3m-YV*I`Q1(K)z4b&NQ\y$hy5+{]M3Aֈ#uL#we.ּ-b'C7$#4y7ӵs`#ʹ*D9[ZYf F $ΓW־x(b`fD͗OfJ3*S֛,kD$C.eyhmy۞i(VhX&jUJ}妭;uPu`P9ͦgSþptX~">?4Jx4(}~Mv•Ӛ`9#ѡʺ 1Rѱ]СFvHoas&w6BNa#a Ov$brGP3% Ɂ_3jgW0st=Phޏ=9% kiog{zVg:Њ8-{҅WJaFs^%ڶftS&ȭi hON bRE6K%0&v$FSᕝu?rljv<ݳ(d}r#t'`g ^hxZa1+տ>nag?_zq-j?? Bitu:CX} !)GC8n`nk>keXDMчOOa7czDfO.g u|CtD `?}g?g?8s*rp{t:r;M%enԠVOG̺ۧ}>xH]\/Pu[>‘s-vХkԧƘk W~*2D -2WBpJFPJ nES e5HCMus(װLW_~,eׂmJ Us'F 󂝐.>N@i|ϯ O95~i.~xpQ!"CBԱeUPR5z~ԑ dErG1zb/-Up|I:.Љ阡:r:RvN uLrzgЌM\m`JJzwY}qAL#fr"ԎR;%CNzqhVc[ոt{yv-8^.|RG`mJ#ѽ8j&:9Lc֓ǝCmd@'O;%~Ը7 ~7 vN )^v[F빐v>]%G@};-h?~~__oO}ƺZ2X{Q A^)iajp}9 TV+>?T8ʚg!*CPn{1F-yt1 80åv_*_7Ao$ vU@ L -OܜYv z3ˮ[]βˋ],37FAmZ37Rf PεUh =Ll) vF5pyb 4?D( w$ț?ʻ[E5I`T*%QIzB!9qфjaODI-F[o[qJzKweH~[`R< [X vYִߠ$۩b*IeuR`|` ~! {+GP`kH)JiFT)*S0ɟTf dZ~*UEZ'KGsR5IK.Zl(sup9Q1-x5AԼNv8ppo(%\y1D䮫 {S.,ϋbWw]w/@kr^׫k7("F ,UZ-{תsMam#Ωu4YJ FtIpmTu,L<=݂=@Wl)zt5hDщpH b7`NwVNVڠFZ#~:M 3<i]'6T^{צsYӚ a6U݈Dq]N*jEq; mұ`}|ol Y!򡶃ہn-:n퀨 À[@#"=aQ6h.!ѿ$GkU5;Ѭ_*+ciz aF 2 /d\չjn!&JYCL^*y Sၰ}u;嗙ĀS;SEWRgERw6^qq}1"4Pvu;z}3(W1gS. *q G LjG3>mf֒ke ₸Ru. yDlF(yq*U(u2Iɲe>DP``\$8(,!iƘ@$~V4 њ z,q#(" 41<: 1O-m4(00(NQdpmb=PN z/+~J!?l[R%T` l{K9( iB^%ʴpb9PG̴P`:醍db-"OJRyxRMhU@zp)7@@{ga5CEyN{OQfBXPFc'4Ir6/y~m}~W!.p.Hi(G `&y-WBP$w24p2Li Bܶ5%B9tXX ӇIm}㰰-'8c{j&j<>e|9|~\uEc|LzIo1mtŖj{#9B!C{ӂ" %|gR|߇oD͠ڶm{עsʨ%>yޝLD˓sO&9!D>Ιz݇iݻ6ӂyVs@{1O?!~ڸs@`I$W(\~ Upއ}MqCP]lfr}65*ߋ-Nid҃o@Aj Gt؁jp 'B~c =·ͅVDn&/#ʪA ^J+ fd5AS/,(}>^o=M?>3=<f#Pw㇛ԈVFE{? m!D/8տ[挲yDQ==>Q2wqӼ{iοx4' }~VԮ*{ sAi:*⊣4@@gE1+i8楰?="Dzh`Q/Y?Ve'?^߿/^ @-$\%kcJkbWspe*im?,2R"ZI[ŒWoi) 6ϙ(^|Y6}⛒7,nys1(xm%ms9ǩA$z@$>d@'N䈟RH($IƜa$va$m/YD5UyŻhhyTl ,Z~J!?l=) -ArMU.AB[ЋT΍auQrO^ҍ7%˯DYr r~Ӥ\kvҽ̖]\ubUU'У< ׷q2k,餆ݑXލCP^]2$_Q{%0t.x`TmyK ֶKZ_z\BqKT!]}Sݗ=Mኰ\2z/~w A 8ʨxAUUG79*O:tK"K|,ϔ]V|o+e66zcj`y۷z+΂dpE11hǭ`Ih7l5Η >60KG ?$7y{b[ C\|=MNa늦[@RbCS_G]\,Oa Gή?^?:O?5=BU}C؅ڈs4B+"XvcU6dS%9`Be,Gv ͖ hn@k)6>koBLss秪Z2λmwN?'OǭX";d>, go eURA;IPlϟ$2z";l0Q܀""9F%9@=xaȴ}2R N<djur3+o J6r:ڣӋ\e\-wlK]T ]Za4}T{`=.5,؜KR6G$s8 J*(l ,t˿2^yFΥP?S`њVsZRƈZc++}Yz]T/"r8ɿo~Wxa"ڹ:i-m0P/grjн͠|G:ӏ9w %\ sSG8҂v@S7ԃD:JB5(݁4#~tG՞i V-&FA6H'k|5*dw ZL^q^/n[_Fiaw7Ha Qd#1( H6D|eFk_jΒ!0wfiBBe}_C 0`;PH"=kF`]9@#e_`Tآ zޫC rέI=*sZ'$٧:w]Ng1 sz=7l._zXS69W@RYɪ*P4@)A 9+xNo,3x f1&A+^!dJkRa 8W%aQu!~;;86Ri# zl3e;pѼSOۋpiM>6e4+3cAZڨg.p#p򅐍:,cݵ/՚љi%\$F@m.L7 #L>ح/w6H1+زFƸ1JQO6+z|5Y˭YO\ty(N\{>Vh썵ˏ?w-բmscN<2xG\<⻡ w ZZ>\6J8|(FW"%aŒ  /kߧ?7,L{ljopͥ算4A9Yۑpm!S=zE5j<(c4n;n}kքj.$䅋h+uWsUcAAv)Bؒiքj.$䅋2Q-'XLaoΒsrzQ4Y즟M^W_eBQ$b%Ea 90ĝsL["B(KL)VeNe YMզ\@`9' *\Ini7o;EQfs7:r})?&%}5\u \"10slhnH֤ t7+6 ZnQ,X/OHJs Pre~Bs 6^E;UlAZ4B' N>&Jz"EIPhon6qvk^8a;FY+e7_Xlr 5p#^`sO/r˶EN`åitQ zrElsݳCn9gF(*K\φ u7AC =f-ծ1JoIҶԵ ^1Idɯi22R1D_\T˹cMt렲VԚ֮^v]\3(5}@P'a]M^)77z@k$u:,o!0%&o뒵{@]ajٞ me:~M\M<"FJZk֌'e{ke+54$mO:*SeFA&(Pw\|Bt2mq`ڟg8^@[{暐5u^au(;"#"i?r+J Buo_)^jA*pv!gˏ YA4s' {yHhF&%UQUdz\%Ր+ >+k}]ZEmXZ,vzڮJZD8 Ek/d`ܚnm/7nw ?}Mʗa0sҞ1 q{[dGh+)7C;тrNx'ɱw;t-'7 ns*7 3=Asp!"p_dbXZ|[X1@ňRM9>rJxX=|.Xsݦ]d2;62#Y'-e{0x&@Yڨܚ ٛtanɈa5F)DZqOOvL莻\\qxSB^_`p/[&ʋєX.G/r1hj ޱ砅,B}8خMqA0ɧ^m9TC NK$0hue]qzP:K%4Xh% ̊!qjL<,JO;`guAރftP}6.poTXrS騅  W*Nt+툹6FʊY lKS2v7LpȒF=g`d|=\jynb'96|c'96|R =Uk)I75VqYNS  %gT(3G-ET04YR2Ma+XDaę>v^xyrNi6*t67a틪K]KX ϖ+)yX̢. U =b*c7!6HG6G2Yka`}ڿk-l,I*a#$ 9FDР jLO*-٨`c"U<%1^ģMtt:!AqLjz'ƶFVK6ʮnw`MOn;?䄨In\xgj\<}玚m5y>:!~{{q큫+!~8=Agz8@i9&sir5]Ӈ3:89Oho;gj*i-#" 7}߈lM טI};oظV Z$DʄH˜t0AֈSHCp52FsGvIyXi݀\zo'֑F| 8N++H= @։L:d8HчVd/~pl̕Qρ u˝!#-8Sɬ ɶT`'ˌVtr9 Prl?3J&sQ8jt0mQg} C?"V:iCv60^* NmQgzthW^V*ЎK̔^&<'>Z^lWKhLࠆ&y]o0h 'r*3` :5UWS Hi.{0&+!`Ʈ3L|T #u+0W'H 1ٸswKl A2_2_/t%\T"IDA6\{|h+IeG:ZD J FC^MD^n,7]RҖd#/G{#,@@VFVNV(A)$}+[iUѾ \B($X` 1 [s|y|c龕848pݞGPk 7)K6/wq]iao}^po7~=& k 2E8Eq<.< 9|-~`gJ~~Lz0şrg6@jKʯNU\pKx5(2O{8. ^ "N-*16d PD8`F0B ޱڨ ofGk\)# >?<|Ig}n/B7۽H9Qk'+#ˤrc .U]׏wZ&]'L"NrVJdB}HGY)JP՗vJJ]ܪ!3_NJ/mv˟ۣM K`=`+ܷnrߺ}&mvrg `A0jHD",YˤkT`P:PKw/*T˷f|ȐWq#%6fo:ںm"BL$~cJBf d Oa^sV=nJk{ʓ 1 1R 񀁊*T, RAfUxni |eR'{dԗm v|29;ZF@>Z"LgTdmGԪxq?LHpmH1(aTEwt154AdUS;XkQԾtH0Ǣ}On='^zFױf1N{s힦{ɫGWa-[G!cx{#F՘rxh0mt̋m&(fJdYC ϙ%xIWIUޙ,pRUVxesb{ UBYvb°YǿJs>h-}V`&Bg \&qEVg<M|`IOi"urT~ ֜tN{*=YqJAqŚYaAf݄uo oZ^ $wζ I7g[k@Kk>zb}a-mɮ l5X b("Wjj[ d` ~1L-̜ӞWc\NN֞+?aA1 e*N5Vj0;l]mMsAgm1֠ wvvKX`Irsv]vyiv2K Ub(90̜ÞxWttf36&wO&۹gzJ07AqNHP'Q5n~ɛW6ڂS,++{82WR5vzFwnH\o CXB)x}u4UD!):J4 T`Cq2X𠋄dN`}TB| ylw|V;6}8f|:ytZ@*濾1v Vt#jlbWIgRO)oR{t,Y"Zy~֭ #T: h^e(X9Ƌ!ݖl4)^˗V.JWU]y01"h#8FTUWtX#RZ輎Wvs1<ЯL)t 0p΃bu9UB칈5 ;"t+JłD6ĕ# h,Q;(pT"xglPok|O+}dD|_ 1;gJ9$3$F#K[iaƏd1HlfjڙZ%c-'UxѰ}vvp`jO(/Q\7M Y[kMRī x5*2*%X>K[ԲQT)8unTűM,βR)(\$#Ϙ74*C@kɥרs^0ߎ Uօ.qcs0#Wn[|~tE_??~ޜ)wc2sO?\/g2xJ\7mBnoW}LwI7m )͋7@y/YJ;T~6_ovVj;vy5E"[eo8q8=̦|6jћ{ӿz2(ZʍK^-߽=P1).T;B5MXsuZgB\ s a+M8PzXsS݅j9=%/w\ɱhtWഺs',0٠ugbZkKMD&4 c[@80>nm"bȵX S(. TqĖrNSl@1cƱR P1wqaIv13ii-fIWP>~2Z21ߐw< a1Ox;ɯ~07=?/VX|x ȩq%ǘ(|d4b{|Ȩ%SÀ`$~DrמQ(ޱrD=.רD'lKO[W9s{mn~^ϻu{-W 6Z-`NHw*Waf AJ`\u?.cIx":zn:9 ` E_J$HUEɖWl)dԤ*{)/a$yNzy6JHKKYZOE(cFPHcd;8f4oOUaC8Sm#M}?WBuaPI=^_}"JsZ.H#>vGyrNk"Go߫77ʷMV#јSf5Wd}ig(B@%St|2?ovgzt|lK=M)ODs5屝.ҧkgs7JcèPR醭˜iVr}$9k^|촜Huv h\{b~ϸJ>[)8G/@}J <)'0WqKe?Sqr}B6r` &>>7Mܗ5>6:VxtEGj};X,L-!2624Bt"ZkpPJ1 ˆ(f*KYHȞy9cDTܡ(zS@1QI iEJqPIlS+d?1hQ- Xǜ"D(3wڌ!6FaA*4ab˱1wEc1( E E@4XɈSpi |`цFCDXl"ʲB(RW^!Z:"%)UJmH&YΊ`TdJheb8-ۺ©TRHe5ZR;ӎB d&edil56.`))4+VL$Hƛ9;K[?])Bf ٣cv)2%#Mx6֓YȾ#pZ8ԅG- $ pXrRbN@a:gz K"[yLٛ",^IJPV2b Gժ JdoV۔mdi1vVvEa^nYk0 >b"[`Fl^G$ay.tz)M=֙-.٥-^%HKlH$W0pN2E8pmƹֱ<??wәx@V46s m{tݔ6-kYؙ=0‚9׫ 6.xj,-Z>PigoZ K}gz[nIs-+E)*G_V,VxjzNəxqs2DA!lO|<|5j.軃ywI}ժx瓳|6qj]i||=WWg:[_^a=>iVSK8_a$PZȬ&t!LM_Nv+$-B[](ok1N([|9*ԸIS5 Nϣ?PIiLp4 lZ8PECҙp…4nl p8=8.Q(0vdVc!(r:%nrg4ܲgs 9_P}Jx:*ps?:ay<^5)7h9zŊu}SFN,쭸Iw۝pjx[]ЕiJ"?6C0>h {ю5{wr6z! ۮe|vVʱs׋?[A9rR ɬL/G"Dj^K0P\H1cJ6gDJeaO͘OyfGLR_͟_i!.q=Гj@Of ,3VM2TVYU2(-e: MV!ۚԉbvRYwexqoUlT*#1aL%: ^j9z<`d  U_8r}Ek >ުaVj!gH&&&m\.Px%%XEnIa \W E};t@ȾT2~*o+=0Zi`Sn[ @P 5RZYb$cF,ʮ uΈ]HʱRMaQ%E5/#-Eb~HF6d-4*(b,̙^l̒5r\yl$AP2{"ס֎m1JQ.Jwޙ2}Bd$ ڳa``ء CXy 1V@ }ن*R4x홟8pT "@&L8(uT;!gБ=ɘQkݢ ⳿޺#@ gdpsa.eXh ga2"r|2 p-l+8B,VXM.Z9b#*U]]&(DY~KP~[ïJ0!ԮD\H$ aJѕjuk dGgBXGjm)+rc~;RͦWΖ],9u_Ӓm@Ѥ$g,)_nkՠeO8%fl#p'ȘU=pX&ȵ6 \I,dU ř̥ptN &{a$y,5Z/W~d [OU3_vl+ jy9q,uM=sSl&yY&跿A|<탣J0U2iU]& yÃD'U//?˫&O*l+v͆k2F7`vŒ / BكI˸Zy/^3 2[\3uB"p Zf."͎5,68ιudݥ.|s)MiR"iGϸK9X>C ԴFw d!R ͧ?"F fT$R&wW>=݃U}D{_?}<z#u" y{*S3:xpp8x$2FgzWvRwLSv-ЉNZՅb㴺)jݭxBc̺rs!\q=)MVb)#.οZ ؼ6zt"X?V+/$7NxwXW?~c+`|꨼H}}?[舅Xl$^+NuaA:Pb;<]+ʖZArk,:ZkpFΜk8cihG!IӴ}j^4:%^2w/א z'C~}{wRsNjIͽ;ͽށ|(."|BR$UQ]kR`Ѧ@+;cCcIGSmo%]:|7WOY0KG~f<3#fQ<gצ|2m riu=7ˋO9pPcۻN Dmd%4a( 3PEyt: Gӡ*t쇒,[.^v|ns%Ğ62kȘdqE b"V `o:h:&>Kb`U`H8 wWrXC1Spy"b^i􂔑 ڧ(&vjt(j?%,2"gpsS@HY#sn`KL׫iCOVb'yGn?]]o+_UEX@b \l?m!P$'p}ch=}- -y&OaB@w'o*o.w5 7;j>7sSh~@'*c~)8E.NIqݙN..[ /`1noWSq왻aƟ㰲rII׈Y 3!`ylcg\rg(lW-gc-22O^s<\sŭØDgE-;PI#k3 jǘe`# cιU^GԞ.hYkOGI01,q2t{g([DžjNm}R?O9#t6Knl45w 8ܽ/W'7/Y>渞]߮2_eC'nH`ӰtG&~{{ڛ+?;/{Wj/>^_>ǹ߹WyWv02]r^o>Wʦ^S?W]+߿S_yvW_o}>w湙ʦDuẗ́i8qU>~~ Gko6w| Wx]}xCD~GN4,7?5K*?ҟ~5dG٭Smc;4)W59fcm;{gys@B ^J:!2Pc[o`X|7. AF{z ic YjhG-ZV~q pav"[4vuk4b 家۟)˚|r^jtkTf+iޯ':؅}G[Fg{dĖsMA}ewͷ,=)‘&1k7I `,|,ԑyYy7p}Ǣb7[2Oo*}W髾J_Uj7В-Fv~-W%ٜMBXrΐCSJj oS4ԇ>x.We0E{`u$*BL@=< ṞJ֫LR2p9|Xnt_=:GG)0Y7I?IU^X֡03b^ia3d+zdd7kl*1udO@--q kUVӚحUpM=P^#e]]&ma1u: !Ŏ7N?"_8>rVpk9 ˴\6^܊=z،( U7Xl37a'ʢȦ5 i5c kz&Zd 5lAySF|,Y`;~򤠠_œ@x{'E9<ߥ A^{ S294ɸƻӔOח~in(gN3bx 1F@؝f@uYǔد0 F/cYG;fhhk?F4#`I1K)ΐW/ ɢ7TK߀cq'j' ~]Yiܱmhz θf'e\ta̚5wPgkΣ5Dyv7FaZO#ukZI)ɼCi #NDㅺ(u,0Cn"̲(OÙ[}úp4O\Xx{s-7eK~|[cWonqgXj_^H/??ezO>^Ou$/ẖN c/ԑ~Gx]ŗQU.__+/jGY!'OW!Rr{׷Sdzͷcu4z|qҠ-I}P>|8yd.nҹiHlsqC%}xKE82F4B k2(R4M!`YЭӿrh|vn ?'VasS2 S &&b*1Dnu[.Q9ĜYΒ, ^Ro~a{?.Ēqw)ݱIVԒr=/  ~{`$roDíѠ3]]x[ZB78sx3 jh* k];t{x~ϛ뫘3}ţn_=T)`qΚjw \UL9C͆S(oj?2rO嫋݇| -aSkU_ 4RT->T5&[ D)"d7Ο]Xw/|H̾@) GeV *VQ1-x2j "{b1s""W q@EW^Z4UL!5j)R 69pF(&HUdtyqàZ/)®1R9>ybTWcۚ.@bf9^&n{8vYT?s"h(d%"gRn:BDiC EB!Yqp dY6JQ]J7'cW$C)ŘDAe߬8s0WECc#;6`:eN.oJ+%Xv6;O^7 AJG{M5 VW\4%yȹd TԑWi+ߔYa6t_|ICÆuicbڔƕꓺP{ [ jzH}xߕj'ۆFq0ap6Ȯ]S2 ^9:bϳB-~P?%crƈEP`]rY @ŖIJGSHJ/JDph cRhP0j7aE JdKIб8u0.$eVoMB[踪ABVAC’9Q50Q9RC 01 NRʧ nzC鰣6(JTUVzl_9spD18 Iװzk1t}J^Џh:*\4<1TKrpR V.xJ.K kL§Dž :B[ 3wV Ke');IDx>nԳD(ua(a11cfO͊R'"2aƕp.?硤f I)CQYvkP#CvjѤ9Wf "̃O=S dK4r;В墚 Tsͻ oy2窺5c&hz&&ZceiL,9[o3zKa2W4f%A6WQ{:'E&Duؚ ian^'0"&u*Ӆ V*:m9pFaf*& ԥX !40JJH%*Vi1saU9hST_NF[޶J9줣x7  T$1!̤N$XVMcB69+nG4⸊WU(!IΑU Fk֔^neQ2*= :1_]8Du _zDP&zabvP.b~_/fT uVnU?(ȶ (G_`!E)\m9QP2ΐje *ozf1B/AʡSIm>``XeBlXpʼnpwXĜz7twK$d7`׏I,n1C/[GCKhPn=I"JE1(y8FAq%xc7OCw=߷Ȝ =9Fu|j,Y@_"smÞM%g7ɍ-=dӺPuz&{ņ\CneVϳW0灛QT()5џh8/%t*xS5k85IBf\YRmkS#E<ZժRYzԗ@׺gfq0?.Kp[%sy(:q6zU=;0E+8uUܜP;ؤӴKވ\ssl j I!IoQi$j$֪̭%8+pfP-$ZuJ~J,{FR]*ϡy ^ɇ Ge(  6 bTE짪ŐEO5ʮT J8y'5shk ^i.iE^9] rnB%~yq%h(l\ WвojJl6}Ugl~eп[CC՛D ruTI$9*#\D3Z\M"nSAT"rTa$>9U(PmVnͩg7#UiZW!Q釟g*p~rL@H,[4T.R/IfUY&~MT*(1r%KHgՋI~Uarw&iF7m7|yqmTB_>ky?eA.e`7! Y]y!u6ep >0/Q40swmzY,J} 0A< ` ;>atHͮ#J>Gb.]U]]^霁 2®;7LLnVm\]9CU`r{GIR'H%/bRn KΈqw2/VN'b4":^K2<6V1E%Q]a(Y(xj%ˊuȊ*9Wrs%9u.@s]rIHRoWy«{aQrtU6)U, 9PK#:mZ}\ǂNm`B MA`XϪA0ަ?h |D &Pu#:mtv4mmSć]AG|h}vDqBrh7wX~=SjD PՐf⨡_$jkh8R԰^65i5cCMD Piyj':bD8wt"8\l\=NePclj:@ ,j}`!pk$ |uF(1g j,5A`Xk5&:^m :m#( |V\{DlYSUh(kJ*vj%jhMI+5RR8 mv7t63@\A(Qt4G_| @k=JAW8a)]|%FՉXGf4bza/cuv.p/qD mVPӄ'jXK=Nf#65KjIsu.AXӑ6K|փ$rR $<m/[XtʯR2R *LjrQ:W2bȠ~^8I.!k  0sr$3EGil](aC)0deA#[ڶ#4,y xN1Iaq%Z2@[:#fTT@( k9QEx+QcA+($oavY[94D@W$KA@h@j-h6rx2e+K]̌OSWdrIن2AL3Yd@TJ嘴A)KQ" 2US38 j$6GvCdED.Q)kՠ b 44Qf6Y\J2t,3\ vC`21`:d&r4.yXⶄU#u`Lö@FIm`7`x9`f%X15XM"@B6& XLVr60p }yR4Anh7X挣L(<X:d%A;7Ȅ4S hv+5'`b$28[KgLc01@8! 0M$x`#%PyrSF/ \|1J{Q- wsXn&ʿ%jhGڙšn^?qPˌC^z8qvQOzA #QGǡF5t3G:ȀCt^PC%p uC8Ԥ?>ݬb|޵nyZwiyWۂrKns/VrҽixFu㙻o3Mb?sX%ӑCj@PJ~%cϺOLtUz-os{);38{%j>~CG/w !ipZ57].(Jkvxg]txNZm d[r7&!ZLl8bxz|tR$0GzT]݉ˇ1an!,Xs\HG\'L" Ԁx:|r'fŵ9Xi—1sog7qu~U߭B߻/'8,;[b^Az80wR{\`ī^ }`B0HX P<dt͕ŠW#t,^AxPg UnOχOhR(?O9kY2U.+og7w_V.fଜۗ_*['xrݝ]^ͫ Uwyᬥa_~I+\_]畋hf"I@imJDPʂAĥgZYNᐥc|yp *q\eu Ac8rk 68Yi\c> SM1Sɽ=|) Bz2aw7l:f=zSx<4ZS(&n+*,{:uJ4v-5)a~]&ӹT:p@? @xu>jv (-n]HgOJtkZn>zROQa..~y{f0(ggS9/'te="BQbN&M#dA:।]߯WON֟/E'ZIn$0<$f):Qa'>y3(9P.js] 9X}&h_,?-l:4"8SO|}Ee }:ᴉ|x5,̃Jt.s3*ۑ'|P[xjO a^JS3L(3u94Qgc6bNcBgH x~Iݬ99UU KFu6JJ%e&퍌)ͥzOvYGi5~]D^ڧ-ԔaIڃ.7w4]E\[vbف/̓zXܯmxĿ\盟KMλ'Ͽ0_NXϋsz.L y. hKnmK`a%!`LL)n@Lg-:]bB)F+v2MrU} o8`2,P~RM%mCY9X,d=Us"h2pwJƕU8k!pԴMUw%nRZ8`tuz:YǀS9x>̓,Y*)*#`{ICn&w һS[iVBv0-r3%hWevZP]?r€Rң-U9~ L:Y?7&ɏVLe~R&"UsD%Ƭ7wW`X}I1;ؤV Y{aF0rb*\F]gpp: yl;}mexf7룄uڣ}ÝU=JQZ{ GѴR Gj7;{]Rԣ/pnȊ cs9p#[YoOz4K*R=殈dRVϬrY}fSr4jE@;% +MbflE_3q%BLA`LY B3Q@P!3ʘמo`'f++UW9U}Sfh:p'O%P_9'GAQ0Q1fa s 9UMyBu=Qaf(Amf@4 #n5ɴt)a(c`Y öʜ!ِ1ZI+lL!Ҝ(!J̹8)I))J2RHd$b[Mu!єw4b:Ss\r kIxゃ O>@Zgm)f-1'0qKu$ kc9L#%(IσAKZb,'HkII "Fؿ &̹x:8 j؃2#EԒΜIgKIl),!MAqAr` qNXZ+؛u06$" cwmmgL%fA*=x-%*'vYv{†([^F$(3MHbҍwf:_ KD im#Tlv<ϐcBaX [GUp,ISdDQ#C$EǢIZ Yg5 \0k@$'+phɠ葲 oN|T a#oKIQ,`<2d" R!\y:0zYw)AJ5%j˩)謓og:R .:CH@:Vޯ.Ҕt19YӚ$6 :'Tݥf9;:k I$3Z@{ 4bCbĒvL1BV ? @d2T" G1m+Ѓ\T&DS#Hx]+(N'=3 @R@u81䡆yQ) !1PUAY@; p RzAzH$"`CRB{.EZxf$:[+(:eZ Os9~1ֲ*k؋5)T@WEWteݥPt -tT GEfoCj.е^hHt'50FmkC8FcAИ:CLGu VE C)IBՈhZ@PQ_zH]ٮ8/ @ ,l\"^;ᡨP!5*RlTIPG`Uo*FqӖ& WEH\J\ 2AvZ7I?/hBVA2C(@RHug@l3jAm\jձp%@=$BaTʠuzD^qHcKm5m9Rˢ4b`Ͳ%h"qG#y" +UuRK QjHEɐ\g iJ H 1J$$) '#RԂ#;b>gё(ɗ &\] Bƙ,JEXmܠtל)J({KT;J*]M(>Qs%EqߊUj ea 5JHt>+T-!vN$RsڃZt l L0jF%4yN<1c+5h4&Krx!5 I$y5ܣXDHi+ R/B#&类Ԛd_ [IJS2$'坠54;?NC0ZxWf9\BzYkZ\J%3Zv3DClZijV><$Q 4B I:h'j L vR.Wx=6c~$3?y 1yJVr~l f# mZgCYe=1r.3F) (He8 "hDauJxDYI>o.F0&T0ʒMÀ  Mqr(k'F淂Y,翽/{'wR-h7pCG8:t)iHL /)ĤD[g/(C8[BI)06lO)V(&~b]\FQ1Q #L nu?Wbx!o蛅 z9s$C:lq~SLuV l\vಘ&*dW '!Xg 8dz?8Cƾo1?Wcɇ Z\ӡjlDHfX>XH)1Bk2nn舜S$g3Erƚ,:, tAňb 02,Q,‹=BļxP 8c4YˋޫWb[Ѭc(3c X. u?{D'nhN'{:L/%m\Aݹձ7C9h[rhK1Mf)ޣBҎf.m KpvEEyUewHbYS5.R8a5GU%dqH7QdDEXRj"]i6D>IvK/5͗X FeDKѽc8bCJd!r@CpWP3,*▲VJT.]*RsoE+K)?8]{7z{F@I=DcɈFN4q9--ѣY<)-eC.疚j_EZ}|bGnҨ<9yEZ/~F1.ÿ +цOpV9{ʭ sM:@Mޔєˁ}S)`(S>elI^1BIɹ(S0_F16-5(Q1hJ4M&8)gT#ASCYA84E_&z4cJ̀)f8ZlN⌍Ds2e$>$i\yz?KklfM"A82T%mWB=+ARJ Ahro 8.:v2J@r$Mb$2lV1ڎc7Hy6v4k$CwEZ;S {-$̎fpTTEjh ڗF3nZky#yqǼZYߪ4_6nV^qZϳuY6DR[ۚ_t5]Wߺm\AˣYzG_zUZ/Nw_УS)9!-dRn~ ?qY_&p\E[kbWɛ 4=^*}q#n8 nzm,9lgo~zucHh9eлgb>*!q|X_f3.}>y9?WY?ü77Guě꫟g⟳YvIXOtyQu2&B%7!/\]zGo/]vZlv+g,15vdU/~ܰXsİ'rz^86Č.Zɍu?5ksИwׂ ~[?W6W8 .w4<&\2185[ L<\'s$ZD9Ʋ~mlk2_ٟ/` yH[:vz |>Wff ѩ^-ya.߹dk qp)?7z_rDEB̴1,,7V_5c1wc21+\n!U!tz^NjF\^L94,&O)f՟ Wӷŋ?S*0T^}.7zOG}% X}}ӪǯŕHw"wy;qԷ<\I>Vu̍~ڟDERm꩜N#̓sdb>N1*IQ Uz*fyZf.c ~翟j)p.jı-t_uWUw]pvn*Վo}\a.&_֗~!c,+?x4ߴR`kĔyzSosy%ifl]ѸTjT5tX Qg?u8Pc\ra^1B@FOܰ?gjT;X{:G"v/Ȇ.fk=޳ZAv.TՈMи\CxhfL^$wq7ѯ].cgߌo2l&li=cSǝ=w%hEvbN&?;ۇ̨ߜc[ξ]LmNƞ蝬O̓lhj![~'QO+6!%ӯ4|N2C\ʅgaQP~QxYӲ/@A8)8t~@)tt)vzu i Jsaaa/ڎrz-*"d0 &qO7iC}ݸ m<&% Oi2QASaDuPp_ (΄osP8jpvᆂNrp,쳦B~ B}:{OYQ*Oj'$H3KT5tw;2eU:Rʸ_QrKs> 5RқoIH!$tYv7qsTZJ1:x==T\ߎ'kub:LU]v2vcلɮ=ju~!3~w#{{{yzQ >VA'z3ؔRkkf)H^xnRʲ>G7\Dg#8^~M.(!tucM2NSh- X2ߎdSm/L6v5<CymIblRs 29bb[EFY7ÌS利e b 0Ơ:'>;ۛv9r$P;SgtWvN uU򤶋Y떃O } md}$oIA|Vˎ{hb$ /P)%%FU ؐzD8 hnLg:;y=\d*L[@G53MZF3ˬLJ5mDQ=CBp"B_F;8۳}V6zc.gW Vg5!Vʜty!>/p1dAR.:_ CW46QӽbxaoprWϙp ipKB 6մh jD=qKր,"9!Jϙ/|_ٷpiFwfĢ=ȝD12?5Yf3aQfbz+]wg櫐DSlG!ԘCc#Bc(+!>-`WĊ>i…"lꑍ~ş`JBTyn/s@@,-k뾢}Д膌gJV(X:ƸE}Ɂ@9ȆhgA"|g׊?w5wD۷U t@wåЄ3V,_.Tզ/dGQU,.Vi!!4?-N7`2$D]M(QDV:d<n\T3.5ת5< To YF#.Dцybqӑ`bꚁ c[0,cf$u4"$qoMu6m Nrdjγ`cChȇR1޸Yg59Br@h (m sCZUsy?䎼lZ>Z A? Zѷ# (2#|܍^JT<}^ruZ򝂫.0$u<խ]Dі|o%U-^Ff7C5bE4e&0U_c{HFZDΐc'HZv~2&E։voRNv\]*^˴7'JtPvК+%dϒ"N<^lhBN,<`xQ'}+lUh{w:95)PS'\ʊQɲݏ)8֭|a `B X%!KJAT YJ/^d)Ji5hTp yc9م1c=:IS`4IFZ'6P¦[sedczI[KZC7#@"' " $6*HMMyks07hZhך%#MdVzFwZp(1P6lA Q'QB&T FRVvْ^TmWL(SV=k^62 . WRכzHz;Ba"5וͽۊ1tZ ц~=ć؝[gȸl- ! YJ 0\F Ndž+ezAB:~؇+ e3h?"ʐ׎5As~ ɿ1~򦊈|ox Bjs6na&OW/@M]=>لE5}7==;kO~̖O2yM.t3 s'{KPu-m8);yӽa$揨+nZ:wqY{Ү#vrXm|%斳o6%Caz|y_|^ljlZ~g=t%sfØ:5n:)PTcnX;WƐMtdpXsp?N#pr):& :!Ni׊xAcM#;JQfBBTK>o(V%\ J_maߦwK+"k7q)N:Б#Q)i "S1:nETWtd3`:fSZ5c9ri;ѕ_`@3;exnvMa!rϵ!׍u56jnQ5>dF&NF_X N،~w^=x= lrSicסA2sjԠ= K5TUvj8g 2܌˪ia} ͟.V4g\ٿhnJ^=|W~þK6k-M&wW!}EʤO2|[,m̷]\W7M起)YDti(𮬌Y&"O;)r^*]V쟗7ҡ/ R,_Ծ ߖ;gw_fWq˷W38՜7k|UM5) `pR1V{vZHΒcLRFsO)n^Z)ڝ3Z_4>].\)B_ _|<>\?=U@җZy͊:ϟw.R߮;7@5L4]GF߾ [57)o\֘ܬe|!1ڈ#ci&L={\9rmfxZ1R=g9>6PdmϢ҈hZ5 =վ"Pz?ooU>CJ\[k&OZψAGN -9v$([!dp䱞ȍz:B[^í5ܺZQ4*SݍױUbU}z^ڲLSxY*j|>3/Z[(уK ,ȡQ `I‘‘srߍ%B}䢖 cN=ۭdDiYboizU]}%H('D7 \:U"oф<[7_J|(C9iiwo_~Ӓ&oxΙF[ͻ䮗rlngO>br]ɩinro7C "/]!ſo2\<`iȾ/f |%d;gVW G4!yBZ;jV7!GƆk9\Ssb^^_TE&`, ziK5F,SnW;of-Iof *x4!::LGnFэd\r70oO}PۢEov݅"dNg*r?` ?Y$OBpiuX Aݡ/?ȍ#bFVok|V!վ'@an-v{$ 2VIcDx>H{md)9"ZGT<-}+}Pƾ"0\LS:>*m;J!\#|m kdeuVp B^U\Rp-,lE(^i0dWܜv/[ǎ\ q{nsJ <崞OօS5mA0rd]8* ^ߞ{r|F>y/]=;6~7ןċ+.>,)҂ܫ{|)`:/+M^uZl k}X\57ƷaUt|*_  J!s4-fRL1 :R=(5o:Knj5g?g^]_UxUfJGuURj+g~nӂ|^sHhlF`ZƱb7 L$>R'j'37]n(.A7ъDʷB':QWh'j^3(I&S2&zTͻU$_ C$mr%n' xCp,LBIs _i1QX \ɏBdm|ЇeD'zBbi+ccyL2I ̓U^ >4Ö`؞:j+PY2dt c3IVMFX+;[TK&J;[ZkbrEK*񠿧`@*uZ̓`*C%{0灨Ia&ʷ4}_BYE&d4z m3bʊf;+9y!x.HhR {A:*Wv=0(!Cx%|zVʨ o&jl&dBhrIS$ioS2:!e%1j9$i6LX f,͹QSԜ14&DL@d/ٖgD23mң @H+G_{:&zbW (a5bwC[r*z*0 m5{z>6Q1;㉹aGJ/V\I!5'c/:Z>6h g{JsnTMŭTY<4Ƕ rc;1"Υ҂? P]CӁO;E5q:PBe^\]p-({E(^%գ',}@X ,њkXqF~ ent<+bMswBl`UIS_8 ʮe2t)gP^ЪTT 87 #Q^ۤ:fGĔdΙ2(a?"ڞRT&Njoc_rA Y+zGN2LsF&faҙ&iOeaĘzgX_7T2DKIRȳ)yf y&HƂ!K)Pm@K)4r* 9J(LSAhDk}9lL2v8aӒ&Do"۬bHJ.)(Ѥ zQ Z b@:9^T >D)C4:%F v2١Icؑuj˅0[%Zm(hnVk趞=0~⶝aPm8m'`TgAi>S{n d1L^b 24Fx$7&)}V͞wU,ɪzl[\5oܮ~_3}e-weaݡrKL;瘵}=} ?9Y gi%kIA>9BKbVA锾#ǻͷaɼ[yeKwB&fܣXȥ:?#}⿽s|.ỳ_ۏ]k<&+? n]+ew~wܐ[x@?=j,bܳPA+2gɚcWB C;IH>*P!)oP-6O,C"xT9 `Bк^&T%Sag`v,a_Ö&J%o r4-vS{ faakUb%v*JNdO>CNj]͘A)C85L/,Rsd!h١Q+斟{4$ԔDדc!dۓDQMAAɦZ~56UXDC5&;,SGG0Tt`$\|rN^Uz4"}чr"U&[7xFqIQu'l/M<|:PϿ :!!eoҝ`r6 蠥ݽKb"ڪШdz#UeE >YW`UI[ V"9 : 0uU馒|]vh(V‚bm6@TLeZq1%fEj̻o~W+^>޾Ww?Z7uK/3nyNVXTOF`HDZPlBKC(l@T΀m);˺rNJcIMDXUAJ)-ڙZ*J*)"p KFueA`Q'7t kl"dEZj]_M.1߄}nRA%J*-*+ע]?ըa@70ޅ:H˃ɞcAT<x723nz-?C)zc"GnkŮpQe:'"V?<X/)uF7;X{ZX'D,{ܶC4Ԅ yc(oc q2rإt?26#hJ>R#%wjGxv%xm>M:i 7x*DfⓃ?v'iG:/sI]ǿ_ #  _׎DRȊtm}ݕ);W*)yJexRj'Ttҹu9ߤ$圾 `qPd%#'u̍,hAeo tNEz 1a-ΐܘ YkW{0;ϩd3 0~k-=ԖU{i'=|{r-cy(qftZa,8xIwo4* {Sx2[aklV%RK0l1A ;׈*zkf*$]MMdୠ4Ul9űR%Ol7U2 ǒgd/ÃtSҕCWE>Vu]5GlLS{uCT|s]Y/D``DϟDc"L¶0P%d o)9FCC9R(JnJ#)5Muj"&n ],ͯ^-nW|Z֏7>5R}Qۍ+ځ7P;(.jki  o\Spb㊉;TKF4X meW}T ː66`C  ~qÐVe?Đ%;dKWH,eϩb/JZ{~|hP&+,!ʬ8fC繄Zകj[j匛,ߕՒ`r5 zMDghiP*,N#D-^=i4j13D X2I@\Ows];O]gB, CB}=~؈{K or}k&KshKa@AB젷(*q3;Ժ[~?_Lʎ܃RyM>$/=~:f`xPC DgâjHz<ӣoERS$8Tgw^/j] _-i@u(K@ZU>}n2Q;3m{N~aj}[[y{SVޟ>46enoU/?4ۏ։P3 P%Sr/M¦Yĺ싣^8~>Vll13TShK]vN>6iϹ/:#-4r/ҍp88D&B.VWtx ++Ix'&SHB-*gl״mjf췍&jސ T JD-a| GXVhBӮ:@J{A2A^:Pmdu%U]@qDsJ_8b;rN@2Uf(n= hyX# MP\?H9I W&ZB&Mݛw52n] A%[(iH5C!):i:-,ΏCnZ3AJ R6Rk-FҗS R6b Bš҉m/f#fU׷7}Z8<d-1Y>C1m-tdѴ)or@]> VtkV_[}wB&jI/HI p< +Wnn \ImGiS`JL, U]˜^qaR UBbH+'uj~ґa>/,EŮU GƴV! \Ij9I|!1'|~NRщw\4rv,8VH8LMM ❒b:+=:oFh@TꌓA$]Fhm2m Lu2rhLp$QA’P O MJ 5s1 o+M"`6vQڂjYAMIWyt7/rEmp]C" \$*%&4;* Nsue1qta❤KVkƀM (ӎLd7}*dD+#bz`A.~CoC7) JUGv\6DZ*8RrMhXT%t $cz,ѻFw\ O`ҳ%g]6/oxß?{Hr_!q Np;_!w5+f)JKRw6S=%J=O;`Wⰻ^S@2j5-6r fZZ9bGl+ F VqBs15T 9n {j3R(?݃z{ lL C,دcX\hZr \>Ԓkg%ȵ!p 9opaN5x L ub!2M#fymAQĿc_ M'c; ց|q]\3__ Br'|H>|54A6~=;&QP?| 7.þ]ܮ?{:b~C @X"rx-&xUa?|[֡7&vy Nnf71hqKTR.lŸ&47m+0UM^bAi/hA) /"a3zmS_%uVv˘ \10aYnc8:;=mݜMʦD h<&H]Ҳg'G-F0 i(Ǩ;n>PьQkG.ꊄ-.ضHSS۶IkkK,Oz+J }kNR_$7'<;$(1a6(j2*"{FC${wп̎&XF9,1qqTNTj({[iZj |ӊ,<֘_TzҘ}kC; d`VnIKP5qn}AȚFC nbw N`)q3xxǴ ge  oMxy`)дtv ,,M~˰]l;^]v Oa]"MYKY OiIvD!H8Ha{Tn)}P[n [Y OQ4|tqnMq:M"EMK[2DsĴX[SY9irSVؓLUT *0. 6s:''οh %wL|Ke⤀Л)/C} 5 ȥȥXq)bƥb:v.M+MK.iJ+$\pL]c 1$&=-TPk)u%AhƙPW; P1dGT[Cab3|Bi0@M(J $KPB-q)1)YӇRFҸZS1ZWͥ<Ѻuk KyuU@-cLG.=>.Uq4yXPuK^hӴ^|ZK]g[VL5{`qNOIR$MH"Jsz$I#*l Zrqr2ť #i*5l*S{vJ"riHz4RQFԢ =H;'O|_]g&`!:DRdB||A{s{1fF! t`Y Gh<.DWLgV8;`Hs݀c'.ju[l/NOE~dy~>)"LWsKlXCZ$Qv]XCo!Řc\ ag^inD{vV3l5&jZRc;D/ Nt0vm.RN=I:C(iKvyC+>Fqܽ,l4"R?]OiPĨ쭹4T Rj}#1m4Ѹpj qDQJ F^*_*!L#B[8Ր( e*C=TXD# ՃI}]tW0<IT*Dy*?fĴY"$m Ңo:EOKPJtETMdGD#jF@shU񫑂"ܕ9(Jv*G Tòm :dIuCQe7~ QcJVTI0FH^NN9O?BI6 s jEȏ+޽ J:oqP2XL\eJp.5VJ^lFl)U@:ite̊D$Ipj=P{i@$q=+EwhF Kr8׍}YmqlU>tk`FI,?T47ɣt[Q2[/7Y^t-c rd5)ZqM@ủWiљӛ7sNҤTMu>{M%X%ScNr E-ig.ۓ}X&r9C#Th㲡+Xq5QEaẌ́+aW9/Zaw֫DpdI(fD48ffR6fM=I ,}FR%y]qXah*Pn},1ezh 0#CVt[ƧknJrmu16י7cJ>f5XSOdXHz0*u N,xNˣee^ N6+4{#& =5,oȲҺo0T%}oϞ=@Zow^Ŋ&ĬT!|?Rz4ƶp0ܫ@Q5Ji4w Ӧ*CA/9Y#HhPչʘ> ! ඇ/)$G\)UovPhSA H&jp6 IdXh~R]YR; EaҦ$-B[I*c:uA%Ío QI_U{ #vbUX~.̱}zo7nj~VOwXK|\⣿ͮ`_suUCF䧿}Ywfl1[]zn x>lD axlʱcL.};<>(86LK qVܰ5ɟlm' x+:yW M r7pOWspn,r5d]Eu\] Hc¹8! і˹*,!'1 S9R'@r 1-)I܊+.o# `KPgy1>]_pj"1٥ U'WWECH6F_9Vds1Ls3XDZwzu))Bʦ5H?Eq174F8&07ſE}pg ӈ3"c2G <IW : e$FI JJ\  S BCıE1m7>`L[ . n rfg3M0V9"cE\S żs;Ki8 (p+c[}##V<BRPeV{SXWzlr^cwqG,%0r_[>.v "B҉~O/)rHRY`X"wvfꪧjلVE2"3&Qt΋W9M"J{8јb+a [m9Ftj$*pe50[%$+,W/V0`DM 5Ntm\-k6-ۚTN墓>"<%!ʳJ( . 'ZncjFj5(Haqi&C ΁cё2Fщ`t1 z0CpM6l:EŤC֦1ˠm)XqpܸE>ȨZ5t2)HA)DZÙ0Ac,Y.(0h"̺A|" -I,͖ٛ1ȹ87Jv>(zTp-*ƹ(ͩjdǯC##Xs>GlWi¼XQr#Vb}^yh m-{KA|`iMw5IoIlIf\%I` d N8"V?m3PV*'g>N,pմ(iZv̑ӍC ElNK}8;]3`"%[[4BݚmA!XHnN&[>T5VY&1[^&Y.^0pU HΝMw(bS$!NaX6)+J^N༚*p Vp^hT.l`D1S3zI(dUZ[JS G=rxh HP|z}+")! %aDâ-1,'iYe!Cj٦fdmPuJʜS՘RH0i"\*C{˝ɮcX٠ JKJ35褮`X q2*VH-o-)UyMD8;SոJ a+J>\ل vFj/ jE4ԥ809,N ?]Rz.ko4]ٽ`Vu47z5t0(< .5za%{n_L # ]+5K2@zݛ/ X5ٰ N }E)YY\e!FԶY|nlXY LkY1@)5 i*:xzN?0a'#Jp&(S+\V]C ]u^5UdDHƒDŽXF 1UӔrFOI-?&g4 p;*qp-%:+t ASԲ㉳EFE%{SC8]m0wJ "NM6F_p L #B)")/$Ζ V[RTk*3* jrf1YI-1)+'(4I(T+syo;F貋N+SL ^بIJ4%ל 9vgTS*ke3ꅁٯP_8WR5CʈJH#} p- 0_IՕƼ`tRؒ< *R%p.Et7]cU2НI!7cRp+\IQYG;^uXe- eL>v\5|EXc FT|v2Jits`m*qBjkV8|!PY203J)q q\uS[ pgNoB/UO/F}:VI#\ôP(]*<^}w= ' vxo &Pn _0 9f|S̾k[þ1J "`5mB)ý{:E퟾~GXc%Sb15y:![jWج&Ⱦ_썢י).nh+ Q*V]R1W(g`.y ^Ii@J}S֤ ɶwSDGs EVVPB"gc"9Exܟ+s@t])LRkV.xUKA`+჌QGE~}ɈgM䳵APugtQN&ڎX9'"rT0 ?a?+@DW-hmBuYV`D""$lk[,¸Pӄ8 kqVQJR>*W"y\-1F nU( THLVN=*VMN? wJJ@9+0ȠFժZI 튂t٢u<*jTB,Εej@x$7PgW(nd L0A[)·RerH[/%q5N!>l9T2NFEkUʎ$$n/2^0BH9D^WcamƊAYi lN.\ E;()ʀ` 'Wͭ~Fڂ7+YlZ]K툣 ζDtӲ:p3H iPvߞ jWWfkh_廟^+ j&_Ǔ;'U6ޠf383hnSTDXV;nuڼF y \˝ATKDX)63D\M7/jk>Z̑yv كJ=b# Ţ/K0+d7yfs0ϓ>`/G'_3]OGp|+{%aw$^u_+W>CRFbl)A}+nb,B,,dvr0:< m ³Ӯw_}կ\FeܹnIz\$w9Ŧ=ۗ$k޾Vb7W󞸵y ,V~қa\G2NNc/pOAK1RPcfUBB`|18C<-==@Qmy@J km~!?;s$3Ɓ`؄'B|#A sApHg`~6[HX|9D0Awzyl:vOn҇ՓD榉Z7B͓$j 5C!z)9zietiD{P4Q#M^P~u[>D~o#N|y08D+8A#A G`'xbzF2- iݶ}Hc5];mZCMN36}iF~;\`|ym}{\Iɳ߭\Y;t˳oÛWѯy-ُ)qN _2C08H1_O[e'+]wi,v[rp~|#)c}FըVv8QCk*rW9ˆ*sp&Z-WR$i6K9O,׳v9ng󲸘7/˧pŇrAQ~ UaI[eV_f!}o[M-WXc$Gmp WJR<#;1=@b!F2tϘo]{ DH2e qXb?L/v˪l%4ݧ7NKq;/o,?.s>ٓ$>=݃y/Ųӛ=|*͈^/~8/x)o+zqr z7N!x 'y78hpֻmH:Y>JK pmmm7_cHE7JXG_ݾN*{MB$QV$w\dPIb1 XVO?cʚ㸑_aevv(܇"䈍XxEN/wSn`l:(q1v2HdI KX J`R;[u~vܞ,^vO`~3# +@`+@` Ԯ $"N>Om u 7/"7+݄8R6 Xk%[~uwWvwJHY5Db bsa) ,hᙅR~%X;,P89``RE0uN 3fHΡ1e:P ruN_gb|]Y U^Y;kktb07V!>i,EmVR1co)2V.- DiVɂ9J0S$`wTI#4tL%f+jE Dx ./C{1c){ߌ1mlݏ(ћ1f0$5@[+ZYlIx Ggq$01dN{yUzhi/'.,X2;HSN{9c c0߲A ٛw^1}C {QåZü l07vɯVWn1gH>^kF.~ȻAYaKqwΣFs,k8ē*wG!*js|remYgNj+պ~v;d҄Ù]٪֭O7JJqke7n)8,¸<[_oF:˪%ղ[p7޺UJ#Ec۲nD3aBVA)}Fu΋i9mk+$Z#c$)r)EјʲnyBN/ 4FÌ˕m;z8b-(!-Y .CɎA$.ܒӫxsd>h^+.vG9 r8}PwʼvVl.?`f(X?g;or;i̕=?XkV.A "ڸ5zHjmc\tC p}}= Jyh(eoW-,|_;)LK=SvBxx\zZ wfTn&}fR~K|L]J >IpyƜ!.O 7fK ]]zHxϧM;qvH)T) 3V:BI+(4K%O4YmBL^坨E 6<3~Xyw~A1&j7v̄~G5rozkqá_#2a՝x%FK1H9VCwbxHƇƘJliLy{܏":CIrSj"&!E/΁RL;r}ӕRBwL݋kC$'g+/_ڙoՅzjI 1,aΝwor H@E^`+swjo\{_^uN-o/C4rV '@VH?rc9Q 7-"%zя1&@12VJ QR\#\|`㖌Nh"I-<U,Vi~ fnvcm{T3g_o–AgOwQju\ͶObn+rZ@`\i8UPlDI0 Pyvnq?1,'G9"iH026 $!Ԣ)ךYC}k{$&) _;WʧNH"k3>Ę*XMڼLjGNAFM~=`{ITK:xH ľe ģ=+gGСy4^=I>)qi6(j)y Bx(*#j,Ap8ӌuO#VN^_vg=\#ZDž#g=RV$~X%gQ9md0Lƣ/UD8a`oՠRb!dbԄwR掄W6g1#UHCL[ͬ,\ d7^J/c g1 }'g$ ]wsK={~z !]c0 4,4oJ59㲵CЊJ6D?~e ub%Bc&Bwk򺂰GWqH;bK-{Z 8RS;¦8F9SuRsFVMپ!!`pn~{toP*2Ў'um|/'a0a>9$|("ףoҲmk#IuNNYXn>6:P|:i/śvxӮoMZ` " ctN2TNB{0؛fSӣUz[NQ郞k|@6*@K_ֵ߹&x"9./FqcuDT8]=>!UH[cdG)F˜4׆<1!RF,x9U΃=!?̡QqDA$'CNl~^깦B,~6^񃤴'?P9j1{wdqi\,N8{'FyRDc 9Y2 JDp J58=,C{ҙq=YG]!,o]fRJ@x9MWPH+PH}#0GT֟,% NUcm 76`W4X1I@.rS@SS5B0mqAgS 7.1Qo]6h^h1F r-eG '0뭻2vzM]䏧]gAUL(mxֳKΚoփ'uf A.U{1gr&ߓ ^sΐrJ4L]rƀ:|>p.FsHx.5z f3{`{ SC.o314SSfw89tF{ N2 DoN՝f 6y~u}63F{t@ð\&dc5" izPp֮Pn%Pn{w|wg=m? uwAnkqlkE q87P@Ps'd|%Fa)!$Gvh@(A@ -B4\R-i$F$9I$֛,XT*ꊐ[ÔbC5 $l_F3@aART"Etbk2d9T+[/:֐gSg׌F"n̛%;gp +he\i/0^j*my( 9 'r 9HO^CxJPHU+öR>ӠU a##M (|`<)%J젅̔:;5xD)L^Z ͗~7DGR&.: E̖H:L@prB+jq; BZӛJ+ &`ZUiMꚒ׶k V /9bA'4S5^ zU۴QV"M8M*xWdV*僲iL v¦?֥EhY\^Q调1,Y,?_9VDN29r߆<5qu(4RNk$BHb6!'?J\%MPGEf{fOr,'֥NR |2{Գ;XoDl505LgL%C-L2Ң ځZD;I/v[m5t )3K!!2]*h"ڈzoJ@'VtgbA͆*(!G=ϕz$Rk%X)H!Z( B@(HH<\E.`&@BkaH3ƭDo+1VA`<3͓(@`h^.0=(Th*=x`pF+[ev0DjBX%GLG{0J@2%6xHCΰGz({:8荶?Tvދ?*L2{ bWwVO/tC0G_BBI_"y`}4-Iu}t70 [>:{g! Z1Q.DYsv+  D8 ^.H-DVwv0 `N aHAׅ#%Pf ˉi ˂%$ 6X2DS˶DBI奇Z'Fag<i$R#xjXJǁvxH ?{W׍dJ/;,~@/`2l VbK$'mIjuKn}sbY<5KԤ% 3RGqdqXrΊEǬt`2YK!g &8ݢ^{c,gs"ze=Z-4:{ODOЂD$BHrIУPǖ!$ӎX{.8pHu$ג@9DDMV xy;>@ \:OvlqRI! uB{]J E' NRi>0.hLROEPo`L^#dE-l%DmRKJP YZ F1J pv' L'iu8i*=2vVxR$GF)@dBdF'3sBZb;waҲIK0)-c5uscT$EDd"\-f&ĒsݽWG 9Vx4Lhe:nDmfaa0VkopB =@~F5,&o9chIsABɀHOcb@Ka2;n'p#(iS,e-kic ;AsVRt =N4'* z|.٤r`42ddfԱ)!<3lRqx",r ƒ͉ɡA7qb :_:C\qZwdyvZNs8Z8NkjE7.:{gY^Y^ =^Eh޸H'h N_>I8(6 m`G󾿢҄}b}0MZ*.ԥX01F$.:%\d!|u gC$!57/{CwOic9\M:)wzc󫿟OW߽5a__fίon8/ ϊg>ڛ/o>}s-s 51`էZ<16ެO_/YvK-7tNukO23tAY .QIwJ2m@ qլqE&MC͟Zc۝,qWӧ=$Ѯj!5z-a#PfbdQ`uBVT )`c”F{Bl"D+ÕULHq!=uD!b´5ls˱l.s? } 839UFqJyZ+k  Lh3mD$au>c$b0l)ZMu~wrmkG{:Dp3{Vo@wGRa5O uAsg㒫 M=_7P1VQ,|sMN6i䧭d-=UVTzjhP7$7_2|[_h\xc~%Rygޱdȥq(ۡm÷.VRB / E_$Zm8৆߼\)tV]|Ç?msUx }DbWs/OLڗW?Cf+!zr79FPQq ysm{?0mTk=H'3,o"g/iz=D؇izxbz^=|R/3|畇Z24ɛ~ۻޓ~TAhTtfJx1˨FcmQA`7Snh픭)v" a9pa'QITaէm{ Z4QG=Ege}:*3{7$Cf4B^ny-13@/>X!45s9qzmgk.( x8,NtxEJ^Q2ƴFP/ɑWCtBr*#2nv|5ǦUnBYbÐuq:thdzLAX'v4[\x?Ѯ8,'/ CXL\J*J?we::9=H |0s)|{CWo8XU # | 80SpZ;!5ON@p8[^1 ?eɆRO[b6ʝL?dلd &!*wdsˇ{2x{7k~QzGYgQz r_M* $- N, * K9$dEi}.ΠWee*g|Yq1|b66\~y~auz|g|*dހlR~,Qb4|cd.ڐ JQU;)K$D * *$_ }Gd$)RC]9liuAniu(%ku!)D1]Y-I 7p^ե[8RTxꢅS fi8㭷-:(TXi@XihN )lkIR8ah9!*nI\1yIiNĬ͜j*#,h50 kT#gjttfV;ZxhhTbS!F(4Z#{`a|}!CMgģtSa j,|R1MVIu_,ઌ\M}%BLj%Bn^g9d^߳EB=jDGE N%a Etz f~VYz'=QbXu,jbêJŰcmp1.amH QBz:,@bX8Vm9}IuYaEx29)e4hYOHlrݧ:!:K*(o _M3*8J)}eTP^vdLyİ:nhTbQXq \Rr:IYg3}Vz8¾/򇾹rws;٣GD ;TMeҧ?ݓc3Zܵm*wUICՑ$TZF? ᔻ%!Z]oMnhnڻ#X[#$ZRt ֶҿ*]_~Z]웭6ͷm> NG1=0Ru=mBȁe+s%p@SSpǐrD % (2%ҭFMJl}3:*y]=͙L":\q:.P)b]##LEk5UwL/W`%2z*9 P2'׮;K\% ˜r#;9mw{ZfHɽnw,SR_e4^iBX{HrNԖ_,)ؘ,Kf{~ڈN;$a aL6P j7 8,MBd$%cD/I9j/pVȀ(C՞+F 樗HጏĹTI8v$ I)+8 Gbh &be1yN QK-@% 0(OaOy6-3_y@.fUʤAD#JnPR]% 朌*;OEw&\2>c3U(eZCM?+$K:8MQhF)P59sώ^vvM4^L0K: )M\I,I3k_%|n=׽+t=VB/A[!>+Rt򥺲l&[֩3q7`%$JHG^|Vͭ~Ɲ?dMdՌL>E|EǵÎĄp,S~)?˔)tN..1 Jy `$;ompAt0M)! 4AlK![򯷫wחFxifѮu|m8:m޹[/<~yap+%暲qղ& i%wFId Ye~s Y%p !V/dlCp< o7<93deRaM{czHʼb=*}–pRaB͟7=!"V3J5Z9TXM\LZmOw @ΥG_OU#"1k`o XbOK+YnfU6J9U,f'ti$U͡k>t%iKlA=f*٘Nu 8J="= {xfT]/AFs>K~ AcǨ0=6kkGw}=4AH%];<ܴ6 nJF},TܗObخ G/n|3MKJܕxb Fb/C֡ŚZPuiB:]ՆW@%u_%iY϶NZEtZwҲQLCK]\hYJxG= \EhYI{\t2'KE}8E1۸ UC-뺛h2MfBOٰ"Ħi>>crD@ܤe_h>If +gp(75^]ŸzS 4-=s+=-+-zfS$=R18G}鴬|-OR%ѲɴUތhYJRCYT]ʧ%NyK:+QeK tE˧9l%EgQڬŚH9 Av>D=}̾[5O4Npf|3s:'Eow&].i&-+Y#ZV/ܢe%޻h0H{W{H:7Y[HG2D`3UF?e%p Eв&\`NeiLd kuz>}dw2vhAڮCeC8@|}+]Mg'k:YqcfdgC($Մ6҉ܖ_E%)v_p芆k%;n/p$utg>/&#qa*tJ0&'YA) e(bV2y9웻1g+G5\1xv)eݷ%h{ZOI4Z 6e=<ޓ*։@ { ewZ;e c)uߕ>m_ &`?yYi&_2R5Pp5z:iB=}*v }O^3TdӘ:*k![o 6`g# iHZjM0mgڙLgXGiH϶FC:DGiH4$)4|Lb%<;@AECzVl+?H {xHΡ!% rsMyg% i=^pGvGMgB%+ 麛h2 e!mUyfH }1,NF%5έeyfJp!'(J4[kNfhHSe 4-0s>4@OCJIg!TDnr4$NCʧ4$ZhHӗt*^oB4$,2KU[ E#)!VZb7?c5&@2 (J;\3A>aOEƀk`8Gkm ](ggDsZ@6İYl"2 ivœQf_R)dL_WfKF@ʭdF2c4 Pэo@Et q3fL9a@ FF$¬eŏly*.|u qw5*,b L{;j HFuf`WD?Cw7JX7 gF$;3/.(uIBM;N/)uIRM8i#Cki.˃=* Q96*5s߁ ś;OKQdGs>Whֳ*İ5,랗sg28݆LGqt+Du0ʓp @Hl=-nOK&v9uIH) lW=2bզq}o.d%靭fvB $Oݒ \> :4MO15ug]]AmchPs, A6!= [؆4ӐU. i%Wp]K ѐpd?ƐFѣ8 )iX]ΞnӐ4& øY钆xQ0#_Wl\ }GKhuhH}4EѱKCd@P7"&}5ˬ!MhWLI4Y"M,Ԭ|XlP \%:2 oSf3bf@= )-zf$=S\>nL!SmyB-C# iLC:?]XP- ע q-d[עɡV;%(L 4:PxE't1N:}bX= )kH|L^4 Zȏ&,W5wA@@$K\$А&VHҐf -z/3!PnJs)nU7%^Or|((jPQPh][H Dj ~ . m_tz73d((Mś2PPF-Uϧ/ZƢzpV =ċk5U V{jWLz9r$ԴcmY[.+I%y} Fߥɣ8/qvp5"E$Սw|qQPIrM8#<`(41ofӳBCgEV:hQ!Y0L};Q m){n({_8l(X0>"fk]NZ),`-zdH&6ߕg _&h'2Me|) XU:ks5LqPQʢj?m+gZK`ΒX|Y57 q+[wq"}+ѢCaIrvkҳmӐb `H^BrYH'I z"q,tlk_Lh f1#ERI,D)t&pG FWG_Q0[Gc͊B(l_tBd ^@aL B}1î`: dD['ssE &ha"'M)7-_YP!MzRZ0LHz8; I"y|($ݬBʧ($ZˇB7&OgWz3B!i \bO;*Ʉؙ*;$[MKmi,Z4E7Rv>x(-8 *khU5=Q$l|Y㪲T$}dU#xO&4V3wnU9K h)\5j h@ҙh@=dNNя  .yC%utFiVwMv1ͧ%f6| lCS1wigNVmڤjj!̑ 9a35Oqlh7 pbF\΁L nJLlJRGwm_Ĉ`V<3;Y A4ݳ:]&VÏқSVViDޗjo-ЀR Ѐ քi@>5`\94Y[Ӏzv PF17F3ݹ$H+Ӏ1Duwv!",h!K):Beoу] 5*lORmA;3gx3O~2Б[uI9?!~!~|h/Pǧ1hˇNo>n߾ӷ엿>0#0seΟ%P&eN<5,]C޷>ڨPաM2C+V=uvbʨ,<1Eα$1pTֳmob MМs7YKgfŹ\fV&1!o{ 3 ak?]\ɭ}y,Rb.!qt_}\~=Kh)~KmS_>tO~ixG͟0W?]/_e_/΀ǏMջ8 Z/?T'V!3Teeo*BDnLJ.g/y\:qIv[|N'h)`K0+zirb6ٹYPs3i}dwef%o[}S3%BQ%7`hrz*̜Xiۘ|@,J!IbԲᓦysz]>փЩTH={J(o_ Bg 3{ f,S( ⱴ[ ˧;Nhhtq]r.L'ejYrB'.IY`|&L{cEjFi6Uma%;P[ffG5])2sྔ; : WV-TY *+/D9:1J~u'w=CI,)9G@nʺtpJft+q(netG.fq3>.4 S4`ZO_4߆4oK0Uw]]{ŵs4u!ԓF;T)z|eȍQ`IM83rf$5CSI2iK\n[6RGrٛ:=C$֍w|Y!Yma;=|&cP)rXa jH( |Q|yGY.tܪ;W s{PgUC;wB 1bwF ĕ{xGdWf=ꐷw99/&s`ď*s2Me|)%.jzjn-D`NkB%]5gUh=K`֒piCKBx;-Is;+rƸƣi=zczc빩ε7Uus,YFmU϶ZhͽJJUE fKw$ؓQ~Wl+?*xpy[og6C*8!% rs Av;_Ґ}kK)~:YiHEh$+ inx!yh-oNOj($mJgU`X~5qH`UdX\jH~DFX4aNBћ}|@~ @ }"uF+.*a$l P]1-l jl6X ޏ(Tϧ/,%/p|PxnL"4=(L6ڡIyv<.(3IBM;ݴ/ł)3iIq p_EϒѝJۗ@bv\z6&IҩuIrM8#OMoy-Jf{bcYN%y0/豤DaWc 8*H @3ڕ9.ZDeP s ]9 s㣁!Z giakaN]Ѯ 9Rͪ5'!ͰQ؜0=}24F 0=q yCAĨ`D:D?o.ӓN0=*vk:LO9!ӳ}:Nj1'BQ#_5) )GIJ EfA)<1) MZ0tklRNƒiDCJ(Ը0p$>9 )~]$g?+ѹ?I4~1!OFLj!\ƀ60rO,7 Rwsv6U7(ϸCJ,JZC7}h5hJ$P܉()5h@ $TÄ%1+l8gxVо#)(HL'P\ѝ-|4Hx]¡rA<]P甐R䄫kQ$"=]*) TS'Q yna%p~ӥ&MsÄ.e"d::mTZ:Sqt W'F<]C-w(`~"[Ε@7}𯋟 plڭo[[Sb7ig"n^}qY4S,Ϗt[ڷҵon+ݽS]ʪ"pXVA6be*[F2Z;__6:_7ww6z/h1C7 ~AXk6_s 曋M_n/nošU{)&eۤ1)!cWk**IK@.$gϢiJIE!, re!ݶ8KZWcɡ-r% ɧy(X;۞`"')XDP8_ H3HaUB$ EavJox0 _!SRv VPx<`řGg:*cF jڽkscc*$5tm',ZJLBff% :XvU2/nhaLNLKYB2 8'&]*II Ë,z#Ua1:pr|UVqqz82.j#bbS]Ƙacۧx#Ƣ~ZNO\DI'۸5rf]$*LrtOHx$#_q 3Dڂȣ{J#PƑ0'FQ*=G.vzRt(e,`s+*ӿga=m~[~ۖA3f_|}Qg21Q[T8x00*zhhFeQ%K2fKM0VQ1N`=70*&GEBt H"Ab jl]֥ĪHDpX򉾹C{/@-Hxui7Ar8t^:#49]zf< [L`lʰu&!Ц"\=*i ^\0%N z1dsuvꘗl.}<*3;'hA]'*Y v[!fF]t u V@u]!Y<5{a߻f/ 'ً%2MIjw:u80fyxuh"--c[}ZFH aThF F5\m#4*eemS-tQb U cIx=lꪪ<XY89-ʢ-ڂ-nίjX*3Ӑv=NC%tէhHOӐ\/̗EѤ3p$c4iGCŽ pZ(B-e:P!-}:Nj1Z(FQ<BNÔB6"?MىZ$÷줓l Nl1l=`c ׸Hk ů7D<$Ǚį Lj.<İ0}>-|2RJaBp*[)ۦr~9gM!FEYoaK(xhũm)OςOdg6l|oVtBχz2hb'/og)w1,BAB;z^S!$g '$ d IAW΋H޺qT vu"x٥tv꠶.sJ|O{8?Uمt307 V̌ wkYuOנ|e 3xOuXYʻyd5u!a%Bѹ[Ŀ.~*)k^-dG˓n`i;.܄H:sm!^bn>B q|]ko6\/H(EBR@d"I& HH& 5*i+F'+QbN: ٥Sl]lsj2ۢnMI_/FIK3qXgŮD :Wl:|E8{ӶOGKomlN@R9^"t(gθFm˫uL%Auʣ ?5uoTC 2s {ـQ1"-0*z<=(A exP%J6[-U.\v&<ᶱZg_)}s\'8N!RH9o-HL'9Ca(yc8 jWL` U N| GplCpWdzJdK$I-:mUQhܴ ge]x b4 @˺+ku@up]I>+%!,t|L+PJ ->-#m1|FaYv%=C,`[ [(.%7 {nv3`F%C8wKhTsJ+23 $.0P(QJo0b% N; ig4$WBe p!9Q\eg"TGCTJIJz6J^qUo?pvom#BAݙhMWO?x>~,? &>vkߑ׾#yt8p[m.7i*7y,So~PPg*h;H:[Uř\}^oM d.)GɈʰIyU"#ZdrӃ9!内v 9".fij[rqkgi [d=s49 o6T;@M1*텃ҩ6>(-B-%1H R>\ 8}GRҩ66(B'D@v疅S 5.(B'2l9'Q)6(\6# ̉MF_2<jƤ00MΪ_!ptVUX~RH9|7~~:#>1]^3AnWa(XTZӭ6t#f:FTƒ=5O:F*)khI(@hBAX| jXv52'[MT6t{)KSts49o(C6P v䉠cT$Jap V:F`Ũ%QK>`mNbQTj(J(LWGI6^)J6 /MRkTzYNFQ`4.F4:'iL8qsbI @4$0:9#,~muR Y0X񫠑FQV{3ga=m~w_zsm`tpag6;#U>qea~e!o8,t,?3amyM.7-+2o!'1f -+z EH .հBFˊ*ZMˊ(Ҩf?GEf%Ɉ)LU#kKRUR\԰,VBQG6N|։cHg 'y d%|"×O:''H~J0#fNu )I(/80̝:;:@˗Y$e.% d\B(%\d-59qjȏ}8ul*5)0s.}$#Kvn.eF\琡jFN 7Həe NL˺R=pp^1,XdDzlkJ׾}+]w[iUD(%4(4?+Υ7_E T A949^X-ۃYw_J*U`2ර?A2cי t]-15ChYs?~R?\bv{bKS9TcJJߕxj}<&+qu{.]glDmoh{TQyf<9F$@6e=<-+ZhK &e 8z{"qW{C:NKb=x+9t=LҸ߯7:d-ީ'^+C3 $Ο _d 5&`YYk4iJ4t1gY._@kH}^ݜi{Eq(կgN]/Я$]UL=%+mN̤Hz8?%K"H 锬|-OҨ%2RGɔU0.teHIQ9pVu `ՖU7M_Y)ڍ++z`!箅7rR#X߱?M vTcX0 x+eBYLHHg,iD`kLU"dlY(HE4F{+2\5YݟL ֆf׶v\;{ N=ylY3[m[,$0u ̆+sKĺ7&-NXz=ٳR ],IigL0υɒpvkؤfoHI(QgHl\zsR rKo0֩d2Q}ntL!76+pkx)` 2fuUxɁkxhJ"M]INX3\<\'AL]*ѯ U._ҔgDp$ pks>.NH(͡!Ig(Z6-CL@0XGd j7c! BQo-5AB)g>QWo]uLWIC44ꌓe L3mO˺ *SPnF0a}s^m8@Zm$ZwU09Ve&Ak͈W֮|z1,'W~K41%/`1-#n0|u[Rk6}; ho-Vdžj%GIKf mt%6}-KpRdeŷWՙ +}F,+V96BL\3X'XMvn~Y遐聰'6aU9;x;B"^]"HuNQ f(r@"gCQ a2YS-04tQP4(Xz/;un@R^{nR>15yf P.$IݹR)$ф#6ץrt$P.ِ_e$.| G`pSڀ88֤c@zTrফd_fj,*A zgU"r4$2"]kഽJaM]-Яn\1->3TΜR((sJ[38sJ0::YhzL:<]f,D(*n|n]Ʉ_I' l=CaOP*fBBtPF% )tV3-9ӗtpS5۰n&.닰8)ΪE_|Y !iIuNC/h(==yG. ivœ1ڢ7 #nRh즴8[M)oݔ^0W|>Jmzgi('{x+ y ɼ!Gi(2aG(x2Y(ۜ2"r \%xק{c*x'/~C-j>!OvmM57vXU5V'-0O:|gﻻ>=PڣnH_kBw?vǻ쏟Xݞ;nW M%Ja<-<{3$hHAo*k(ZT¾6}UPEI-CWə5r'm4nLzyZV4 F?x th֘U4+j[9S`|l:hMg^qd~0,IJp4+ug}.C4. @w14k>>$Ή3PJh60(u>՘<^gf6Y #9YIfgs<@Ea&.MZr\VhҺ{n}5{qAVE$"|Lș)8%ڏbqKQL[JC}AR<Q˒)/Y6C)T0>>4k؛ :<]f,JWPaRK_ʦ%N4ue~Bel F6Η$hV,4KX6=JΠ6yS(l */8lBRihips`0F#>8l|Ti|~{a_AV̷ZKe鶃7J ,"*n4gP$VzTpYsbt>}z5վyxc,K0 -]@Gע$5G~R~}axGK)aN `cnQA)jN|v wA7bFF 2朝8pܔ"]aoR8M>oyUÕpE1pjzwCm&Cs|O}wo1m\C_m<ř6`Y=_5í!v-?v(njb8$-VmQ>FZр2#ǹM ΛRb]!RbRW,dE&hS~0|4 _ S?dg _aBS")yNfnExb@P?3 yJx{{BI%N_b,2+La@}bj˜yIbĘLElI nG4rY>rqm(LtY8k ʨi42̳nw_ 23~8㔱h}Yi 2={p%Uo-}ݡBB C? ~ &$XĚ̔k N̔n^,CPVWԀݚ~݃$A>Dl&10/*_ '.jXs\W!ㆇտ݇O>Xpݰ:1cu^^}zã_ *D%dzxtD2,\u]۶Pv؂뫆,1w8Wy쇴oYǗqrLE-iSsf@b:g @Ѻ3~vUiM,o9I1deΣ)6\+hB;dڞihos8S=MJh&>zAs(:&'n{>LO;7"ϊ\ܾ<5I?Z^ x]#eԽc52T}V5bP$+tnղH`7LyΑ=`PNG(+9uH]C1Q*e!gǖSD V}>xcpƸ1)|4,M -ҁ[$h`n.RG.A\[(eZ\%4 ؘϱq!*M}B[͎KXSzir gªn8V@VWL¬c-U!mTž.,}/)g#+ 3[6\޽}~<+@^y Ѱ$FTi 2a -*Jnq{l7ӫx)Vh5+59?}RdReM hSjƝmnڤ=ԍ{F+T1?rKQ],w=ď:8pp(ӦmԼ,܋[3|,ԸV 2ɋR$۷iK/&mAc}QXrN0Rp?aD[g7X`nGh)L芐[ɮ-K0ݢ`ưxc(ٕ1+sn3$1W|=u"L{ʜq,cZF̻i*o[4U}[[ I65:jWָNMro-<f$)iHQW\׆;jjGZ HTSCJ݊0%RU3GI5Ĵ1 AS4$<0|QNs< )1Œ0 OEbCqQYHH.3lO{b Qs8HQ q"^,)MRŻu){f9oJ|eM PRY4+iv 8) 1k7rHhQWZ7Xg ^K9yI1S !sn%o+97wDǹ)< N։S=.NIrr>UZO4D,,A b,AJqYg^4ϵ9HI>w&_g€8ϲ5: ElؾW?oR4dn*␗5A[¦aWdNiUxq_ش ߐq٩s*'^< y,8 ߫ٲGz#~IP}/rbƕ;/N"g܂qs1&QnGĺLT[<%W chl 31|ss~ty%n㐽Ap7F]ךEVFکipb,_zӅM޹i1w8wnDV^0;J80c=MAPVb$u+-|zG|wM90#EIH^j9cCޛsJA7|iD]5%TQɖ5NӇ3&T"/bPR0$)"]-4\/g$yM+QW"tW;'ֵc^Qü btR>C+%(ģ=zw\Ѱ<]`= >Md@+ 0qJSm 0Z擼$hӲ,$TO2g- m68ɗH%Hqx dT٪"KZUOIZUoe6ZM'`W4ivœYŒ*#2Nq Gs3)Ny:>gH>֏< 0u,EyPy С43yPeÏb<"dpLGD.Vh5*Fe•@UkwfgQ*D n8LM'עJQ.ڻ,P({R6]99r3&l b(5g!d|-QfyƏf5J=?)2ykKݢKF u .vqxPOwT ! ]wlh[8%gݞuwC#}Lߚu(]aCmD8m̃ʀyPXj;^t4(_ -Qs4Mժӎjv06}հR_Ur !fPFN/(h %jX%=S>)ߔeY0\:Zl217^ݡ2? wuI.guh[mסm:n[wƇ]ͬ-5ce:2j}@-}x])|A%Eg!T.-^Ӈĥ}u7A/,0iHO /?yw=s[;_i0/J?f1x*k:u-lBVP{!z3 -65tS LN~%羑3d$iPBg s#E| Sgѧ8.oyN_NuV#%(d$M︢Ijǫy Rf#EQH|-J|EĻ>{$kaZOĈk$ql@87V`}Ϣ*]"6;+}fjU$OL" Z.\nݵ!jYEU#<S*qAT#el;`U= . c| !Jo?';8hcYi/ .eԴdZN꺛 F1a`Rk<S$6L!0_T< _Wsyr_X ֌{ޗ}33|ȡloXj!w)*"=;DYi"?eļ1 jM[C&zײcXU5]u҅WϑyHRYq155Uu w盶Zߚnzl^4mАB XugiH^@CiHγ"=#> G2u,ϲZ5=Pd #RTz;+:CJYKÐM$(rSQJJe !0QJJ-sxԈ q;+X efM4G.4uH]_ߘ4mDYmIdbI! CC=.CJr -`>*Ĵ"Ͻ`HI.|0x4 xuv7Oc!Rra2֒ϵi-)>yvf!EЪF- J;C%4 ZϱU C5%e\H#Qm Hg-VB8;&}~aU͏I XU5[p+@85./+dE3nMhKi)+o/~ub|I0۸g0S)ج V8__69cTu{еC3Sqkn~6%n1U0;rhiemd;ů ;IdqI;,gN"K7JX,7E6{Ac2bf GȂB3zuJT=mOD֙[sO(wI"S{BRUuE7tlTRê5LI(7}&JY_4 .ez YMFixb(j5D1wo{} zAWM÷݇cݷUwCctBpA'W^oAsO݇/+$C޿^exCx" ސxFC?Mc7.e}QY-V j9';}v(Dc, o/oÚ^ zm}2L1t:J(rα%W6-2-k3k$n>&pKL0/RZS$jlR4u.H]fm0AW䞹 ִe VŦs&@ Qи|쭴X$ϵ[ nY`QMǫxKQuhGEM\[ڔ 2р& i@,[bDLӶ kYVOìHX<8$߆4qmU-x̔*sZ+&;+}iU,Nm:O8b:^4;QbN圱?b_lLX\ı;?{Fn"eiYT i)I}];ǎgfAaKFn6e)Y ;nTO%yޢ6WȪu}gHȪMym聯U:\ l/s*zʢJ=lɱ~Q Y mpJp $% Ok"_NE-IBKJZ HDKFk`xpbc'Gv[tά?yޝ8LgAiرL n?7?3y5y`_zv`2d0Kv:%Ë\=u&B]݂*nW`(J|(o?+3]Z1(lȪKhYuٞ/|UWw YuZ"Ό2h=1 TX뮼] (tE&gz I0"a|ڦkH- mUu?` .m#W%WB U-jVUY<#Yn,F4KePmTV#PMS(5Eǒ*jt%ۈR-7}K RT0]/m]E}#lpc嗇M k_y|g=<9ޢi~l~w狏߸+q)Բ?jI,?Zo^rBכ󡸽_7v͍{+౲9m-??I]`6kGf:RٞxEVbC]wI)`a& >i6=oհ)"VB*c&"VD&))S(!D [CKyiIHCas4})`R^&1-j.G ٰ.ehWYYS3ehr4 \)Ъt\m<6~\MybL.Pֽ7|&Hz@a\J2nprā4"T' yv*2U$ЉbɄpnf])aR~J]nXsׅ:njPTQF 왞Q ;X3h.<`#t WZ MC!BylF/i[7čd]JW PA(Ef w{-~?.ڴ-i޷s?Xl]EVst5iZ|Nf~~뮋.Ewop%EӱRjh V隵cEQnrU-~^/~ReW-b?ze..ҵf5כ|sq}뻻wu7iXm4ltiTԍBE]*Z֘YYŢ6MU⻱dld7jۙY;۞fւQ'iYӲ$--k#epsYYa0ZVFZ{6$dLQhY^j?;z,:I 3Tڧe>{^īqa@ԇ k:)-_d%@NJˊ.5b<(1/PԂTh!G4t1;G QY,fir9Q ehZ7taA8-/*NO QtOJP MhYoZ*/XEɀtM bnZ Ӏ(-”N)(L1*96r 9IsC@KKAQrZx66|vZ<0PzZVd  b40[i`>o8_yk)r6QA?=Q*u-h@cV6ɓ֕eo  cQqhXUBazKeoMD54΂Rhc"y`&i%/?o#AۥvYB)-jS 8Si =S=pgABmͳLi3^B)I30"'Ťl>&g7Z M5IS5m]"QS` (뢒RXX۔-<#Yn,"c,y0X;`u% eDq355*]k =;{)v8ٳ!{.խn=NcL`َJwyK1762ucH:w[d׽K$ [ *cxSBrU,j /XbDN[KYafX0!o MmHP J~12p2RZb{8,zABTDP(a<^0X!2뗁t ɂe4 `l VlF> `Q1X!:D è.aF!cmB&5pճI- quIm,I҂Z^ҵɌ$>u70w,m՛OwŻ~}2=>N@ټ7<4reQ;|m&Q2cDa"1< ~Q쥐!|W!OM< .ESn-VEڥum[࠳U6h|>04/>:q1s@ND/ů9/ugAjųZK35Kg$GO} OxY`q5cp4ȔԬK0&3X0P uk*XאnF)AWU6Xuѳ5E½\k̀ Yk9ˮnʂ'LC˺ U&4%AyVC&HL:YhY"hYtݵuJ]+][ieUl+[ $FMɢBdhVTK][-H ꆼiYe~}h&WBXm4ˇ$kU\ elf E4=@TP4t[VM˲Eπ*A`Yڕ8'lеM 2XHya:e_vl T41tY:ghYçoZ*/XEQ6Ҳҩ66-+@'@"RXL]J(Ըإlr)jN)fbNZ@KQ4iY E`>K:Z9Q/X~!*2ɔoK' \Иl5|*o ֩L7:y0$z>ی*zDDmwS8ԲW0+ӥ?֞l3<`]yQ` P6e1򨇱#2Avn&"V@Y*en'!Iq R sS5X[Kc6R^hT eKi5m-<Ydl$˾o;-kg㴬FYhYҲ6ҲX 3R -вKkB>gC2+5eyut4-+huTڧe>q%;4$u|ܽWb5n\T/n0oe#k=wSMڍjۮy`]c}dߒ)j-ئX6RC[-]ۭUn@h%3}kn?8%vʼJ!`-J7rbh_:-xRd(J[̉h7to%%|%ƴ`6̯Yڕe^̯g\ܸúx]"ž;pz] vq^8.j:3Hh;`*7c*ZɪzHZ:զ&dDHHZ: Ef6}MKBIHO@B[dI;sqOsYvr4\۬y0,Y0m9=`,  o::^o.jD1V[ /%'g=~:{bxYdJ8_+Fuw$,r/M>ˢHO޵@C,C1jc`EMʰ$Me?ޟ@F:E7д^k(:u٘bmr W.m|b6JQ:s0]DJNΜe !)v?poc- +y d: `PN`(6cFgXDqp$ev^  }v:QDHk6\%xТՐm@^ c/!ˣ.Ҋ. ٺv|:NTjrU38  :܀Z p:;|x1-^鐆4yx) ƉJab)ߴHҐ)&#JKC:nr d;2&G5Dc<5l)GSC׺tـc.Wm2 >Yܪ(S<3QģeTsAGDhIǴ+ÙVTi`:!{:͘ zHRI:զ&DdCI |ÁM *lN:6Sl`SlV9X3|)0&=)vhV ,_ҙ Cn_fyPIItV jlnIBf!N@S8_7 {`&!Kߣ! )^- ("NHC+_J +4ȼpN!Q4v:REr<.wfx1ةiH1H!dg:OvktdݝcA7Yr׶պp̌ q :Q#qq֠ ٝ@)-/gwwdX' JV,HaG'L2XWs!/-Lg15v[QV\vՠ;נmB- Fh_A ñN v.=[!]n ܝsCN(9x3B/C .|Ff!]ձ(ߕwbZ-:g2v&2Bfç@U`&Y+X7)p-r5,/F#X/qqȮZͪ z~\Ye@k5}O菿8pCjIi*|HS CPO E4ƶ+ twZkpMSjkD@.j.TY+J'11~lNOcdVNHYB&d(Y:栚E&LNJ:D9L?4p=/lz$PyVZ5|*;S/_7A?1Bl 9hY{$le췰`{o!o8xFEPaLs' % vJTXdB8*F0TXJt<ĽlŒwfD1a;cZ3^uT{@:q*GoF:M:W\NJW'|Ũ% $!jẋ=\fLćPT+'dlBp>T:٦C 2 "P ekU{ŇJ' |r󡒮uαovfuBr0T*̬ H X)gEGōsFg)Zz/"@c7)NGz_m0^fJP}0r&NeQѽۣi@9 h@v3F,:1FrJ-02ގsBZWզn5WkQ֮Lk(%c  Mzj8$ ?p5H#ptYhA*LJՇz_ Ҹa)KASUCJHxBѥRzV@JeQu)ejPT\i2%W-4MW@(OƓckO[-8*e)VaYSՉר+o|/F8Xx,xM6 ~l{ǹrWs!S5/ϟַoـͧu[}՝ ?~mW^< nVnڳM#}~C~˻k|X{AFƱoTqT}^>@*$>?vm+Y߷/cz7s. ?kaƻOw_oo{| %m-X7Vk^ΰֺ h^}l?z%ʼK!-C1h[wv$(w),mPz;%SѸowlUoH/HsFWiŗ#40'c2Bȧ,=1ˬjTd2Rmh);32|]GˎJz ᾭt6X=z=7 ֫^9:Z-m,et8tڷŇw|*0>HLI^ݟM몎'w3o_ң] r ߻&k;h,Ƞybd |r~!1rNIIo5GM"C vP:PH>fF@)Ʉ0p$/lRl(#~M /el_btb0;ι/vKp{Q'Z9}`Sf ,>]&?O[/uXh'`eM;|˲O將_˒iIً}_[.Y_^-0 ^R"3Q"i@2aFG뒁@EpnC6 Pa PDnG&[5lrC4ZoZGM֍픚zrch&SBd/VT0W&HN%NNTi |*S4r&HȃW%j(O]: I4QYT A 6mJ\34!szMmcJjs// [Es۽:#n`\9@cP`K нbhze mPd@0>_gVJo 4Bp68ǭa,,Ή㼴ɇq鮷mVTԊtp/t jBsT^B;q}^3=|7I[1Ԗ\[ٮQbT-}U|#"?3USWސsu1R<_^i`3}\8rH^_Ǎr>8VXik _tfKH8,S`}"oqH"]!<) ͵C(1>T+[ʘb.N=(pBNf+1& EJ4٢B0YT+JY"=.'iɄ]U:V_єğRkŔE@X[Q>Ǵ 6zF=aSJ8)jC""VHS`(*6c.(RnteYjK\Nm'ԖVE«243*F'H%$xU>ɖWˆZ:hs"h8_:{sDe3, "+~F`I'cE`9ۿh>a]vn;#z|l|+"_<. XmFVxu[bFܪQO||?Z t'o<')rCB8O(BSIHD=Q24ꩈ")nMvP $V&E뱩"N˺{5:L,R.BքFjllSVw jL 4ď1RjؔHJx* Z1V(ֈ+m4u٦D"|Xia6?{z8;'NE2,R`AV=.;=A;.֗c?EpS`1;)J--c?MOEOxڌO15U6/)F'R<,@$P>` E慓rHdW5/+B'4Bّ>3%bivSֹEch Q,Nv>đࢾ@;ܰr+K/koams6^Ux/=}?w_LO7 .z/9[)?Z6a@G7\D`HD Pp4#"Y1l=ܷb96(`cW뷳Y7xF# 0 n77=MU_HUFs!9zu,UvCEf;x]]\\" ^8$p`tK;Թ 6 R7--9s۩ ayl{Nn;i@] 3A P3tl'䢗s*95:#ֻXnT!Ѐkcnc'D+3hn|DH4\Sy1g+l=kuE]1w5,LC'CP ,n)`+&xhO}E)rʏvv{33:o`X|kCA|Z%z?7C-}37C-}3oiˇMS5]@c*]W}O7wb;3jaS5<8;|O>p(}0 O4n5_{}n>MQ*M[V4}>Į ZV#DֻEW UEѾҷPmT]ޙ֏jv*QDMߌ'Skd'ɡuhYϾ=E; }aee\ZN\DZ̾[eȍ'^ qẃ4TȢTԊqⲽp.T>BEeh?iq.1 :͘ e@7͇s63*B'2ta]$`(L3btPR{Jwl ~Wl-[3;)I$&9f,xIBD¨E=`eMN!9-~~9'ٷu_js[ CF_Ɔ}hӼ)VI%.a`j1"{ 9`9S1"]> VF9htp6VD`v3rptsXJ+aS]]dfU{Ԇ՝Ծخ!ҹ3;B>1ubQWA jqƫ@I bN:^8N/?G^Tg EU5ӫץ֦pv[D^NbY*sWB_s6l˽IPdmw1^ R|ޟD!.wSʣJ09K4!lĨ lS IUt݅yl7>`A`_ԱN#@-(Q%GCIo||`ϕ7'_|}vKő R$v}Gϣ?->Y4նd4 7z6>}n5N?v%KţO> i_0p4⒙t9$Zz;ݗqmlPUFo|suɛ+s+_҈wш4.ݵG쭴.d=ubecNOlܤ MФ M͸I4Q}UAgabTMtBKz\[7h &~(ؚ+*C9~@r gpF{pm3@9 gkH--a!/ғ~;+VpHZ_]߻rutWdPQaZխ$OBz2:1cjN#ٵɷu4vPk+|vSjlr#з/J Q|pU2>9 Ocm k;7p侑T'A·BgoDCC}c-4E?_3[hZ)x6тn~w,KL"b޲A;d9t&蟊܇p/aOk:>u#|/q`hԐ!5!ݝ\,u"aw0{ wŃ-+cNdK ցn [_4OpP@zڎ"*: }%\ CDj%eГvKQ^I1uOqE4psd33LtHcN[ctOfyڌYݑ"8z3=&kYoJ7I6f%ld%lOt# $)@IMh c({d;Ǐc(о'X1N &{NaRE 4Y#cc َ5e70mɍIï_L$2Nx^`/|QG34w_TlR l2!GJUqF/Czχw?w<0t 7p#P}qlŠӞ^HLN}6vCc˽L7݉ɛM?ͦKm6)oW!bۆRr|FX! 9kfFhZƀ!R83W2!K泓7Q$?x.tF?#չ̭vɧ G6FLrRʝߋ[°J]^%)L,eJ5y蹭E隶WڛatNz"9= 6Kt9c(<@YQk㏾,sKtum@=7mMp~0FE2lu=M: ;jð5j HL/D_Ô~IH 0,/)$#5"8\ZD9hdMl_^8ŰQc4IҌm |0YYq `׶ߑvWaϻ[!cꮆVykOL=5 KRGw ԉ3툜|9I+xsyYiJA-q` &4 HpS1pmT.o4Z^ DZ{s]N t\ g:uŒ\kHʁ uN :\*XqNVH{{̷zڠ^xP1Ӎbacϩ5+#s0(hQbAm^-<38f{ڗZ4ҧ> i\Kǔ;}/ ;$ip_#UHUJ'UgM'ɔEYUHB^ءJbRRr[ zC]% 2oqЇNB DmմM8'<ߠ'9ߘvX:e 0Vsm^vT />kLHQ0B`%5ӈ}U p)k&kaD8460bu6+4Va21G90XsMIʵ=/BG݂+st5[[) 4D׾ߛ=8kD{gA6r--(ztRHz8@I#p;fzhY>V)T |W2tv5/K(PXTb0KE[2 (OIHI:•Sl\F ~] &Joѯtچ6q=/0,ʺr]9իz`b}eSsџtvœ({R}1օioa/Sp{(|%?~}qfHi~[`Z2n!ɤ"Af0M*H =$5xJgX(1i|?ML n`%9qwDUwT7ͳFU)ߙz?bDX.4+I"ݚRJ̹$tdoK x; ӷ:ĺ7'8w$Ѧ]|- cF*ŝg@WMW _n߀ྨöܝfָf( h4JATiYwES-1:\ؚus)RL@z̡7eiHI+Ӳ;$SqAz*Sn/qj{ɟɟɟɟƒSb2նo G5¦m둍tU؛IFO,܏O!O&!/-`)M%Ja*\(mSBi4LۤI:,PBu වE鷟nQ/EeY9^g\|YLzmU %-k)!)ao-Qݧ&8( ˟eEgX[%&N,|$6<,$ + g!$"!0b&AQ$7JYL.F0V}B)_08F׌rDns{ 鹅C^]-G)`P^]{%S(sᨮ$&޹-(Iiݖ8`;IR]N\^ZR]-u8t% 5ւg`h&0d"A=:*y T`")XC\ TjzR(`B)ܺ7)$lON,H9$@~Tޚu߽n XgJq=M{@`cW&A=vH&^z}S2UU^#+ wAx'%(ѷW0GBW)GǿW@BgI+[$&/jJ~zHX)Hׂm\Zg{TWW1;.ͮAՉ)1G?"dܳϫd& @V;S2QbGP|1/N+`甕tyZ0#d&vbId;i =pksT۵%:l v8ŴALOfJ[S9Ig'3T L^:)j4j d$3M$OgW 4z13c& J2S>W 3id GM2,㌗&)JVF%hit`M\":_xٹ;ɪ ؂ӥߌ\ lޙmSve[9b㪴[q9Q!S;|XǪ~|8{}#Ogǧc&1}ؗ yo&kjNjc3N8=ۋ8Z05N}fkGkrgm>}6$Օo5ἱX8u+I tnK'S{IjMy[rNi mMLHw}"v{ݫQwհmqY=~`-r){^ybppk}+0kП3ެвǪ>`y~P=zrl3`c}~AP1F+Ӯ갵#˭ x )-h[47\!>!>,d?= l. Ue>׭hwa޼{nn KT|qWǒt}.> MihOC}7*ڸ۶껖;mT5ԁR0T+ ӽq(?=Pړws?!2ޑ8߇:=->?w??~V?$J˞Q4C>t)X* ۾i-cۑ>v@]h>2َ=YGG/**O:0oƓskO  ,4F@: ]( 08`ru/K1X;D-+Y]e\0ʬrc)-k=18b2 _çƛ8ìXQF5d&`YtY$ikZ|]1t[EչrpMTtcԕr$SpjR@% 瞎j+dqtu4:(YҨEƇ|ؐi]6˧zl3ӲɴռlRtEJkJO>Ֆht6l‚XI5'T3l`hMel_R:ɤ[ϕSl'/<@v~lU g!iVuWV;R:9ޑ 61)5{LX5;azĘ 1N{io쥼J!cluTؿ>3o {ᒨ0eB2J'Ba&*r잍 +Hd 탒fF1\6k\0u`.nfk阖յu+F\[["vtulm=*T`d+g ;<0aiYs?45g4)Xrҧ4+ԓ8H*i#aTw u' ZfW۶MW!ɞUt$z;[#gxiYG^e :\e}} eMAWcu`+=qZ-ˎxdX MH6L -+Y{Ҳt.@[NiYwF_Ʀ&P\XCVVZe3[73KYYiKrtb{Es9i%տͽ @k"]ξ$]C/z:Q_Ȇ|,]Le,Ѳtj91<zZV>VeTdk(-kXL:<]f,D(dLXh))L\¤MJm)XR4: 02- S{aHɸ?{ǭ"<,Yi OYi,6g8v, }3GXd_1lYndHTz4EmxYvEfe%]~Z8Oؗe/-fQdeMN@R9=8uf_ h|sr:`Tf&՗3"éD1"8G 5uoWCXYXU`m\B;6&G{EqJKl`A~/H1hAsɍ85`&=cS'v^'^A*ŝD<-^AR<1qCiuR> PPN`佮>_WD;%ZD)v(M>z5ɅsAWB@2۝wKAH}VǤF]b*j^u;Xk{XG02T1xe`ԹYL7(:K,5Vk-+/x’&vU]]U^yD2,"ˆHTRZƮȇF:&-uR9Mi|k V:Щt%qm(K:ƒ46ʛӍ [7TuYNMe2 iohHz[BRzW4"hH&cR/v ^sGC{wƐm`C6-BC V.ꋎӐL`ӐfJ4ٝ{6%A-2[hB7W(?8HJC\d$'eҰYS$kiS5AS2e9Ѹ,ʐ /fdv]!MZa=1GjiHqqB8 )l }8 )N-a 1Ӑҩ )FEVʺ5MV.^j \@tɂʄ+xf(\{Y$)錴tZVrr,-GCW9<V6'~W"kE%Cky]Gy}wgfUGϰx}WdsG}YT&m;^o}+|d LuKtJ\+UZ窮 4%[r+m[yK燭/՟/^ܼP/zy'|OCx'۷;^՛fG7i=z )Meפi*!+T{ふ ZD.IuSsb;[mUйs.m0,i3S9Œ|`uhY{>NKȢDepZNe:V@vF#(Z^>(pY=c`CAԳвad_tgT!-kvNfѹ$?ލLQ&e=FQ1&eM-FS#&Մ/㧣S,+~4Cp- [BX59aJ农7o#N1Q?} L 0/Cc`T}8(F;`TUrJSSwft5@ˋb LBk!FE{E~\|DUt% :[PLa+E <5`&c'v^T~ uM83I5hi,YäT;Zѓ5aT)Id.8!hɱ Ѣ.qxh1%h- k{wZ\]JX{|jPk^u^ |vS)+3|Fh Yucx1P˴|:wEu8UuaI_. Ǯ<7Wd`>-suE.z=CZV4NibƮ;x;[]_/jls "UiYS_45ڇ6iШ4]>8Ɠp.PTQـ+7\cL#3u29[6J%ƒS[K@KMqW@q; řO[Bx. W{~,􊨌FKW ѣGLܱT GS Hދ%l0 $ϩ&u6K:9ڕʑ&5u3KSy4n>KZ 5v+*GӚ[lE - j<(lN?B䐣8rx>q&Ny>1j2|<ݧx1f(LSuyxHNL:fDJʼns 5+w)N'tI<ytBay#3V][8F%/-kU㗷5<؅E R2<%Ō jhy78MsYK_˳ y‘<+\,O0ӶOy (|Jxʣ4E+˫̈́j8Yb *ђ6i5ڟ+22(1?^itPYPtaB֪Dʆ#HrZssI{t/H bN8Z٬AA<1P&Hq'OK0b$Ոq͗6*E"=!uBFlz5D#r\F:OU.Ëv },!ԅ!.U ro1+=bdȇ/t +q;YV2*gs <]&=/sa%"'+oCO$/"FȀ|+!-#mWZj6F r-ֲ8q B[Nl66U3b,iೱ$zsԥmu%ZZ#*j:e0L@W[ZUwf`I"Zn_ww7e=K9HWͧ}Wǝ`[[_~7z[B> ^珽|voog_y Lϼ|{ss{?}K͋+^n='g?߽ Xټ6姯~6¾ւ/v$L=wn2kv(RP:쒎BLɺC(>>2X x6yҏF+a]R(JLPEFNHEZa 6L19j#g`fɆ?5v)KZbf*RN#Ѵ(׸} *qx(R؜j!pz(R~U¡HT(IECvNŘ (\&"Smn(RN3"l NlJ)U5/*\'Ҿ|zvPHGq EKEӑtEX$6reI[8Mk0I4Lz2Tbz@]49aJPa֛}3rt\O1\myԜ\ȭpLB< E=0b2(*9]P '@6Π^ Q`A:ig?WdF2{rAa.][X`WVFؚlxmE+?O춽:l2aI4IsfN3ָh%j$TR)AS!ǚ`_8+"Ցwpk9U^a_+ Ǯ򊨼Snp. l@fY^d뽲]i!3uIJ݁j?Fljҕ-rֶZa,1Dl,XU]TMbW-;c[J'%8Ṳ+u\= XCv>bd,M8&n{bb3o`s =)ԇIn_VLoo~<@?zq>޼mo pL~:q5-&jqM!m0RAm=[~*GGwSnľ5ڭL=td vq\˟3b$ׁѺmKUT*\cjƕPqzz@cCQ6x8K:F@3o\<\,6/]gm4w%/lr92Ek^%<rD:OM+ ǦdhWVW9 'ؤM a<˴Ѭ*aEŀRq6|) R_?׻O'WbST`nCK$:jtm@fWBHPZIIt(͉bcX -taQ5_Xkpc': ق&[m _;\GEU:,+Ƚ~({dJQyx:2zXPLI_ď*g}/ h?+9~џy^qmnzחEKVm~_~6m~zmF-PMF7s%ug@Mm ;lhM]:2~kVϗWnno\5/zy#?^">nݾ7myu7i&|t'0C y~Ӛ_/A:y.m)ie"u>z o2u>*9@-d֧GYx]|\}T }?WdAH\l)@[S-)ä-l7=yXu޶A1jNL._&Oı!/IuN8I$8tI ʘr*yBYˠ(Lt2WZ2!!;4}l!8'e`+ I&5 SK1թ CR[U\ʠ3 濎84|˼|:V 8]yhg!F8% vU%otULXQikW6TMSږ\nۺ[1bt HMϒbwȐ2S &saQ=> OM{`[-l ) #ȎX%%|(~Ș"> 6)kS`퀱Zܮ Z_e[nYZ4crƒ ;klMP:[V{LS[:UTZ_VXR, L.1/}:׻뻛'!!궽;n$_@ Y,Vl.0Y8״wL##YslnIX$U:g#l1_OO{j]~͇M wGpl>'OހZ)?N?^mRo>w9}+౲uXDաӻ?F\;?3Rog;yQ(O.dF Lx qs#LnA| w i&v$wvVA_>u}S%"#g[?fL}+lZ{D uiw?ѵU^Za`pmiSVR[?ىC&fZLl-OLaR6ZIB#nUeՠNmрi!ӄ<Vޮ [f݃r Lg; ^^yqV d9UoЗA2#tb-&5x|/;ޮ&iEQr̎2k~6WQ*0sNIhrSH43Iv7$GN' FX7=2O:m ۆlon\Řj6JQTYL00qR =mld 8IyjgSug53K(V2)&f ivf3!fR}KgLG726@}W^=cUHvqz0 ĉĽ!qz0a87] ū)yGW'T ۱yx5 lzgsLQ֥޴Ju +hi|,W,F6m.`(8fdX8B6QRG4f(x^9  5 wV9Ev  N .NҸt=Qtr\ZgkcDٙI\V}q3=5ZAc2,"ZiSE->)mԥbPAM`-u+ЖqM)D ec˪?t X[ 6Wbj,-Tm{jƂjTcIXma+!?~^Vk";Fˣz :B2\}b#Vafj C돴O6?},ۦPk|TD)bSBBKu&ͅUl1+F1)6`d&F y ցkeΔU*kᑬ1[%kw65?,pJ|`$x囑AD64tRč(FS>I^w h1_ b>Sf_J)Z}(QNk R3_" tWk?⒞r8HTj*6 Tb U `QBAPXD%?V&t| =mFН PaEv <]odIrlÌq݌qhYX~JEiKˮ\@DRY"DZGH|sFaY+ܻ?mC28 bREN;.*)(a W7ԶlSUc 5KXr,9E_ɪnܣKU~򍤵ڡ2m*Gub`j\VԚwpz\``Cµu1+i2zD?1+:B&"#5ZvX 4&gE_ZiR~]Tji[}*+ 1r7Kaջ;{k қ万ԳZQLV=oy4hm0W6س,ut ̓_n,z=P/̛ms4PQn{&8(N%)i"[&ߘ󂤏 Apca<5m2Q4)=bY'Ͱs#Yt6 l~hgO0'LZ1MvPÆa1.S#Ĩ(LqHxjtM!B-1M: !xvrc*qlSmjqN˝:`AmM'Hrm9a) ?ECl(@d jғuc/@09;idV$'O4"-@6Jo'noaG|r軻ǃwiS9Gf_xehk(?OO()IG¹1"R2G='h$jU22|7ë FF*+ZMxzX䎐O뷿_ W)͎7ZEۅ^9WUi-ԕ)Zg@t¦ vB^5*a 4/e _v/O8YĊCŜ@"_h:qRu 4K'K1U9EPR)ɏ]uP$^*ɇ(54L/C,Ą}X(F\կ3ɣ\c*½`_U?Ղ<؋ߴV̓C{B:^ oHelȠrbP>뮕^wkVgPZС/Zۺ \Q`m)AJ\R`[iK̠Q ʱ/?hؤemP+Rk+f]y.q\ JcikoBX|cb,9Eha\ z/!NF^F ZYtF\Hzv`eGW1E]8)llY;R7H {fh/,*eWdbF9K?k7tS-Gӊ|Xkvʲ4댬Q-ɑ8v1CSwq0&߱MQ,I1nGEl:]Kה'GŨ%٦C wto'Wx1&B(Ug6D"ҩ61.J'MXfqpYB&eE$yr"NP11RI1NONJ=8T8׬Fx0.6"5-é;E`xkz;9UF?vOv #kCKm*Tod,cQ#KqV_9S-PKEFQQrZ M .MY5lBoHNVU]=I@Qj!P`{~?Xbos}^ɟ`[VQ0l/*(ZdߌLKs J@)@xzPd45˵MJi.W^+a( Zo.C֌칝@nAS9|$։w8 A b.=8.c\6Q7oKF9 jPn`@e j-u@`qLg^*L8Y/qBk%O-Q(fyd#`>ŗ"RE.T`W,)^7=_ÛY*گ9 Žb8׽(Zys_j:ΑLwë2/0\Wd5PFVzݵ뮕^wJxs*4XLYB.5+REEY`ݰ5U&-\͍|8ȗn`N/% y‰}`+o]2ߤ0mއ1 !8EׂZD(kӠ)tMϑ-EVP-AAE-MmH?1EYRz1Ypͧ1Q7+w8W?Us_se ([okZ VPjXѰjTON6?s!#/|JmYj+ԥe밪8ljGK(4U}7zySw0E&qaWBk>|7/߰?)ЮԑǢ;7Vn:x,~_7os?}^{ ӨA0e7v%*zxy{O6΃Ho%oۚNS;p7$9?m8)(ݓ޵ɠ6;D-ʛhJG7-1M'mAߋS:\mrM4MK8F7m̂[ bN78ےm\>--ʛ6eѸ w4G?:rsG!J)K ¨7Y3(Y3yp>pZ ] EAĤo|K4ux IWwKp8Ξh\Ջ AE=?GEirC?9N.ze"*AnhrmTdq},Љ%r-tAE2KOg2abLu( vfKxafK'9  [81jꤱf׈6ilN,Ϭ]6{8鲧_/*g3_ɺ8N(L͑[L[`-p`0|L޽0<|><'lR*GhWZ)`jRVlm΅m+YIN(fgq|y rS4N8 G$i!D%DIı,Rbdy" n/ '$ H¼ D]ʨ Z&-JA*-=U*Y;&MXDB]N:U,FخDӮ-4V]jM+Fdw7Pel-c>0t~ײıJ9J^> J|%шhiR7^p$tf'VR%Ը:pꇒ2~I&Mk-[MVrIcƮԭVAwuWouVsR)et=nʛf=2dsڒd9Ix<.qH*$wƬFf`ȭYmap 5`V#g.Y -|dZ&@H}m(l&W00f,Dpʤʨ&UFp4uoW lQB[y/X)08 .㩊,Z P001ʭjZÃX7ܶ 41{.d刃¨yDQ]rY:^Q*]@FX7.K"CQB{Yi)oQNIW 0];2$D͏ۊbU9v5"nKnl"c8DV^Ur096U@-v]r]9d +w@QhF|( ښBBɋ[+M4ǏƼLS#+ VZNCZFLnjbXR(?וêQNw}C#kLغ%ӭpMZ۶0gIcIWu yG}5%V4kn86toS,I~ Ӑ=CC%4d !=>OCKbiH[Q cEB,^O1O7ِ sАՓ㼩cYJKiHiFX@ X>2F9nL4>֡JCJ- 2Cq(+JbSc|5S hA4- A hVTdjlT@:~3\8oϮʧ|* 4d!wfiHySѐ%U.o O'ĔNH%Dɧl¦=&l,c6:Z(!Zm3r!dzlap|c}9 Z5~CvVh:N˺9q!{K]bIw=l]݇;GF\57=oOUo͓'wϯWr<}VD=HN>+JN@?d"gB{ gOOl\_`LKWfetoVU[UW7GV#/meB1{#'K.څX8.JW.L'EĻ~YIP399[}fI5[QRN0%xHC7|1P뺽A %oVH\7?bh4+B~̧lKЮ_ɑOyUͅZ+AtW@o;$PF^|T 0pNu{p0oC~Z뻟?R?oۃwFn|<8_ /76on]'"4lV|ЮIbƮ="U먗(u't_- =NzMSmVAcIX2EX 'm{ 't΀قp^Du)ǍHp9[;JE"T$KaEóy6IlGv݈59o|XFe)lV6ISdΩlE̪kZ<0mXƭ( a1[.%2Lzhj#Ά[`˪H?"//llDӆf4äD$aW |Xܚjbz8;m"G 0|MNcE.fgÍupx8* M12UT^ɑņ-~{BS/b30y_=dqxX#@z(Ē0[:(/ A6ɜFpjИ _YQYɧR*rK Ef4j-5GӢ66u6SqmyrҢh$Kw-z@+UDHhQ~33;W-Whe%NQsӿIp5]L@?ǕVNJڮUb]Z,|`^咡a](,u2PQ{ua:=H)ָ% XFD=7eD;'lQb_@ZիRBfG{Pl!VP7z+ u7QL țs*oj穋;4`֖{*Ѭb6Vv%ܯ]JbeiZ%BRCZYAp/ ȇfŭC8*j"47C6=4+^-)@ 5~fήbLb(*gĂ 'l@(JmC+X38:9taN IQt mKo?1/< [~3rhV[8ҙ$arVV͗Ya VdV#\17~Cc6 tЬs #!dX`s{qҏżܹƮ@>b؇_{(fh#aCy nN@ c[RqD*ݎ#Qn:bXURgw&XF#@1 =0C)!hx~0E &3(ɬ+4 ^@"8n-Db>0bDbqT.1,N֥RJ(%e DNthɋIIߊR7^p.%ڸ$}C0ٵJH40ۏpu;m^ $bCDJgbRܱ.vv .A/FF|[HVɈizW!UWQmykvt'Rg(G %ċ)ȴF!!8+!-#mƠ|Qmz+{¾ۺ ӚU؉ TnE ¡WG %XBz[Yզ![U :$vNvVXB@M]_NjKj y(޶HC S,  -4"VNѐXiDO E6$FLBCVyǢ4$ PKiHS b4Br@© h,ҐN)";8׌g!.rd"jkaQ:(4v-j/2Rc*H;W*iOD QmhVT$hMFHCJ h8ӁOC[SQg!T 4$HOCʧ4$ZddNx1{:1 B!62q-TIVBM?(UNIt|@f66?<e7K"^Ưf5uAaY-pr6T=9R/CQ9-P[|!r$@Lj؀[5?hϯGa3ږ׈ 89ΰD02Q`g@ET QmIÃBu8 L&r;>)r[0H( :٬ VՊ(jU][łMF{N^W8$J1.J'D}Tbg#AA4mL riqJE;0"-J uȯs96nރEP&Uҳږm:n6 I%rZ>k 5o89r*(*tsv0Kqr mQ6X3n8kH^|Tޢ7M(YΤ"cTuIpv`,^#Je32E>etЍAKs7ǿ ×=u+ʇ"|9ac0`ӒG/S흍ꦤj[ U(4mmQ* @"jnmߵREcJI֪R;ºm)*׺' 8K"ыmwT}lچZi;Xպ7ʿ{۫'PuLKJ4mOА| +cNҐ>NCr\ oDFѐKg_l`C GU)4nux7!ř4W:!oXu$2}@X^5Ƽ4aJDҧR4"{1Ĝi ta3 VY}$_"#uխDS#HI, `Y|%- eU&ϵ^]YC>Ǵ,6z=")7VӐ6? )B-6. iLC?]ŘGyHR>!뤁~rzI'(HIMXe_B"t"ȗ`<gyP pK@^4}lQ(Z4- NʺPq𐱰 )BJ #dI?:SsÅ ɞ!vwi+3"-bDPVK" zi*JڍV;t7 "PPQv렠" lxq5ڟ>.8mTuݸ7BM7VwF6#SO\q&'^#"Դ̂)VR\ՙcz.qX'#]8/H 9t祓(-"ҰH++ qJ.rAW+tvp.;|\&l5Vxe^`9 yҀK8%2 c+2[Ak0 NT[뮼h"s`yK:J6C>EM*@U+*$AP};ӪGPN ?2Vv}c] 0#ʒ/ǒVPz۰alGXzFS 5Mt[15R,IkӐl{4P"{LC cTHl}eYR&#;{UΡ<ؐdKpRP:<ΛNӐL iH:wzBjc>`U|;mYnoRꑥEH8]@q<-?w?ojݧ'Mў+遅?W05/~{ORpV ['ħZxCj_݂ƆOL!2+m6vSNWδ7'ub5GӐύ&\CuKMw,Wmhc&=?8D|tɃ? f=tPowa;/Mo?i{@5.m;?z3ݟy;ۡ}v>ܽj#7vwժguE uU?}k7ӵ9c7z}3;}|cݐCys}on޶>T`V[@Nt52 xY Yr+rd]~ʂ7&mjo$YA9i{E9:ԃ"%) ?^k&w%]Q詇 4|Lin6?Q z835REOGvkd|LvUI?𜢍l"x_QX0wt}0eG˓'|P| %I1%ͧ(4(,|Djg]s"tBvbAzN\\ll!&?7v qIҳsM@-jӱ׊G$7xHNU|rF2Ǝ֌rɗ &JO5MSX-*$l?HCy2 <H"A(=p$z(>[B0h~P Ƭ[-0]jYh49"3!>LaZy/6Fib@$bgfW'æ\&AM*?3P40Y `z߿ _37^pRC:Bb@8aA:̨)Yzc R> rחf0mI (!5( huH5ݫbkCb/zdV^x4+hUhu-'ZdD9]yBXf:ďy*/$/]E\ҕp-TU]mQzmH}gim]Y̳!R-{j 68 u_, HGң#f+6ƨފf'O7`os?~>eg>H y/:X`ocgޣ//$#47ϼQ@lAT#0#0ގG1XND%K/.kkmo3Njmmlu0)CITԏ?j&IcFsVprx9y(9 5B<<aD{g.]C|l\[YU*t(F[SIMJu!g|i7Y}1Ll)$FmrmNjEP.I=KvHCόxڡ6$e;2HuA2$L`>Wt8L]V+?X0*^cO!V`yM>gqK1[!E?lp(d_%ZJx%2C98n01==9ʈFӋqXY4:Yɕ08,l~_Óň+gu<c`q;0lp?0nGoX}Y>.V}_7ɛSvs~T5ݫu|~xz3ʳ z]w/?yƽwo?~Jiڂ[f~ݿ^ߴow7ͧ'vӒ9 ɦ_>]|7~N0|7>?vw?w?O HQtƪوkMNIR7??_y%)4x4 G+Ʉw&+>VᩇNxI1|ߔhr\lFp#%(람s}EZ{2.Ҭ%f,gZ|8Å6l$JRN'ȧrDZYo`z?:Nj1@./;. mfyNFqQ:%q)q"+g Zk;[׸&~r5/7 Q ~"ƙ>9E;ujC}ol[Ɩ6M]jW- T bLѐaǏ*3Is/qM(ER`ehk9d'CyȣJQ#⌒g'2ym_Mw'JF[X54;;ںկm)E\^h é)#"Y^`q r]bZ=ޜFfh'7s5i= G9$V$ RVo#H) `|FQ4G {EC6qsRǎSK&Hx sI<4‹AMT!72\2ܿZ+\0sGÐ邙t߱xex5dҽڵ JwmuIHvIB|W`w}WjC\ic^ zg*#iAOrQdm^%S[d%Exd%T޶Qվ@b_'ߞ VSAc0t9qiQ[̀\~h1 ac.BJ-Zf:3ǼRwOtj_^-ܺ+ʼn1WEDQqj'Tfz5&mF0z5nTL2\0G 7T76 u>2{agM'O^K`M<V%C]8*H YX+R\.y Dg jZy D&߿$ዩAr3fi?ۇ>~FQ #A 6%ˮa Sec"5X7}Um&n5ڙ\Yu郊h3]tV2][4\9QTK][Vz 4NUf*H *prJoVz;ۡގ[%u{+\WuF`ة![k}];]]CЬ%AR?>I jkBMBPOj{Uz bc[h545ς1m#u '@(KənĒ6Вv ǗQz1hJEqZpZNe {w4V Yt6WG^Tq?MnHBZVP:eE5=Z eήbEˊP0/Iʧܴ|7eej^ZVNtJeSl ZVl^z3q)e%Xs.-i#tpc JV˒FPӦ/,iU z̜6(+3 2T侫D}C;iMR0+&}O6)6ךV  ᠩU+j>XUrm~P Ѵ>,`d=XD֣ǚ"{6 R*鬟YKX6mXkiPdG[Q!ؓ`Ⱥ`,XL1)~%+S8,H INgz_w _ R7^p\deJ=3urM@L#\CuXS)K 8* Rv QqKmJ\)ʡ]zeds:{sUʹd7" '숗LHwvaReN`D`]'Sa,v9՚9l\vhC+1ꮫMؾGmXXۊ6d~X6խE]0+CIԏ?jؤe Ƒ o7CLJ 7QkQ?yZbHM[+pj+^1pU )2OH)1gzɕ3W ߜHKVEs \&9c=gs6[=-Yn6ei -5Y/?CX2EX2`ml{J(\0X&c ECe؂|.avk~ R(czW`vsXth:*mci]r!TFu'bAEfy`btA2H _ƘC] $CY֘36tM1GJQ YZڜ@nC24Ja(P҄JajYp;\phY:&HBȄ7ktr5c$ V=.I/1XTK^>X ^)ddCA:'Qe:V5)S%82/Mje3Bt&5@ŗI-`gb \&s4&3+`F7%`u˪!deL%E6VTJ@Ie[)(pKuDYjȯR*4ee,(F4d{))PEӊR{! ].LJ-JVR)]5mۦ6oZVU ikcQjPn'qv^<Ž~0>zک\tJTz;ؿ2HB)׻o6V)*/ȞG@&wxsD !;R,*O^^jm7~6BWPR`K LtD,hQrA\I7 b_ ԥ(`d={^{Ym3'uc*yϝE&pRbL(4ty/ B"װͪNa)%G͓QSZ8~Zݱ M+N6TdxaߟNx!ji6橭NbS PT+Am62)@'RlȽK(ܻ:Wޥl lAJRl&9HɤF&9-,|=rR H ]9XI4NF2#cqjtF19o Q7ʽR40QjrFc/oo6'+/Ro <G=w#Ezu^ DM㴂 L@Y@B mIӚӝ"}t?oNZ ¸h FeTXjs&x 5@i v<%TYGSO>~:w'Q*cLp1ptbmD@P7X%+.&c)AYt.X'<w|1%/:lwɚAW>}*9.sJR}ۉaX͌:sϖzb edK멎nFduW=K'%륕tk7]+>jj k4!W}T˦6̢b&Eu[Dz(+}Q+Û4&)&eݤ1)>c)֎~DaƒD}Y a.m-Ike.+u]) ֆj1D*<mapUBgcqbz`LNPh wUKiHX4Ak;R&zgFkQ2K `vst: 3A?_Fm C'5FMD: >762亚JYZ'`hc9i CW84ƴX^jM2Xր0܆meiWfXl>ؼa{IYe M֤( %jCz8@)H-D`+waǤÌ1!mX'k“BT{0 l[?\(7eLSmdTNNS1tM MZ lP{ רRG/1Y pQIc)=+ti`R4<(Z<ˁq a$mRso)0]<#+( OV*>=|R%nz!% |[]ݷӷUW}n׏׏{p_\'_!]iʧ^jzo"7'fѿ"1&_+0$R$_+zޮu4_+*f͖ɻ7`iuB++憂k-*itx[dJ&3'cV Тvӫu` YKjȑz[[>z [4H[^j ;vZ)T ;>ubꤤ5鴿?F^bxǻ΅wcUGatA/y'$Cn:)a3b}`9q.!go7ud\@)@DK2\~l|sɲuޞJ׺L+UX*TfkO)``bŗRGytW޿*d`ؿȆ|_ǏA^ lRZmOA-lB,fi$ ),+ji mT(˶Bb7|0KR;NݒVRUZ)Z`XTui0jn Q{BPXf*’(ev<  di?n:MT='O_߳/{VXTOηܵ'NW~zmk\g&Kf÷oyvn?$,#?XMqKj 9ZqM9aVd.ʇc,qyi?׮vcl:QHlmEY(V uHPlWMa<}sJlN!Mp [M[sЪrTݫx~]"PDY.6qCJ? jN}ڮz՛!\ߗٲzxpSjUʢEM rcMTbŢ!CSkT!nTUآQkq7u}N?Lo ׾@fsf]"]MUHꓛ+W(Qb`̪ q+I-y[ ɶ\s~|ud;>a3G I z喼2^wq]O_q]Enil^jُr"~ >Ͽ\=aB 03>3L##H2/KcgO8QedbEiں1 +[ȽV9`4"ukb:؃lWjk ~&<2`{lqlROul{}4KB-{E|6zFB@eL_R"[cl"|ȚòHwcjr$kƦdij{h6ZFdC]`^ -?$ctϐIjp6b :t tanG87z}T$Mʦ맓c}D0ʟ>N!:YNQptM JٴNػFfgdf 3C* Nj ͣFmEjSCWj'5t<|MuR,7؄~zK-vaySw{3/}wu[y1i݇}ig4:K>r1#8>pVIyŏ>3 *]y5J5&80o"=o8zИ,`dDTA߀,6ݭ(Iϟ;EfcfUh !.JcpTuw:PZxj~e+ˎN6sDv>-g_MR;ޝpU[/ltbi0ڽd=:)*k^PabOx) |dHa"'DDјyqچG,HF]ZCJgw,H}/ y~^񹫀w|&^y({7+kU4qK rf^y ļK8T;׮5u\Ͽ__XREMi>Y,kDPT<hK`$ۖT+KYVXh%!`UUUi[(d 0Kj)ߍ%E`iW֦hg ,I/ Q Qk1-u,MQ1dޝ {7=u%DD<"{;Q` `\m]H4"ٻQzm=_T-!w3AlHrd1 ɫ[mFEiHa&}ub6 i<nAg M LwC&!wSl:) )Ȭ#*Í( ȉ3t-6GZlh=C.r+,Y^˦Z ) #ҬHhH3,hrkM: Sm^'\Ui3ҩ6:*H-2R! iH맓#ѐB%k3Eq-TkEhHa?*PRBt2g.UZ/̅t<ה$]Vp:Q4@s)@2ulTbvLYOְ4gZ+F2[4N AMLp/oOi>p4=+JvEIiki>ӋdJ U"R 0UB@mۢTU!hj'3B, -C-Vt&4$+RyUhhPʈK[ii"ƒ"~RWBXHC\"ut ,*EʹKvqٞhcfջvHCa(R Tj݈i1^`5ʊ\jlxh (~W-ȀwD#) 1lE/p "j¼XSQk YZUM 9adiZe{fX ِԆgWDa1c0E6CT4"!)l$"Sm|(RZJ H'o(*/XPE =t Љ*:`;$V&3'%9wZ٥6I2>Œbwn*_vsvR, N F_3 ɓ"-ݲ16e) 2q$wH5)|=hx Lԕ1LFGGaGs1(Iua\c@z!paefAN~ҷ|p(3KЂ2$ ˿a̐Vx[)vg(xM+xmU󢂊k!2w.6y yk-`p z}a}d\;J;ɖYQIcFax-eθђI֢7;5[7>|_gSϼrEtR<]nD׎KI ~TIq9/:ޏ>*s~^RxO0u6gSDfv?QQ,y ]%-S7;Ne"t^8)H;핌U(kv* 7fjiVkhT.Okv`NwBN$ƚW~P̰?ך1C'T fBE'HPyR }XI 1ݾv+E\r]EL7^)%u& 9 :U%R)HT̳K#dᥩ2o[b}>yYגSm,Uǫ8['V>޳'TҤK R)bq  :'I ^KG&`4;/:L<2 7R#HI##<]:0:g?.P %!uW AD;e)uJ;D\Cx:uߋdN\`%:{U8N:  S <'n'jbӉ/y(Ix:/:f)K|ᾷV5&:/*2d2EަeMJ 2Y\ZN+,l%]J%+]. VkfZd(0y b,ir /2֦*R֕ x^TsSU<3S(9YH+}, l+eV|A1_U_7|oAm/6~W)N*_koěE{IUѿo볛W 39oqϜ^_~?>hK½q3c;nϟ>m 2~{v/B)X~L{j8tX/dWL>޻r-?Twm~Yڙ]wF㐒+Fjн gałqw+faG֏M0B T Qp;#˜y^mywEo}~Ƿ>-)L1zei=A9aH}~pG8n8V cXJu`s>tTbp9Vy4z 7ϣKABۿ(%~ƈ($K+J:Aߴ+"\r-"8` "ª,5vSg)i[MͨƺKT(9=_ԏݛZ4}^ۗ{`:mfG0!f4F> @=: = y)8PklZULS (VcۋFOa:Buń0/Fo 2AV4~۔=1IiyϪ^V6nskmu9Zɚ:eUg?cYbjeu~([VS[ Hf+ Dv*U%#0IVT/9Kf¿ovJZԝ~ lFHNʶSgdVO!=zwь?5ō"짍z޻Nc];W?9뙺a+4+e|Mɼ+on]j[_ӟL:WJ9𺘓*yQAŵ {w}(#OV faRa썐 ]"cL@5Qq-2Wq[ _Zxx[rS:OBRn&ir-+va v;wۂ ߱cV" x7i&IX IJHD2IS#K1I 8E rHѴe<*V|-4%ԧƨ3 q6&5!j!pf{C5y,_'OqA<@Nf<dLITɌ(Ԡg#:.Gsɜ-/F5M1 u;uT9fZu~ͷu>O8@ {}H b cH^}w1ko͔h>kFLid*a X`g /V۪[Lmt&ϕd[::,W,WoO?{[v^0KbZ=vaR5Ui oˏ2m#"t4Tpy`]*PUB8Hy"4 h+<JN9#}Y(-F0B z'V@Ev -QKi + baJmc?ED [ۓ?M-XՁ uz_w-m߰\=ays_[b>Y#g^Ǹ E"6QLI\\/*茕fh:cܽ{8sFvթboqhɗ.ԋP\e,@+>e\O7eTKۋߺaLI,E]ϵTRͰ Yʥ0PLv;,RFwxDuU8xft^5Ԗ{$xr2&fӡ7_1{H̎yzr41;ZJ7MM60}SN )fҡMth8"ǩE&t6 j|&RrqGUГa寊GՀRs2;w_#Tg=.T΄ˣ)A #kGɉ^Ƽ8C8Qp C&pgRDjZL_n>4'I\o_$m7gy6=EGߝ\!GJ;nuļ.oz! l{ۃndny;`OT %[bJ~w(|FnjF[):u&sO?f9]\4j՞1k9HG҉^Dj̱l1('2ѓ0CRf&e'`,K)fW;91-y 3(2JG?Y+e6,XVZe]q^ kxWbh1dE%wY,0 Mid]~^0EW?}NKOH:XIpupڈј^A2[t$Pн:u)Ԍ/9HS_ S2]Q2D09? @ ђM'C9r(_ *о5#0g}r h"G|: \85-sSSppjN֥$nSQayxI9]6"r j0Jl ^jHbPP2bv8B+[%Gqi4)*QWTZ:T^S!0Mu]*&t:ᜋ4dm$kQB%cjdЫɂb" >9ՒQ/䗒5$QKn.>#tҚ^}cXI \ A1hrټteua/_OGWzD44rm^cĄ[/|lhyEs%p<9 anP$sNSy vJA1PFE6܎0⻇2h7D&[2߿_"9Y4ޤŎ.1|'Fϟ4Ji13VGDQBQ.͏әD-I 槸4ItrWB[DS\[#4y_/tG4OsG<,5=5ڄ(d$%ǕR;U Ubmٌ'˭ٛ7PC/y}n=k-Km b}儇L[j WnO|UC<:\ey|bKΓ6Mmy>FPsYh% 996OOa1KGT?$F&WH&)GJeaz/\;:| 0Vlܓߙ-"N aEsg$sq^I t2aǽ.I⾴K[Q%{^7JuQ!hS1ϼ0Fh'XZS:T8}`,Gޤ!Xo.5(J'Xe=}-_9s u] Yq/+@UyisV-4 1y-U- } n,0BZIALZ0A#)=I{KD& NɘP,m7q'E䁞|ŷ'wn.f]_0[9n/=RۤX]o;9LgHF66ŗj1j]?EL%;87fJ0=`ۧ?|͑xͤ?XOR>[Pnnl iw3ڔ t e(ĸc{MnCX 7^6%wAD ޭRLo9mUyMnCX 76$FBd~.eFg0MZ]tqnGUZS7UR^R宄$Erjvc)Q,%6*R*ދD '˨@o鎞7WMIzx*Yh#^ϯ=zUA9KuG˳J1lB!J0B2Oiӛ'*c:QMF*O<lW50 \M='f)`<=dcTN Vcw&&]6 ! 0ߵS s7T>O{H1<%?NI:Wro^3þp*U-I_Č(lov PD5nyпo≻"*}@@҉Y\3EʴNJ..|Xy˞@O7ɩkQzidUz^LX-GJPe>#brJʾѐ J)BĿ9P-ƚ@^ל%(t*1HFuZ '. #`MA ȿNiꛧ^("% ȿ]yIoXHk z[gRV RzbΨ}|EN)vO(Bu;rsq#;F(7\w\uxq{Nޝ4ez6{wre3nۓ7?ۻ7՛mobDm=:3jΧ5})o8yyuJЉ0^@>gDa_d[yC3Jq݃y5N[߿_"Y4qؑ%>.+IOd?͢q=yE۳ YVި ]Z75Np M9Jz-j'jS!3zG}:[lIqGQ'ܯdma`n|>tIS"&jޭH[5aWb@178 Xc@Ji53y1( OhR 1>wM>Ԋ 6senUffhtk3Ȩ, Aٜ^)eCN HN4:nK khH_8Rs(iXHz5T@[E&Wڭ83sF T놇DDp csD `ZnѾ" P͌80tm|Qжv.QjDNQw {Z+(ijj5N-:ÑP*tZ\A>[eښjKڔl(Gg͋bCmpM q%3Hnb ȉ 6,dN&!j;8rc^6A2Vx)~lŁǖ%hu3{ E]YnY e  "4 O i8nۨcTҌ7/N}@DŒ7pı#6i"Hq*tb!(ho^j*`|_%ځo.ΣHuV[űX sҋXH^9)TPGO{-Ԩ u>FJ w^40}>HBw⃓G~WΝQ64tS^ݙF3A"t@iv2sR67͟z7r$|tB4u L"<>.2 c`d _}HC@B[͘81V9FKY#(eA42I(ȨOv~r:8z9Yjaw4 9u\XxQ Մ8coKxut}~`yR)mx@agH V0 gΉBE$|4)q Ӄ: 1, PID`a9)^VF 4 xSFmAͨ:iW6yE؊ff4ՉNӐ΋u:f%[DS$#`g"3PW'@?enњ׵ۦnx'[fƿ5joĢmm̻jy6?.sWv ^6iM+}$SL^KX3U]ɢ(˚sօeK)ِXV7onZJ]7&3zA]pP_s ѯ>g+d-T6Mz3v0eC?2uRSU º,2%’YW*ERQ=XҭUXrhbI2iY[e5%T@Fe=<@ Ӳ(Ek%hEʲ+{)[kEfd%V"+J]I F*XTZ5p1%Z-eUH%pqEEf,ڹ;͝k6) vJ3gXp{~hümF5uc3伿Ί+E[9ǻl}ՌVunZvۏ=XWpomy}}jz䣧6l-+bɪ,EEQ}7{RvJRJ!D7CD0h>֙s6BuXiJTy[jT"IwTd#JtZd#,HN 37W2ĴH6XJxX3l%ir!hV"E gfJ6(nx̼0?緆R1 R-x~J8-zBTRq;7ܮ_]!JF - Bd,!Nj$&6"qj!:icP|pϝyÁ @Uu)FTEə8vVóIqz.㧴Y[lta&{<$Xb0ۃ0U!L/Sg'@{*e`u!Q`wf> ՝hꄓ/%CJC*(qCeitԉI %Q3넞:J/q8:˄I6%j޹8QX 7zK05J9AꀤH󢦿JRT8oa~Tۉ)RoEj;%Uvڣ5zn:؉hyJhrjT٫T0dҳr{"c`06E`MTYlE%aPUƅ2JejFS׶ʥ1Ui,j3'$ KWcI4XiU-kxme1nI5̕օpoFiiPg֮3%= &7NeDM &s5]:}Jzes’]*C,G]B LT:"[a'ECL " Zha"Ov5Vf$f6n&`5NH&g4a~?o =p<:MZn7„i7&@E@6M<]N(c8$aēmd:MN-ڔtDtHM{tkd:N,ZBWl0Y$ׄEѩ5QWoPI0ꔶP-ỠKJj%: @0ސRC玃xaԡNh5x4Љ:h S@BD(x %zP-9ВDUvǹj99$ЬT# w"kc$L%0&CbY. 4,ղ/ RT:(kX'yC{-jiQYdw⼢YjD+e/| @'Nǹ#Ӊ!]Nltќ6K㩳 y0ٕ@c!!ũa1}v? 5{atӢ@.tPPJĩW/d"DH6@ˉ\RLrᷪX,A;3'T fb@ySleQyj47<Gxyá;+/E)O[B4쾸ڒ%WCzx#miД|S][?[{le/~m78f@溬sI% EFEaVygVThR̙eRZ6ŻR*-~͢I^[]Ӏkkp~sqUu>C9:٩)سȼL;lI};P85QQBm JBiK`TTUyBͅe~,kEOH)Aԇ2iВC;K .ə\]V|I9_Wwm'Ӌ`[[?{[lKdujwsM?UMC{F^ .Լs?rw6'WҷlMb||g|S}ekR^\m7({?tk}/`%kGӶVT +:(ChX^k7{thkupڥa=+[]kRɣʇTQv.ƠaV1Ea -2i!m×C, A6LѮhZ2*ElN޽wfXMfU5ös&'* +la ѰԲaA 4xMF RAb<V˛y:1 +DQX4? +lҰBtR`Si +bаdLc aE]ҰE+ WHCt T)-`@z ҙrV>Z͗I >5/tdRhuCc2T:;efMJCX&K?L׏§TNtϿ2|p o/l>~̯Փ!6U2F -xcDP<F=l',6Dd7[n s÷ l9ęvIS"HB"X*KJedʊ 78h!&넧|>dnԐ4K% 9"V7^p%ZCG&cL7/z<.QR?"\RD"5RLv~\/8 P =Cd`;_T{aTW C ;OUl79I`\ya'"$x:M=h<11쩎FU*^lRG"+J]I׎:JR07eLeVKYR %C\Q,$0IM2כnj.EvɛT̟noO _E1{ђgjM_\gŕ"o[׏wYn{M]/ԵXY*l?`{_6ڽYi i7wjU2KXrjfLCLi|JiElb6k3ֶɡ.kUX5*+ r[*YaBe Υm0U?Y~,JoH,k^Hy!qMa (7kpCDQRڪE6 σv?)kMlBw}{] Ew5ڀ7 8ySBb%!ݵ5ӾeH48ٞ+~˫w僇 Hq.k1rWHͿקq]}";ETt"3Kk1EJMUUo"y[HG̰'It/.ޫtaN%iVB)EN Jo$E2DbaP~k-)00_`e<[|`6ϛY !]Dot*/XE1" 8*D6Il DBjl$TNmW2^D%$k _gŒ`(>(t)=LIm<(Ĝ[e|tWɌ42)kp u))+$ZK.,|afkakQ LS+s+R+N=fkR1A0 *; Cmp{#y5tNp3Mon{"ҶզXJ#Ǐ+G0& =dEt.uen.ݛU>8%L{*{AY),j 3Zftc Ed]P'F@YɉW?F}Eo71/= (f];7o?>ny&v^Au _W?xfkMt֘]=.ٵBSfȮc%nTQ84:1'!Ug~~ki9tR/iI;{i[h_W> &(yӬ:/J3W GըUy4ώn_G=~_>ځMOOM+uoicEڴєA=KH%`X@fUy842؞JD+r4n<2f:.|xYYa$%vVR%B)̊YIՃMmM,߾[Vֺ6a %W[Մn`L.׶|r߽z5K";S+tuρ9.d3KUfFRuYOOIդTK`ܰH|RUC0P,I?)'sNslNz/087%v{X, #g=_`u]ۭMwb4釄?oSCŸp4֌A L ;óet!tn# 7l4A.8[s/ɥ/BRTQ:)2`6jM ۈ;2X΃Rp; $}e֜ptY5YiʞɂYz:i/'؊j*lҊFL%"zuL7;e wp÷*C@^\ҁ,ukY*YNsOweՏ,yZʊ/ɋ{0Gr B?s%$8${JփT?Jn8Ptw%Wވ.V)YBsC$¥m^hB{N.RLqyOu,hhtXIt.ݐwA\qn?>Or'_/ϻ7V*"~ ܻS%2/cbZ k_YA1B6Z5 #k14#UD/ C_4`:(%{֝C*dfJH1Ɩps4l(mFf^tR&9y`4&Zᎆubֹd]b`Z+:/@Osx#yRw=uO S<T C|{B7*r ? < Dz!( PxH$ =XW?RͬUD콧͇qOg1vqCwtxJNzˢցY*)F1MS`OϬ{wT۷ dm?Mm) =o>|Om& oӽWכXRA uYPϧzF u1}zv1)_QP,SmlqN8=P O)p%YdPl=03q1:#PT&)80&И)J.zZ/.$ F,9V(/ŸLgy!}#} :qm-|ñe[Yx*d~p5C;gcb2w w9DP_(vS(`CڶŐZp5lIw}A3Ui81t#+c(R`=$mTH3H]xEs[. $X͌ut:غ L*uߜpŕAsqHtBT5&_Q1p"AxNH0v(/;́.ix*UA/9A+ jg q~P'r$U޹1gg\/P,MɌ@c>[L5jMe[xK;1ϬH|TaӬvEGFe.u}5zpnkne4%0NҐ)&K@ ҐRkk*oP>Vnl0t5\UJwX%gj4jFbѐLxVa$)1@óqtRhg =U7HEg!NiHۧgW#ѐ .P:e) i6F!餍VsauRlM@C* 9`M!׾F!DӐf')|:DzE73:.=0S43pg & 'as9p042d(|@_??_6< ^J *[Ϧ ;{j(^o9TC@3mP #wVm YK} ͖/1FqWQXQg.W&YYgV5ڝ=NFYq:-qЦٕ,3LFUtBQjX]@ekOK .IwCι&ԮEHory\f3A.I[P;CkPycB<%ڪCgP.<B/i:% a;%徭@LGk]kVX612*cF+kW#wxޒy[Tf1Ԛ%`u?dC5y0s>V <ɨhtB7E)tbrFN47٤hy( zdk87iB$UpO!:BheEElIA0v;ēZ4 GhHd!oB|(EzHmf_k+CeP$_j('יIR oCHYkwWtTL͉](1>UnB( )'}0[)(RYH5 Ej %w^.oT>Z Tɶ|f t9HCTjtE;?\_ 4hҷF0<5rNޘAmU壭Sm|UZ5Hl(U[ ^nT!!'T"nzwBQOU>VrnCI˪@=m/M*0XgݽZOLVhdS64SUZA*rp*xuVrwsL:Bi$*;sbhĕ'O)*di f,Txdi d [m_M|XXGm JΆ #ioZ;-ZF16BFKlgI!*oLS2Ֆɝ,޽docd]_ŻSCUt3{u9XRoQU?xGOxǻKoW)]w¦<у?M 8޽{MiSۆu{'umB=}VAV`hXf8(FzDt`Fj0jgɧ!6A|#7ڊ.Q7uP5Frou{y=lmRF !HmR$0guD{", ZAw,[DkvW N11NHap>7>90q)`8= \ؤ"cyEmoǿ@o6A2>TI~=pþq #%9y|#{rk4c{iEZ'OK]YRYnpuEU6/ˮj 髳{µ P*!ZVcVh{qc.{0€-(, B$*.VmUo)䬀0xXVUP@3j21)_p̋ap(H| Lv0p8 njSjcOZfYu~P X KgOU#LNJяm^Qq 9u1D9q'nZ-D>a9r ݾ䲁 ѯ2:u ]nt<=^/Ӌ[2Nc\:װgXSh K%tT+y:͘Q tm>ۖNim,В,~21 e ilNie4dG2dJ/^<'()qQ}Mz& wI@{7|N1_&9LLq48uN|?l\aGkЈR7Ňho5]~L1XwV 1֥Vfı:{5 dhHK{G3(G*9?0")pMQ ́3A, Hiώ;&2%\;[Sk]j+` riuU%tHG*eO =2zusa=-q@ћQ 8xZ:G f>Ǐ]Ϲ:')yd(ц?!V7Km1o8p[LhN{\ЀМ^ +ĝ;X;QƝ;Oj DXtJF KwxnԱX[cXޗ86,9:u8f&#?́ xR ݑtPuFeIfy!$_mL~̿QX:R) tkIZ45hTY]V!m|gK% 8Didmjrm#TEDIRBRD%4J*㜴mEhēIZ  z3jٵ//CNMwEq%EH\7wmO// ?·/vPy;߾{ŏwM}\EKp'xh?ܭ﷢ǗO%g!ěTڂbo? gEV o/Ii0+ Vߋ>qKI[ +"9_-{1|èf7 Y.P@vS ֕]4sa 02S&ɘ4dD(3l0ncq+1Ô.Kˑ3v/KW 6j(Gs?R.WRYndZ!EU>_'nND=RjiūLtMϴbELGC>p!mN~3!1& 31-<!Ҙ:?&)VSÑ8:9j^`c 6!ZC>0!fN0c:e ZPʵcL 45|N04Os.ީ7:uyr+P؁+cuz0 0 fd;PTi щCp4(;-vqc2lh'3BA 9]C J̇q֐IBx*hG[΃:F>w"#-uOll٤PU|.0U̅4%vzYV3c^yWnQJ<)_9)Y#1GT_-b0֬v@_ T3L6RPTNԕ Q.a uJ@r T- ZMw3\~{чv-Bӻ;__!?ϟOE[aWw_tHf*뢺[>HHǻb}kQ}mC++.ovbǠ] l?{_UڽYVK*ԇ6!PVڦX6ƷZ:e@#R/;R{">rMbԦa=D:$ɪUHafIij#[Q׶j*M]% BI](*YRd*EUҔ5NI5Vt# h'Pfqm}{e«:qQPK-ŵz}/%@؍m*yHIX\Qq>:dq\@ƝjPz{݄<t .Wr#lW6GE_>dOۥЭ%p$Z>7ڒ/Nb,MY\TRұY\ߌX\EFEQ&)jCX:Y%U"ӀN#08sm+ae`x(ɁdIHϛ/D _|9S&\&9/Y\ "/Ӱ6?}^aok\۪7Ň.-13=S}!E:jOd ":{ܔ*}4'SQ(MGS( Uv3hA f@qIdt0ɩq{̘7f2 o% n%X4%(~,)FpݗzX#}(ц=-t pt26ktaJ!qҊ(H+Wdԛ=R DԛzJ:Xr'Q _)1(8_{;@l>sީ.^Թb6bf~+##Γ"Ӕ,9_wLZ3_+?=P| p;23.oĢݞw(j?.K_ ^ZzjeZ}.6-6P9]-[v-5`tAM5BTiBvk(/oĿ,6M(j1Co!3z]H8h }UhGPN*M+uM~LS]1'*X`X5TBW-mKʏJ;+ PI>uOc< {3y`Y[e 0t@<,w²(J)C_ +cѲ}/UF=eȊdͦCˊjVӽGǢô, wĢcZ=nmW$wSp<{$Za,dm5e5FeZM?a 5?Cn|ND yҭN#@6D At%^H,͡,q8j LGFltmQ 5|M<\fLEˊWH!ɬ]2iYT . 3iY$;-!ڜd@Jؾe1t"&>=yáӌ[^;=?abnxrxn:tα0UwT'8LO⮽JМܸdLD^J"8{cZe_c ֺOq `QMc&$ă8"hˣE0E@fKݴ zv6`1ں7N ={xF m4T/-K|0pHmkWa\wt֫c%рsKE: UZN`$/׮ۻOKQƋN ʘa1Ι8ϲ}eE;>@  iKwJr+,V4oݍ*uP9jTYGG# Lbф11hy4蕯KA 0NH")h!Al6ȇ391Ư"h].7Z6䝣ks.SF-eas)զ윮a9}[/؜9)prlOE2U"QUTURjs;N2B4wxxNM>_mN޾1!gUJ'J ٬wp$aZNe,*5PIWsh&g$]@;Z3'z|a6=;?KJ/3=ykJi{1=l?}^aokTm՛vMK?mwq@cFN f_|a 9041070 [+VXHѺɞ!ı6b:0  ቀs )rA,, .7fD1H 97 eUr`!1v |4OwMvd̗DM[MIK(Zڶ.M)J3Gt.mO|bhfձ'؉jء)^%Հ.'$3F=ƹG5R'JށH[(~4J80R8\ 3C^b3b1f2aO cCK|vb':,QH\m/oVS< u1ҹYH9:IAT sŰif0jH%Gf!Q·\w % c&;Eq_TE^_oCŇ { Z#/IzfZK/C- 2~-ݾܷKl,ijW+0UmC , Yt)uXxȖFmC%-~"$|Y/@ gM5_ᄍjzC=ル\jzC-)Ur!S)ԵV(P\Ua=$k~{я0:۵p׾I3^?~tty{>T衤>`⺨o=r ] E Qm^] l?{_YVK*ċ6]U6Ų.}zY5HTAm|}5Rb>7ERZ)')?p*m.чhyKeq(*-X,o/JaFTǓ=96nS83mD7ou>Dt{x 1-"<ٕcN2n➗%48#wCFf}(`[Tz܅Dn<<՘6yh .C{Z1DS-ڐ3IncMh|̑$4v=n1v2$ͰTA+@_I6%k>PnC;FRNǡbxttɰ7R2NOtc,#<-jjw.MnF*؎0i$HL\<+Fw>bV&E(P9t7kX?825ێ;5f#cu8?GU5he4hhyZ٭zh qnXqbhe\Á^VvQ I+[WHcܕ(4Oſ,RF]-j}ie) ƈ)vPPTdluYk+,ֈ]BwZ(lFƾ|pl%h՟ay@߻3\IXxVKXIIٷCLJHn2,@Ptuɲbm J]UIN%ZW/vJ CǤ8OJ/=993G\3,o!B> @ DtE>?"VY5ƁcUIO|)Rw8q+Wjs`a,+?`M޹^eOfkM>0bT1|Lk3k26v|͌(Wl3t3 s`<%l:pr4Kb ˯r7q ޤkSo6Ѝ [svrDyU;1K2qߣb,P ftigjfLHQԂ7&"8 tMKpIXٜ3dPiAvq:Y9Q t# ъta2rԏq)r1X9=PN*=,vr5PAvnyfv]ꇏawowۖ1^r /e즰!{M!p.)K <30T^gN [QE`m֥a,!JPay(Tؾ /UԭHr`-$*Eէe,+62oaSAeєJ+;C^1Hf5 ,HW&rМ|x{t$֙wp‘Ar 85ye:!]-HagL03"cW$pw 9*liRT:Diun{6mx)|膷Tlv*0|@ULFZDr@smw^Zv㱓Zύ [WHcڕ/"=2qrQgal InM-Vi?Hc-,2LLY$/XϨ=,om~iu/dWK Э"X&x~f؜ pz~fAzpc:զ˩y %jZ0^N.%aB"EQNѠU \,n__ #Wf3`b:auٜbYndFA}4 (L86kOn#{Y)f[8 `$,r tX,/@o7#HI,b3XLݟ uxUQK`Iwx |(S4FďZri^M{ـNVIƴ b^8$Ӱc%h֘^:d s ( pe0j%b$H3\I :f T"fĤs̭ӡNkM.[x%a~uWǿ3z(Űfs\uX|r@S&@1…f7>Z[묙kcʓQtUŖ9ay߻2N -"Ii{i! 34ȈReYb6%-T-5ٶҨZẅw%0DeP+P7*(JhJjV"F5 mrNլGye]O=[ fvev#VN̮'Q1]Zb(D$,ލ'ԁ@}GzKڸJw )Uf"]vJ0vFuۿ%?dQK9) tax$) h2nP$*~5ҵ3rfr =[AgE@kΆ4t)2Gj¢ZoÎ\:N\͡PA'ĩ,cX4JLH;bw: 1:͘A(e CJ'0H.md'XSBvt"M3"U eZN #.C>81N/9Sx NeY!=$)voVX77/ ilaN7 c&Nn  璛(TR}~gH$ރeTdB8n(F^0Tm#AeQ$',so7#˗$AM8 *`gJݟoE5uCceUQl6*hȣk;vB^14}d-sTWDN\NA Rkٷ/ ([-H%M ] &u#nE-1|0H33p~؎uDګչmnyJlub6:Ls o{r׭ (r.IP4 YɇCF{UP^uSsH7R`kN7VQ/]yk3/@yK D؛|H!EpJUrYtMJo-SY-Z( )҂aU֪*ẍsx1TjYd+aEmT騩uUhtiI'd!= ["Y&"H1FYlH;9t &tKIu|EnIa3 )ufC\j;C(}Q$CIV\`aLI E:D l\Z(X;M.c\lpja2 AfzR~ȇ%5l4G`2*=9#oNLƘtaͩAĩ8?5bhJhDCB(1":͘ (&Y/ImIԖ(̒y)Hqp RB $@Qt M+c-/FE{*@TjƬ@N]w%Ys:))~~6'0h-vV(<13GDBߒ+>\~9RaG.jn Ӻf. a-ֵWØ2H\U!EacaLJwl7{߾}?={] s'ov%SW}:?[H~y(tۊ-wkۮuǶWĝO/{'zWN~>{}>b4T֎A ZK[um2mCm+PUu *J[kc]< ?'U[@[p]45XE帀FOU٪1Ee ԥoK5[7Ɠ[BFQpf|](} c3[B*}?RmAoTHz˄w c+Yf<2\]|!k=:`@?krK-0DgX0N)&R2FHYh7l?Wt|&9\+Ȓ3r?KW G7f2G9qtit{#*?eZfu]As3T4V%dvEUfW2&gvEweBf1ٵ:͘(81:^4ٕL]1:DXPQfW:&evhsi+b30bds C.ġhC` ^И) [zh`E.ꀋi:aKy|`vN@S8a>kL,}7!+%dޥ5Y=~ygH((oShRT\B8o(BX(,z*ƣr(^bIv3h˂c4 2pQ9Hݟ's$s< tSoSMmklXk5ty M~ـN֩|!CA.u*]AB ;zUѰw)"Ay'\?T$VܩȊd̖ "^8F}PR v֩q vӽV`('*:1݉"&jmB)?:|RNJZ:t9ֶWEQP[9M(5ĀKH@* Tj%d3ݕbK_M;Xߞ{3_w?]u#_7/Iz?dZKZzk7u6@5*oRh?*eN7TSiK qԕڕw:炙At%}?;>t gmBw4Q|[~*mMIASWCJHxbmRebK]wW%Ej+ǪЭ4-?{Fef":/{ Pg= eDPh[b?L~ [tZەc4?smɡ7v V#=kƄlo^n}ӌ9s`*=c7OlA' j=nIxǏ'늱0IR;+qridE;3u$WJ(L D,IB;n~FTdm-HKyV-2]{k:`o5}_YVs܃m.tlq^4@D^F:K f;/j }@0Ca6FX'M,۶- BjcT6n6=ƛ.d0IK=]/NxS%zee׷Бb۰[zg#p@o]QX)CL& R$} ?)~s՟Jo1}ڷ|_ۘ:5Dn7ByрǗ=y_G?ٴwAocx3ar7d&E޼p\uTW#@L.Wzr6*f>$,-qz$=\b]a;  BT<ǥS%u9S%bVUmȫaAW9QȘd 3y:YN֠Ndz`X:y%3WOT͘dmÜLxg9wf`xòμA|:P;stBmdM:MbSl ql{łBGXuE:Gcqlb0Lr0<{-V|\Ga7vm`o)SZTؗxM- WgP6~i^,q.h/ˁg`PfnHrp8e@1((*vٷ`VWgPLX2hB\/z(?=0+gP[aQU+7fU1s,uq{ƞ؀Y0䨗$Qމ3S$pL$NΪ%6p^ɢ`MZGy93qJ+I_H -z5T1%7|m+LIŷ lMh'*^&JaL!u03^%BzAzme\hgP^zEJ@= kWŒQ50(VKZ̠AH0UG+f{v|ޔo8usoFn$b-0EM6Vd$) 3 5b[RJAMuD:"x'$ +~O:IY=ᚾk[26 E玡Şض"F@jƘ ʀG,F $s 9 -rŤŠjyIE[6|(#-on:1s4d[6l-TZojt@\;xinT&k'},[) vBf+9&s@"M5Y&N0JT0nFj]es5j$MJզr)jIܚ4^5,mticIgehB]#^TBrx2ftq1&CQBuEtJ9E$h_GDi^DQN7UО&oI&G6c)=<^[ 3r`/}St2rI;r uZ?HG?S4?C̳i)-oR7شQx+-ț&aN:nXy 3f%ȂN-a.zzxK_@zk <<`dꩊrHigD=UQ$5ھ}P 0kr րR׀z2Z QdkFgtl doBǍ Eu6mS hj8\7:#Eck%vƪƓu\y6Z8N)B;ox^i sSqt2(#Q1+؉wX_r6dU*<*AT뢳ZOZ)X B#է7M(I?d~ZKb-.һa-}O}h7FlQUMס7-xcۆB qXO7qN F7S_Mχu6t{B;o>/ۛw~GWi S+d,T28!hDX{9Pܓ #;냯o@yI/H)9dDxrj|IIJiX=FÒ; !8Jy:+j!X;IP:U Zbb`=ʼs0췖Jd-I&|~$+ζocO*XūI A^y60& "&XI"kWENTA*F[a__>Tt3eֱF[0FwTV;ĠN9C[_CU%mP6 BAF9p+iL:pyV$l/ULX9j*4&#OWs!%4RȞ4SL%=9$]Y%bjn"VN,\9%edd\2:?$d9EX='BW;;٢!^6nkQ_h^ 9#hQSmi!8.?}65GӲꇏa7>oK˧7L)Kz/0һ8,ý_,q",*+d9`QyKDD*ʁXtXTEXԾL(,*`],*hÅ7f&kբ ֪-ic%"A'<[ujM j8RzYDid'9/uU;z^!E% 5y Ky"oIJ;~8LIF͓%59ĄF'Ҿ&OShkMN}: OPbQ ͵WZr<`JzJ)eΣ]PVRP[tb1V2)++X,׉`b0[Zzk]wZ:$e1vi[Ji=YԈ뉼鼳PZNhZqɤ/Dʚaقb!TU9M%J"e6FR'zTzkk zjFTַsDeIa<9FN$Y!ʤGߞ eP9@++feEQl%a -IV2?R!nGW$Z5(0s0t*RzF:fVj,z/ +յ*4~[g˰Vk%U/Cm߯v*{Q!t)1g|ۻOO1|?E7TҐ3N{׾ 0x$lE__54}L1ǾܮE_~d~}>؊^m:"Aѷg~voR:ݟ>W/߱rqKԆFX,L65bfsLSLaPJZyle#E2ݎLz.tl~}3uAI)~ 5„`j%l=qUM]|Ik Uz؈bfkt9j a:vyFZ6X>-mLHPIz8-KE$ny?VNٙnYjYm|LHf.q3fb(jeSytT9Ne QM:MbSl^l,Idl]rt҄Z/DyN šLspz@5~gu ayXY֢`Z\7rjQ_h3;\blyi{gSaLe0'Y\fzaH+Iuק"m^T 8,B4IuuCLHdA-|nÀ\-h4VMWl@*>?=0-T1N#AgӆM q4} z;A`)cϤLlAX/.M\祎Z[L%u?p΂`$ڸg%0mX $TX9.IH'mQJ] I粳 BYvM&du}tk^ZFuI@Pô]#hQXUC1ɜ0$[;)ꁵ]^`׆]rd@`8"dxeXeύ+°|:tU^&Cn֋Uj4U79#n vIG{g;+k% -4ٶu`^02<)apozTC0IB Msfv P@  Ey)_PQ['"k2΂ L+%qeQun P$5VlX|k"xa %)LyC$m'0y dLGC^bkN2tp~^ٱ3SǤNC130I#!x.c]wlJY6'Ȕ/Q̉bKQuYn0[^:%F.K`5MOCz*(.0eت" Y4k0rR47tCꐻ?I+%5W*bYdpȴMwi>mӇctDmkʬ`,g*5j+4f:,* Q0B0j,Lx4ireRZB3#IѭJKLF`K'!mm{dKtt@ I%LCXى}pd mhX IxFKƳ!Z~WWڻ߳Xtbَ4wBq $HsSB*c\QiHGݔ"k1* wIJ xk-~_n8(ZuLccuOh $v$iưֶ. jY~&#e1f-5jV_۹͉^MWʫCU[Y'P#W6@Q弤j3(eVV_};l__=o_ffny}՜}n}b}Ǧ =~lycK}lb}!_ uaᙧ4v~Re 2Wug`ѹ6Ućsx#Dhnt$e+˽7+*"DM 7ݫ!$LX=I\!V\FTE6bF"*^9Jk6\yNg 섮3S[רYƊMDOuۓmd ZT ;r^L%Q#`祓T&a̛_7 8,y^ (by_:W'2XaH cP<=ŵ۪ KL8+YQ  -{hV 9wl8(@["A#]|*$h7OĂ1־x2pkBTPZMDzk av 0ȀgQhZ+_ kG2o"#gٟ6UӐjj}F7|1+n/E6l}ƫٮv.fn8~dk;5IU|ɂ}u ͼT~ܭo\{j߯yHsu17X󢂊|/ykJa!9j̞{Ϙ'1{a&u x4 eD'.*J w|Pdۊ[dĝ[d/)aVȣa&Lhr!E2IJj {]w~4i1N{%ha"v~oh m~k-xZu.wtWExPA#8zc6OGWx1(j8 Q9`dAZ\ơutqfSI&`+]jDF~Cٛ<߫mv1kC[w#!n f0`Y#\G\ 5gl*ih3m&~-[`Eܺ{eQ%ì}q޺|d`?0DzR%ӨJ&Q%d3_%A *4!NN (Q,YdZ0gI ),15z^C&%2pcAd-W]OAꠁ`~z Rp^)!y vz5xKx}iJ%W祓֘0ϣ 8z/NOo۵q`>/~ܣ:/w6 bxz"^1'2[yo}*umrڪҾ:u"*WtTɋߴ(FGuDwAZf/iȆW Q%[ҵҥkKJVڦJҲ*(@˚J,2a@6Fvh}?ԋ*{C{6iJC/S읚[CM<vLWu~^W׿\jhۋqBK+T2!ZJn&+u7n._R6KJcu6/4X?]ڄQݬtW! _?"SV4ϩ4v<*nT z~ \|7ID0>EnXm:ic",̰h.f`DQQͽ$' 6߆BCJRdM H@쬥hrcrƄX46)qǻ%hZ<"<oA-^Dž( kx UKuÇNDvwG4n3Fxwws!ޛa\H۰u?;lBCu"A&"4 .L'$ogO$ ("hp6nԨh3NDJŇOFJF5xhV)'_W`܈RknvknvޯmfJ(.IB1VhUjgd6Q*UeY.Jw |J=4_R}*xڻ!Q17xɶIwz}łqMLǃ,W#?׈op`6T#ZR7p1$ ›~ӣ\`s4B" Og3*3׵`y^8M 1 QjoNc=;AYE9/4 4d A8F%z'tV~Jw^bknoKRQ/StF &|3`ĀC}ӍUKten鮒2[HHH |8Qq/::o9DZ 55M{!1Ηi FO ~>G*ŗZDJp u$hFwA(+Ib8 ]љPB*J(d iTsYg39# 𽴽A(=|SPWiΣ#fa 1iC̦X[ rc(;`s7^lG)0C!g>ƼG^ &Tn@>~+u5Z'lnYfٞlډ]HU(EYyEFVU2#IPU+YA0k Mމ}?+;7iRrɬvÇ*scMg7nف|ﯮ?O_6uz,=+/`ã7ywb|& /v14`*+AAU+nm2%:G ;r¶3L8{5{YS =Z۝/yz˻]#}8Pߕкc}.Ғ! sySH WzVdfepd4wG(Nmⲵk tFBm>/ #O/b{}/9_zW:J괰'Y34Ë @ϧ:wޡ=s b ,]X0~/^2DRݛL/P8~ap0^ض.ʷv9"k\qE,V V]ڶ++; k$ueP_3N;#kDGu,OOu7eѨi+QjmPSO{[ETw'>=}tm/ƞhT D͵,Ovd7bqv}~ή!axk~1@#1mQ}.d :@@m!m )k_KQז-d:5_O@ca(G-8RZZ蟂#hy {PH\>(Ϋ|`q"i>]?ʰᆲy9c}Ys@_ywLāWqwcRxW}KFKDNRCT*^4g&WslkS1#j`AFvR{S{:DFR*eSAbK)]6^j tK M|щ$ Qo=Qu1)7Z\F5&2u"n'5#7q6bP xuL@ؒ+0A)nmq%S9G[Y.`,jIrbID9t {I ic查#u٦**Ǥԑ*pfҰR|J+}LJ-g+=q+y2R܆wc\~]M 'W6t+oDT I׻|6'"6gz1?U G \2;Ioˍߎ'tQ>4h۹zՋăE;҂U`9 ϯ_,ܧ`Ԓ 9){Segs`ˀw+6: y G$J`lUHIg>FޱPLq2x4$4G+ li[)Z6uCˤ;C!- ]RC[ʽB^k(09r >8>\w`L 1(~ vQ]$64eD#JrhRj&Pߚ!Oئ D_"9QIbAb !KSCH Z"6%1 . zӔ,!0~jҡ$EMrg,H+|EŜЖ|>"r^<&C1t©s6nн`N6/͜H.t,)7uh:Zwnh:M4k>q+uk>]otq/@t5kL9ɖz 0@%I!{KDj^j^l>ѐ&h0$eɖ{qjh#\$uADa瓁kgpmdZ)5%dH|frIUAjAx1l6x2j#Bب;f9R7@}4Sj0 n,(- EU┃73醻FKwB`_"otdJ:W!7sSWy[U2kCǰ=-1 %E,CO ՀLESo3l"L=Z- 5غe `E7+w0c~Ӏn8WΖt&l3X1zٲYG64{hYSXuLqPbhex1>||u)uY#"zLJ9:\(<+ B aAgJHXݚ[ѹ2CGј°*ZB!tG;3$T'h-oy"\xNz^бUާryݗg n,l٠+Q@آnK`)s )GD^]_ y,RLyœaZ4 ScC㥻0A^l A0Do4 x; F/sL@) gة+& 3jbL-+JR+Qu%b{bʚ>a zh` { a= VZn2" &g/saNJN (9Ot#TŞU"QU`+Vc-ud{۟PO^Y!aRle^uB)2 Yrȏ" nXz0#^q ~mmNŏt]/K{l6P^^_] Ou~4,CRVݍeb 5\sRbɐנ-sAߊՏr5y}dْ SVBlχ\08' > = LN1Z;Ti |~*3WoU;A(æ[N)FRo/6JxJ CI,}<`1IYbO/o?}gEApV6.O3Ozɖu <>aihV[)s`-S;d|%4Rb&I )+0!sQ) j`Kͩ_rH[͊Bg┤nZ2pfN|c c LV}|YS m HԂ͐)s}OzP#0M>כK+%]] l٪M:{s%A܎҂TW&bќ|:J0܋7 F}#/_zU`7oNq%$T kΌBA; =qoޞ iLAǯU#I!8RvKS_y'o -V!{!ȡa%Etf/ۀc[C:?>16Gfh?`C+j56r}FΥ0z9"\|MK+7BmE?v0~Ǜxrouֿ~p{~nKUp /K(I5K|Dd(drm+D`9{?ףT0%#t_O`@·~*NP3]]΢|/뺵|9$qhOl%\TZ[$f5gwIozjdf?+eVLMS|nY 2F(j,h mbV2d< [?e0 iK`"`xHߋgP& /wOGW9?kԸy(Zno Y4>0]/ѫx(8z㓋yD+۫ۻ#]7[vX)vdFՅ[r|3׎&Z>ʑ,D2ޗMDt~{Ţ 㿬?lcэgVoM8 i-{oq'o]~e}}jP@{M- zKJ=J_^]_\]_[%E: \TJ5 GL]Oѷ);,@kYy \ ջ[^z>Z]U+,Y_O'"-=̏g/@^eVH,PPsVzRԘj]oѳWsɳ"\&Mk۠tph={;hwVr7ȡ;rL|'IO73q˾}`L9:,-3S&S15kꋕ~K۷ \g{PUH-p`EV7[zBr} L@2t txxXOC {}_ k,mOTV\Vv7[F GLᇪa A#n NkCTW+qі`ZF{@-HԨOKXbU-ث1P;<u;#͠@~*fH9·|2W5QF4H8:=i%Ia~&8RDVf:M$*/f8~sfBN /"yto=,YBdP3`$ȥO{"y3<;cC 4؛iڪ @f%,xZ12x1U-;c'p7u8S5Й-Dk, S9ɀs6A?l Rs*$8:'ʼX7$ƚ¬~l44D߿d;w{OY*f@IZ lgOFU|/Fgϋk_o6We,PfgNu>vAR)ŋ\O|\2VS]eB ]IW%kuc{<ቅ䤕EI˭/"̹9Ǹ`,*CZQJUh=K ErQʓ+2{_IW] Y1aoGVNu1Ds^41YYzTw4^=ﳆ3;gr⁽ 6(A 8Qy4!&N/vCvѷ tN|zU4(ܟ̠ct{G֠6v0Ӻ:.s]mNI-*JRW-R@QR )߬bwQ`؇c،G? Z0>Li /YCys'w@'[hWp,hULMܛ5$؆Ⱥ26<{@V(:]f< ՏhX6g.Hp8yҤu7{u͙HX[&ް0;]_,KKK*Q KQIc *"F+$ıh* gg,a T]/ B >Y4khp~h2\lf=$)!U;>]рJ#)9*+c6dq4}@̍\@=wO&%@!'fضoLuA`: $SɨFY+V놪bYݘ$"NK;oXn"Cg]@J=|xg3t|-ӽWI_1ؾoNZtׁ; Z8ѯjbfr7 W]תVWu=_^L9ը.-=2?UGrkGo ܼ[|^Vc<\\j.Ym%رf9q^7 5?đޝ#k$h3s%a`jηR} ͒#)\1'k!lj#~Iy5KWY3AQ I6[0=f/ᄈZᝯQHq̿F0+WD~*#gSD!$˧NW5+o+dD21׺<:xN3ȺrPrOӤwJ__ qdvYPCt pkQPPh0BRO+8DIG y"8ZaNqe"RDl!8Y\- fwNf#&LέR,&ٛxNJ.EyJ|:sjj,|_o:}/{X!^L}H3ih^*BO, %!h C4R;>T%$ l~{p_؜Jh_ W=p4*[F0Tצ"Q)p$qww㠱OBbuG\vvKe`͒FAn *KӆD3Jש{Ub,DZ[njwOEN|z7Jۛ!ٗ|)`5>@7fj *.~ZۊaٺDQ.׊#Z,Q6ƇYK}<|eࣷkDpxv~;ݏF ZR4lKydQMZ:M|ihBjcw~~1Ty|\;H' LB`o?Ǿ pթD^RTc<+`o,'נkYzy;cy_Ftgc<OHf/s@f%=er1; f0\L}6^'xsP x}S%:ypm4s3v=|tJF~!蝔C^YEɯyQ-jγ-kn dcM-kqʛs.su[d=Xbkݍ#/CĒ3C5/zYY )"n/ތ@ .&x:yWA<,g\[sVJC=.8(NSF8Y{e$ql(y\RэTcʟs)ls`0R"8Up1@ -3bmVZ1iE6gm . h<_u O -FJ`Oip-"H0A9H&i52D^4hX՚+RB4B(K,pp@|ZA@8 qp@8Y QBˌEˬuP k `ò1mr#8pցjI1i+״pDfX"!I`HB<8\x7F 6cu9aťڀs&MDyTeʼ y SYr05{`VyPFJ?s l_N V]Lg,8I}r,[dPxNcKѮ*r)>&hN=^ɾW3Y2ókߋ%&y0asRt%F9b4\0Ϡ׶dWSXrfv7[x%|!T0~fqn4-Wցw`) Us{UN팴\I"wwh_uzSm9?NTyD_'(XAwq`F@I˫hЉjNcZb(|&T s,:9or7,| n3Rg= lRb2O{;1:W=%=)1˔jpjJLEƎRbpڶwUɆԓ*^' խ{- ޹~:=M'6'l̬9~p~fw;^ ]1O͝1gyX~%qp7#0fGcD ԙl(}@34FRKF>nOPK}uy*D)KW~^ceu# EL)u{Cu Etv;gZHvڭ{T(QBBqm/S +Emwv!=w`ܹzSHg'{E .qaں\atC'Zj&ՎJW)l9^Z\G's@B4TjD0ZvիC:HVV'LV ׌cI~}=t8L7R\HS;v7iJ<#Lj~֤m%K+[fG}0x%]:jA" sibMGs#̙: J0"\)X۩x5cpHy'Ĩџc]36S` > )vŽ;KS`-ěm ?%;.R5МDrJ\dyh: 'l/d7i)bˀ0pB Gf;'0;~t䤒ui~+53 !>u t)W2Kɵ&-xr!Ƞc@D>wʰ G/{)4x!;7K?HbԖw<v_KbѦ3ӯ-?i̛ZFcX=xη=_jV;?r|惌k{u#h%I牭ywT3h,ݲYoy׻-[R7;6BC,}4 /Kn=|GŨv38k(~ԓFJ IX+EVEAy VB\RJ5z'YoRmadaH& /: õX[IHRXlprE!I,4­ӁhMVIYX#IĻ0-4];ʋ%c3nfE׼bn?4yi﷽VTWp)[ ~%2Je]—b/K;Z.y "PDHPtʺ1X;Uw0~,QY?V_0V3soOmZIޠ=Ry!BֺkT3bSiزTۊMݿz'5q +^KVOIFe[1rsqïH b\5)cFygnNw^ Zů㒚V~J` 'p[e,bQO^kl3[SD<P慓0QEutZz0 9 ~ |Jq{Jہ|O!M0ւmG6{J=*+Wv @%"iS!Ev>L9aR,RIQlG6'cosE1jpONmpXq(zmt Aw4q<(TPw?.$.2%ߩ״Gڭ+)M/i[Rmh'b[ڍ'#IM:Q#zu.'mLD}$)O ey-xc0S;6L4F*wT(w}}槲I<ZT"{\Io#ow? wmq,~Y`?- @,vاQsX:&Hzg| [=,bU}M~.*\Q,!O#_||Q;Mw5vv-E} ʛ̴6yLK5s,N4?:w9dfx/fk{b?B:49 +,֒ipB2v;h=6z}U[s8J'ޣ?TCP ah!^0|&Md!Q6D9x TAi9k͟7g 4n,z?^+A>_V 1uxw p;i:tcBKyv HQU^!Ē1putEjm|6.T*nχQ* pZ۾[oRIz"Q/ UUq5Fk.@t DCf"(tk JY.!)Y @YCX^T`|E k$XMEЄE'eJd#)jʐMd8#hk/ۥ;uq$'RjqGj%$;6_ҵyAܡJRS;q+Vq %|>&"IJQ iHjprOg8xpvwTH o0L}ؿn&"u$<7N` ~7kl6۵tZ[<[]+}$*Kl/6Yj$&>B۷Ty%:=J'fa݂BB;?xJIbc[|zdizq+c&~15b4_'?rV<?gg^4=N9xS*lS)cw=7xmnRQ: pNcu辅Ï:ٸ^'~&uc='5=5 Q<IAӼtVQRww/.d/tzucU"C@tsxw]>%ɯ$/${u:1&1;pJ`nhON>|ѭW-c1wZ娅E}i Nvb*Rcm*4i+ aXy@ֳV!DkPdYl 8!bgh$'VX挵hVPb2b %Hh;KZ|[ $G7P5s(֖_ Bp7}Tnf)<8~["C >3@[x 9E5?<0'Z1,o:8uڑm Ő棬hP~vv8LH#"ͧAU; ||2R `@q`eQG>#C@\^nHe4h21bMI;SIkW{izYP}(M[!HB DED @{&YBepRλ4ൻ7컋k<'WJL#YbGpgc+ZDކXb!AdZ0 8z$kHZj0>.A2B)_W`,%vHm4 d!֠39Z$BlTl%HIBNlR ME(TMk4xt SKXE;g#u6( )E &T|_;٘jMrJ;LH6X!`U+fbJhbScK"S][Rsh3ּ*ggSo*\sΉozENIkץ}\ZzֆA(i^k2R5vvYx/Mb@$k38Z$Ӈ$rY2TtQ ^pdW%mbyp@}NmdTfSy>fU)&Y$2Qv}";l8v(] `Eod.-29) m كQ/HgE"!r*Z6KϬL')& 2gEYƃpZ"(ʛ[n"/<CJ>]E*F9yu / yy"zӥD`#_eIKVW/[ʨ|"*gBh)@* sؒ4.Vf)m?ln6Z=l7XfC+_xcWEӊUUJ4O}UyVBl=Ȕ}~9AM7<TZ|vyw{uwͿ[:gs;U_ƫ'6'hxw ֋zIVe.?u(Y[~V=⏛=2m<*`GeNfTBF,h_eJ"9Ji@58o~O0-f lB4BllOpqJ2p\#@R0mvar/home/core/zuul-output/logs/kubelet.log0000644000000000000000003326012215153042033017670 0ustar rootrootMar 07 14:39:16 crc systemd[1]: Starting Kubernetes Kubelet... Mar 07 14:39:16 crc restorecon[4681]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Mar 07 14:39:16 crc restorecon[4681]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 07 14:39:16 crc restorecon[4681]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 07 14:39:16 crc restorecon[4681]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 07 14:39:16 crc restorecon[4681]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 07 14:39:16 crc restorecon[4681]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 07 14:39:16 crc restorecon[4681]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 07 14:39:16 crc restorecon[4681]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 07 14:39:16 crc restorecon[4681]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 07 14:39:16 crc restorecon[4681]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 07 14:39:16 crc restorecon[4681]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Mar 07 14:39:16 crc restorecon[4681]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 07 14:39:16 crc restorecon[4681]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 07 14:39:16 crc restorecon[4681]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 07 14:39:16 crc restorecon[4681]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 07 14:39:16 crc restorecon[4681]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 07 14:39:16 crc restorecon[4681]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Mar 07 14:39:16 crc restorecon[4681]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 07 14:39:16 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 07 14:39:16 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 07 14:39:16 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 07 14:39:16 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 07 14:39:16 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 07 14:39:17 crc restorecon[4681]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 07 14:39:17 crc restorecon[4681]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Mar 07 14:39:18 crc kubenswrapper[4943]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 07 14:39:18 crc kubenswrapper[4943]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Mar 07 14:39:18 crc kubenswrapper[4943]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 07 14:39:18 crc kubenswrapper[4943]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 07 14:39:18 crc kubenswrapper[4943]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Mar 07 14:39:18 crc kubenswrapper[4943]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.481882 4943 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.487416 4943 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.488395 4943 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.488423 4943 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.488435 4943 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.488444 4943 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.488453 4943 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.488461 4943 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.488472 4943 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.488480 4943 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.488488 4943 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.488496 4943 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.488503 4943 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.488512 4943 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.488520 4943 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.488528 4943 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.488536 4943 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.488544 4943 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.488552 4943 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.488560 4943 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.488568 4943 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.488576 4943 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.488584 4943 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.488596 4943 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.488606 4943 feature_gate.go:330] unrecognized feature gate: Example Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.488614 4943 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.488623 4943 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.488631 4943 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.488640 4943 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.488651 4943 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.488661 4943 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.488672 4943 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.488683 4943 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.488693 4943 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.488703 4943 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.488714 4943 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.488725 4943 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.488735 4943 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.488744 4943 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.488753 4943 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.488761 4943 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.488769 4943 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.488777 4943 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.488785 4943 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.488792 4943 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.488897 4943 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.488912 4943 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.488923 4943 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.488956 4943 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.488965 4943 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.488973 4943 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.488981 4943 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.488989 4943 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.488997 4943 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.489007 4943 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.489017 4943 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.489025 4943 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.489033 4943 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.489041 4943 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.489048 4943 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.489056 4943 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.489063 4943 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.489071 4943 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.489079 4943 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.489086 4943 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.489093 4943 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.489101 4943 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.489110 4943 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.489120 4943 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.489128 4943 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.489135 4943 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.489142 4943 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.490635 4943 flags.go:64] FLAG: --address="0.0.0.0" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.490657 4943 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.490676 4943 flags.go:64] FLAG: --anonymous-auth="true" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.490687 4943 flags.go:64] FLAG: --application-metrics-count-limit="100" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.490699 4943 flags.go:64] FLAG: --authentication-token-webhook="false" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.490708 4943 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.490720 4943 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.490732 4943 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.490741 4943 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.490751 4943 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.490760 4943 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.490770 4943 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.490780 4943 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.490789 4943 flags.go:64] FLAG: --cgroup-root="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.490798 4943 flags.go:64] FLAG: --cgroups-per-qos="true" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.490807 4943 flags.go:64] FLAG: --client-ca-file="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.490817 4943 flags.go:64] FLAG: --cloud-config="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.490826 4943 flags.go:64] FLAG: --cloud-provider="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.490835 4943 flags.go:64] FLAG: --cluster-dns="[]" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.490848 4943 flags.go:64] FLAG: --cluster-domain="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.490857 4943 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.490866 4943 flags.go:64] FLAG: --config-dir="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.490875 4943 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.490884 4943 flags.go:64] FLAG: --container-log-max-files="5" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.490895 4943 flags.go:64] FLAG: --container-log-max-size="10Mi" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.490904 4943 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.490914 4943 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.490924 4943 flags.go:64] FLAG: --containerd-namespace="k8s.io" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.490960 4943 flags.go:64] FLAG: --contention-profiling="false" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.490969 4943 flags.go:64] FLAG: --cpu-cfs-quota="true" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.490978 4943 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.490987 4943 flags.go:64] FLAG: --cpu-manager-policy="none" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.490997 4943 flags.go:64] FLAG: --cpu-manager-policy-options="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491008 4943 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491017 4943 flags.go:64] FLAG: --enable-controller-attach-detach="true" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491026 4943 flags.go:64] FLAG: --enable-debugging-handlers="true" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491035 4943 flags.go:64] FLAG: --enable-load-reader="false" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491044 4943 flags.go:64] FLAG: --enable-server="true" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491052 4943 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491063 4943 flags.go:64] FLAG: --event-burst="100" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491072 4943 flags.go:64] FLAG: --event-qps="50" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491081 4943 flags.go:64] FLAG: --event-storage-age-limit="default=0" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491090 4943 flags.go:64] FLAG: --event-storage-event-limit="default=0" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491099 4943 flags.go:64] FLAG: --eviction-hard="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491109 4943 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491118 4943 flags.go:64] FLAG: --eviction-minimum-reclaim="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491127 4943 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491136 4943 flags.go:64] FLAG: --eviction-soft="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491146 4943 flags.go:64] FLAG: --eviction-soft-grace-period="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491154 4943 flags.go:64] FLAG: --exit-on-lock-contention="false" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491163 4943 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491172 4943 flags.go:64] FLAG: --experimental-mounter-path="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491180 4943 flags.go:64] FLAG: --fail-cgroupv1="false" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491189 4943 flags.go:64] FLAG: --fail-swap-on="true" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491198 4943 flags.go:64] FLAG: --feature-gates="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491208 4943 flags.go:64] FLAG: --file-check-frequency="20s" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491217 4943 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491227 4943 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491237 4943 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491247 4943 flags.go:64] FLAG: --healthz-port="10248" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491256 4943 flags.go:64] FLAG: --help="false" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491265 4943 flags.go:64] FLAG: --hostname-override="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491301 4943 flags.go:64] FLAG: --housekeeping-interval="10s" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491315 4943 flags.go:64] FLAG: --http-check-frequency="20s" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491326 4943 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491335 4943 flags.go:64] FLAG: --image-credential-provider-config="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491343 4943 flags.go:64] FLAG: --image-gc-high-threshold="85" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491352 4943 flags.go:64] FLAG: --image-gc-low-threshold="80" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491361 4943 flags.go:64] FLAG: --image-service-endpoint="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491370 4943 flags.go:64] FLAG: --kernel-memcg-notification="false" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491379 4943 flags.go:64] FLAG: --kube-api-burst="100" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491388 4943 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491397 4943 flags.go:64] FLAG: --kube-api-qps="50" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491406 4943 flags.go:64] FLAG: --kube-reserved="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491415 4943 flags.go:64] FLAG: --kube-reserved-cgroup="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491423 4943 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491432 4943 flags.go:64] FLAG: --kubelet-cgroups="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491441 4943 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491450 4943 flags.go:64] FLAG: --lock-file="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491458 4943 flags.go:64] FLAG: --log-cadvisor-usage="false" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491467 4943 flags.go:64] FLAG: --log-flush-frequency="5s" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491476 4943 flags.go:64] FLAG: --log-json-info-buffer-size="0" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491489 4943 flags.go:64] FLAG: --log-json-split-stream="false" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491498 4943 flags.go:64] FLAG: --log-text-info-buffer-size="0" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491507 4943 flags.go:64] FLAG: --log-text-split-stream="false" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491516 4943 flags.go:64] FLAG: --logging-format="text" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491525 4943 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491534 4943 flags.go:64] FLAG: --make-iptables-util-chains="true" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491543 4943 flags.go:64] FLAG: --manifest-url="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491552 4943 flags.go:64] FLAG: --manifest-url-header="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491567 4943 flags.go:64] FLAG: --max-housekeeping-interval="15s" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491576 4943 flags.go:64] FLAG: --max-open-files="1000000" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491587 4943 flags.go:64] FLAG: --max-pods="110" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491596 4943 flags.go:64] FLAG: --maximum-dead-containers="-1" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491605 4943 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491614 4943 flags.go:64] FLAG: --memory-manager-policy="None" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491623 4943 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491632 4943 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491641 4943 flags.go:64] FLAG: --node-ip="192.168.126.11" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491650 4943 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491670 4943 flags.go:64] FLAG: --node-status-max-images="50" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491679 4943 flags.go:64] FLAG: --node-status-update-frequency="10s" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491688 4943 flags.go:64] FLAG: --oom-score-adj="-999" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491696 4943 flags.go:64] FLAG: --pod-cidr="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491706 4943 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491720 4943 flags.go:64] FLAG: --pod-manifest-path="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491728 4943 flags.go:64] FLAG: --pod-max-pids="-1" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491737 4943 flags.go:64] FLAG: --pods-per-core="0" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491746 4943 flags.go:64] FLAG: --port="10250" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491755 4943 flags.go:64] FLAG: --protect-kernel-defaults="false" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491764 4943 flags.go:64] FLAG: --provider-id="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491772 4943 flags.go:64] FLAG: --qos-reserved="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491781 4943 flags.go:64] FLAG: --read-only-port="10255" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491790 4943 flags.go:64] FLAG: --register-node="true" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491799 4943 flags.go:64] FLAG: --register-schedulable="true" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491808 4943 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491822 4943 flags.go:64] FLAG: --registry-burst="10" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491831 4943 flags.go:64] FLAG: --registry-qps="5" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491839 4943 flags.go:64] FLAG: --reserved-cpus="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491848 4943 flags.go:64] FLAG: --reserved-memory="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491859 4943 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491867 4943 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491877 4943 flags.go:64] FLAG: --rotate-certificates="false" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491887 4943 flags.go:64] FLAG: --rotate-server-certificates="false" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491896 4943 flags.go:64] FLAG: --runonce="false" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491904 4943 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491914 4943 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491923 4943 flags.go:64] FLAG: --seccomp-default="false" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491960 4943 flags.go:64] FLAG: --serialize-image-pulls="true" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491969 4943 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491977 4943 flags.go:64] FLAG: --storage-driver-db="cadvisor" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491987 4943 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.491995 4943 flags.go:64] FLAG: --storage-driver-password="root" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.492004 4943 flags.go:64] FLAG: --storage-driver-secure="false" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.492014 4943 flags.go:64] FLAG: --storage-driver-table="stats" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.492023 4943 flags.go:64] FLAG: --storage-driver-user="root" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.492032 4943 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.492041 4943 flags.go:64] FLAG: --sync-frequency="1m0s" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.492050 4943 flags.go:64] FLAG: --system-cgroups="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.492058 4943 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.492073 4943 flags.go:64] FLAG: --system-reserved-cgroup="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.492082 4943 flags.go:64] FLAG: --tls-cert-file="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.492090 4943 flags.go:64] FLAG: --tls-cipher-suites="[]" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.492101 4943 flags.go:64] FLAG: --tls-min-version="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.492110 4943 flags.go:64] FLAG: --tls-private-key-file="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.492119 4943 flags.go:64] FLAG: --topology-manager-policy="none" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.492127 4943 flags.go:64] FLAG: --topology-manager-policy-options="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.492137 4943 flags.go:64] FLAG: --topology-manager-scope="container" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.492145 4943 flags.go:64] FLAG: --v="2" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.492156 4943 flags.go:64] FLAG: --version="false" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.492167 4943 flags.go:64] FLAG: --vmodule="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.492177 4943 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.492187 4943 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492408 4943 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492421 4943 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492440 4943 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492448 4943 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492457 4943 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492464 4943 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492472 4943 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492480 4943 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492488 4943 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492496 4943 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492503 4943 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492511 4943 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492540 4943 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492550 4943 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492559 4943 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492568 4943 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492577 4943 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492587 4943 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492596 4943 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492604 4943 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492612 4943 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492620 4943 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492628 4943 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492637 4943 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492645 4943 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492653 4943 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492663 4943 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492672 4943 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492680 4943 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492688 4943 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492696 4943 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492703 4943 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492710 4943 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492719 4943 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492730 4943 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492738 4943 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492746 4943 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492753 4943 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492761 4943 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492771 4943 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492779 4943 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492787 4943 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492794 4943 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492802 4943 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492810 4943 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492817 4943 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492825 4943 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492833 4943 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492840 4943 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492847 4943 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492855 4943 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492863 4943 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492871 4943 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492879 4943 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492887 4943 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492897 4943 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492907 4943 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492915 4943 feature_gate.go:330] unrecognized feature gate: Example Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492948 4943 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492957 4943 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492965 4943 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492976 4943 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492986 4943 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.492995 4943 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.493002 4943 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.493012 4943 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.493023 4943 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.493031 4943 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.493039 4943 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.493046 4943 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.493054 4943 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.493076 4943 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.504563 4943 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.504613 4943 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.504719 4943 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.504731 4943 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.504737 4943 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.504743 4943 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.504750 4943 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.504757 4943 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.504762 4943 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.504767 4943 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.504772 4943 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.504778 4943 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.504783 4943 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.504789 4943 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.504795 4943 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.504801 4943 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.504809 4943 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.504819 4943 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.504827 4943 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.504833 4943 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.504838 4943 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.504844 4943 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.504850 4943 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.504855 4943 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.504860 4943 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.504866 4943 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.504871 4943 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.504876 4943 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.504881 4943 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.504887 4943 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.504892 4943 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.504897 4943 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.504903 4943 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.504908 4943 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.504913 4943 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.504918 4943 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.504941 4943 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.504946 4943 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.504952 4943 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.504957 4943 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.504962 4943 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.504967 4943 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.504973 4943 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.504977 4943 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.504982 4943 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.504987 4943 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.504992 4943 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.504997 4943 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505002 4943 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505008 4943 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505013 4943 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505018 4943 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505023 4943 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505028 4943 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505033 4943 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505037 4943 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505042 4943 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505047 4943 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505051 4943 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505056 4943 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505060 4943 feature_gate.go:330] unrecognized feature gate: Example Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505065 4943 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505071 4943 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505076 4943 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505080 4943 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505085 4943 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505092 4943 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505097 4943 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505102 4943 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505108 4943 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505113 4943 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505118 4943 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505124 4943 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.505134 4943 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505316 4943 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505324 4943 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505329 4943 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505334 4943 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505339 4943 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505343 4943 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505351 4943 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505358 4943 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505365 4943 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505371 4943 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505376 4943 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505381 4943 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505386 4943 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505392 4943 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505397 4943 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505402 4943 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505407 4943 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505412 4943 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505418 4943 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505423 4943 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505428 4943 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505433 4943 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505438 4943 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505443 4943 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505449 4943 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505454 4943 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505459 4943 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505463 4943 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505468 4943 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505473 4943 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505478 4943 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505483 4943 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505489 4943 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505494 4943 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505500 4943 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505506 4943 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505513 4943 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505519 4943 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505526 4943 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505533 4943 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505540 4943 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505545 4943 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505550 4943 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505555 4943 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505560 4943 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505564 4943 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505569 4943 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505574 4943 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505578 4943 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505583 4943 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505587 4943 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505592 4943 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505597 4943 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505601 4943 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505606 4943 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505611 4943 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505616 4943 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505622 4943 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505626 4943 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505631 4943 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505636 4943 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505641 4943 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505646 4943 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505650 4943 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505655 4943 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505659 4943 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505665 4943 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505669 4943 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505674 4943 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505678 4943 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.505685 4943 feature_gate.go:330] unrecognized feature gate: Example Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.505694 4943 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.506184 4943 server.go:940] "Client rotation is on, will bootstrap in background" Mar 07 14:39:18 crc kubenswrapper[4943]: E0307 14:39:18.510974 4943 bootstrap.go:266] "Unhandled Error" err="part of the existing bootstrap client certificate in /var/lib/kubelet/kubeconfig is expired: 2026-02-24 05:52:08 +0000 UTC" logger="UnhandledError" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.519798 4943 bootstrap.go:101] "Use the bootstrap credentials to request a cert, and set kubeconfig to point to the certificate dir" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.520115 4943 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.522237 4943 server.go:997] "Starting client certificate rotation" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.522291 4943 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.522460 4943 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.550974 4943 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 07 14:39:18 crc kubenswrapper[4943]: E0307 14:39:18.554424 4943 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.75:6443: connect: connection refused" logger="UnhandledError" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.556250 4943 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.571780 4943 log.go:25] "Validated CRI v1 runtime API" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.615327 4943 log.go:25] "Validated CRI v1 image API" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.617477 4943 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.621866 4943 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-03-07-14-35-01-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.621915 4943 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.652390 4943 manager.go:217] Machine: {Timestamp:2026-03-07 14:39:18.648567635 +0000 UTC m=+0.600704213 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:bf22d61b-d7ce-44b5-9dd8-1a46c0f78104 BootID:7a039d39-2e22-491c-a9ee-09ff4d277e8b Filesystems:[{Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:1f:01:57 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:1f:01:57 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:67:c8:ea Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:f1:46:f8 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:32:05:7c Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:ec:4f:74 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:be:29:5b:9b:98:8e Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:fe:2d:8b:2a:3b:c8 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.652773 4943 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.653083 4943 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.655093 4943 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.655566 4943 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.655658 4943 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.656106 4943 topology_manager.go:138] "Creating topology manager with none policy" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.656129 4943 container_manager_linux.go:303] "Creating device plugin manager" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.656831 4943 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.656893 4943 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.657406 4943 state_mem.go:36] "Initialized new in-memory state store" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.657617 4943 server.go:1245] "Using root directory" path="/var/lib/kubelet" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.661220 4943 kubelet.go:418] "Attempting to sync node with API server" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.661265 4943 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.661313 4943 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.661340 4943 kubelet.go:324] "Adding apiserver pod source" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.661361 4943 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.665965 4943 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.667016 4943 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.667284 4943 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.75:6443: connect: connection refused Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.667317 4943 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.75:6443: connect: connection refused Mar 07 14:39:18 crc kubenswrapper[4943]: E0307 14:39:18.667420 4943 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.75:6443: connect: connection refused" logger="UnhandledError" Mar 07 14:39:18 crc kubenswrapper[4943]: E0307 14:39:18.667448 4943 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.75:6443: connect: connection refused" logger="UnhandledError" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.669236 4943 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.670454 4943 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.670479 4943 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.670487 4943 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.670495 4943 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.670508 4943 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.670517 4943 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.670525 4943 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.670536 4943 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.670545 4943 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.670555 4943 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.670565 4943 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.670573 4943 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.671557 4943 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.671958 4943 server.go:1280] "Started kubelet" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.673680 4943 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.673686 4943 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 07 14:39:18 crc systemd[1]: Started Kubernetes Kubelet. Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.674809 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.75:6443: connect: connection refused Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.675403 4943 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.676102 4943 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.676159 4943 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Mar 07 14:39:18 crc kubenswrapper[4943]: E0307 14:39:18.676442 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.676645 4943 server.go:460] "Adding debug handlers to kubelet server" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.676852 4943 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.677019 4943 volume_manager.go:287] "The desired_state_of_world populator starts" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.677057 4943 volume_manager.go:289] "Starting Kubelet Volume Manager" Mar 07 14:39:18 crc kubenswrapper[4943]: E0307 14:39:18.677182 4943 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.75:6443: connect: connection refused" interval="200ms" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.677445 4943 factory.go:55] Registering systemd factory Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.677471 4943 factory.go:221] Registration of the systemd container factory successfully Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.683064 4943 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.75:6443: connect: connection refused Mar 07 14:39:18 crc kubenswrapper[4943]: E0307 14:39:18.683214 4943 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.75:6443: connect: connection refused" logger="UnhandledError" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.683831 4943 factory.go:153] Registering CRI-O factory Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.683909 4943 factory.go:221] Registration of the crio container factory successfully Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.684033 4943 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.684118 4943 factory.go:103] Registering Raw factory Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.684178 4943 manager.go:1196] Started watching for new ooms in manager Mar 07 14:39:18 crc kubenswrapper[4943]: E0307 14:39:18.685733 4943 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.75:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.189a9606a8eeed5d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:18.671912285 +0000 UTC m=+0.624048783,LastTimestamp:2026-03-07 14:39:18.671912285 +0000 UTC m=+0.624048783,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.690006 4943 manager.go:319] Starting recovery of all containers Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.697409 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.697688 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.697877 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.698083 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.698262 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.698436 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.698627 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.698811 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.699044 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.699290 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.699483 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.699665 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.699802 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.699916 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.700056 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.700308 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.700363 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.700387 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.700406 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.700424 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.700441 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.700461 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.700481 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.700503 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.700522 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.700540 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.700561 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.700620 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.700638 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.700657 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.700679 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.700699 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.700717 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.700737 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.700814 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.700837 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.700855 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.700871 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.700894 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.700916 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.700963 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.700984 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.701000 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.701017 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.701034 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.701086 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.701108 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.701127 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.701145 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.701162 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.701179 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.701196 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.701222 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.701242 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.701259 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.701281 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.701300 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.701318 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.701336 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.701353 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.701370 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.701389 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.701405 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.701422 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.701441 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.701458 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.701474 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.701491 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.701509 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.701526 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.701543 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.701562 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.701579 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.701595 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.701611 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.701628 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.701644 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.701662 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.701684 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.701702 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.701718 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.701736 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.701753 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.701772 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.701792 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.701812 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.701829 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.701848 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.701868 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.701885 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.701904 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.701922 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.701966 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.701987 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.702006 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.702023 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.702039 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.702058 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.702076 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.702100 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.702116 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.702133 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.702153 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.702173 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.702210 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.702231 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.702250 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.702269 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.702289 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.702308 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.702328 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.702368 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.702390 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.702410 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.702428 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.702449 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.702468 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.702488 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.702507 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.702529 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.702548 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.702566 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.702583 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.702602 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.702618 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.702635 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.702653 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.702671 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.702691 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.702707 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.702727 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.702747 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.702765 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.702782 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.702797 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.702815 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.702833 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.702849 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.702865 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.702881 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.702897 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.702914 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.702957 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.702974 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.702992 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.703007 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.703022 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.703038 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.703059 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.703083 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.703103 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.703124 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.703141 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.703156 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.703172 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.703190 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.703206 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.703220 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.703253 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.703269 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.703285 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.703301 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.703374 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.703398 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.703416 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.703433 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.703451 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.703469 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.703486 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.703503 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.703520 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.703537 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.703553 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.703569 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.709224 4943 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.709653 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.709675 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.709695 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.709720 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.709739 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.709757 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.709776 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.709811 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.709834 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.709853 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.709870 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.709886 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.709903 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.709922 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.709966 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.709984 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.710002 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.710022 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.710039 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.710057 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.710074 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.710092 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.710110 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.710130 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.710149 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.710169 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.710188 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.710208 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.710227 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.710244 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.710260 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.710278 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.710296 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.710312 4943 reconstruct.go:97] "Volume reconstruction finished" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.710324 4943 reconciler.go:26] "Reconciler: start to sync state" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.722734 4943 manager.go:324] Recovery completed Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.735392 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.738013 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.738067 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.738079 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.742701 4943 cpu_manager.go:225] "Starting CPU manager" policy="none" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.742743 4943 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.742776 4943 state_mem.go:36] "Initialized new in-memory state store" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.750721 4943 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.753840 4943 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.753894 4943 status_manager.go:217] "Starting to sync pod status with apiserver" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.753958 4943 kubelet.go:2335] "Starting kubelet main sync loop" Mar 07 14:39:18 crc kubenswrapper[4943]: E0307 14:39:18.754025 4943 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 07 14:39:18 crc kubenswrapper[4943]: W0307 14:39:18.756002 4943 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.75:6443: connect: connection refused Mar 07 14:39:18 crc kubenswrapper[4943]: E0307 14:39:18.756087 4943 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.75:6443: connect: connection refused" logger="UnhandledError" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.762254 4943 policy_none.go:49] "None policy: Start" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.763102 4943 memory_manager.go:170] "Starting memorymanager" policy="None" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.763137 4943 state_mem.go:35] "Initializing new in-memory state store" Mar 07 14:39:18 crc kubenswrapper[4943]: E0307 14:39:18.776594 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.838508 4943 manager.go:334] "Starting Device Plugin manager" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.838556 4943 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.838570 4943 server.go:79] "Starting device plugin registration server" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.839055 4943 eviction_manager.go:189] "Eviction manager: starting control loop" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.839072 4943 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.839573 4943 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.839669 4943 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.839678 4943 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.854676 4943 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.854799 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:18 crc kubenswrapper[4943]: E0307 14:39:18.855730 4943 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.856236 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.856282 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.856299 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.856789 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.856830 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.857380 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.858282 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.858307 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.858320 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.858328 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.858354 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.858425 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.858428 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.858580 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.858664 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.859382 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.859420 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.859434 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.859567 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.859690 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.859735 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.860656 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.860679 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.860690 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.860958 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.861013 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.861084 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.861224 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.861246 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.861259 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.861361 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.861491 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.861535 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.862111 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.862157 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.862175 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.862368 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.862385 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.862393 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.862467 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.862512 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.864037 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.864055 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.864063 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:18 crc kubenswrapper[4943]: E0307 14:39:18.877858 4943 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.75:6443: connect: connection refused" interval="400ms" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.911983 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.912038 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.912076 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.912216 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.912324 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.912383 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.912413 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.912444 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.912465 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.912487 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.912506 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.912539 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.912559 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.912577 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.912598 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.939413 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.940868 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.941016 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.941124 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:18 crc kubenswrapper[4943]: I0307 14:39:18.941225 4943 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 07 14:39:18 crc kubenswrapper[4943]: E0307 14:39:18.941716 4943 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.75:6443: connect: connection refused" node="crc" Mar 07 14:39:19 crc kubenswrapper[4943]: I0307 14:39:19.014205 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 07 14:39:19 crc kubenswrapper[4943]: I0307 14:39:19.014261 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 07 14:39:19 crc kubenswrapper[4943]: I0307 14:39:19.014280 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 07 14:39:19 crc kubenswrapper[4943]: I0307 14:39:19.014298 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 07 14:39:19 crc kubenswrapper[4943]: I0307 14:39:19.014325 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 07 14:39:19 crc kubenswrapper[4943]: I0307 14:39:19.014343 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 07 14:39:19 crc kubenswrapper[4943]: I0307 14:39:19.014361 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 07 14:39:19 crc kubenswrapper[4943]: I0307 14:39:19.014378 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 07 14:39:19 crc kubenswrapper[4943]: I0307 14:39:19.014397 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 07 14:39:19 crc kubenswrapper[4943]: I0307 14:39:19.014416 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 07 14:39:19 crc kubenswrapper[4943]: I0307 14:39:19.014428 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 07 14:39:19 crc kubenswrapper[4943]: I0307 14:39:19.014457 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 07 14:39:19 crc kubenswrapper[4943]: I0307 14:39:19.014514 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 07 14:39:19 crc kubenswrapper[4943]: I0307 14:39:19.014478 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 07 14:39:19 crc kubenswrapper[4943]: I0307 14:39:19.014469 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 07 14:39:19 crc kubenswrapper[4943]: I0307 14:39:19.014540 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 07 14:39:19 crc kubenswrapper[4943]: I0307 14:39:19.014543 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 07 14:39:19 crc kubenswrapper[4943]: I0307 14:39:19.014562 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 07 14:39:19 crc kubenswrapper[4943]: I0307 14:39:19.014584 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 07 14:39:19 crc kubenswrapper[4943]: I0307 14:39:19.014633 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 07 14:39:19 crc kubenswrapper[4943]: I0307 14:39:19.014649 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 07 14:39:19 crc kubenswrapper[4943]: I0307 14:39:19.014656 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 07 14:39:19 crc kubenswrapper[4943]: I0307 14:39:19.014662 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 07 14:39:19 crc kubenswrapper[4943]: I0307 14:39:19.014746 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 07 14:39:19 crc kubenswrapper[4943]: I0307 14:39:19.014813 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 07 14:39:19 crc kubenswrapper[4943]: I0307 14:39:19.014873 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 07 14:39:19 crc kubenswrapper[4943]: I0307 14:39:19.014877 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 07 14:39:19 crc kubenswrapper[4943]: I0307 14:39:19.014964 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 07 14:39:19 crc kubenswrapper[4943]: I0307 14:39:19.014965 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 07 14:39:19 crc kubenswrapper[4943]: I0307 14:39:19.014974 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 07 14:39:19 crc kubenswrapper[4943]: I0307 14:39:19.142059 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:19 crc kubenswrapper[4943]: I0307 14:39:19.143808 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:19 crc kubenswrapper[4943]: I0307 14:39:19.143867 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:19 crc kubenswrapper[4943]: I0307 14:39:19.143878 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:19 crc kubenswrapper[4943]: I0307 14:39:19.143910 4943 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 07 14:39:19 crc kubenswrapper[4943]: E0307 14:39:19.144524 4943 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.75:6443: connect: connection refused" node="crc" Mar 07 14:39:19 crc kubenswrapper[4943]: I0307 14:39:19.199353 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 07 14:39:19 crc kubenswrapper[4943]: I0307 14:39:19.216203 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Mar 07 14:39:19 crc kubenswrapper[4943]: I0307 14:39:19.233175 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 07 14:39:19 crc kubenswrapper[4943]: W0307 14:39:19.250619 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-e6308e7c6f514e8caab5538828f1ad905d2f1893a4019bd8dd6b854134b7df09 WatchSource:0}: Error finding container e6308e7c6f514e8caab5538828f1ad905d2f1893a4019bd8dd6b854134b7df09: Status 404 returned error can't find the container with id e6308e7c6f514e8caab5538828f1ad905d2f1893a4019bd8dd6b854134b7df09 Mar 07 14:39:19 crc kubenswrapper[4943]: W0307 14:39:19.251274 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-3eeb0e02e467e25079ebbb80d327ee626187861d5d9b53545439dc0cc9cf00e8 WatchSource:0}: Error finding container 3eeb0e02e467e25079ebbb80d327ee626187861d5d9b53545439dc0cc9cf00e8: Status 404 returned error can't find the container with id 3eeb0e02e467e25079ebbb80d327ee626187861d5d9b53545439dc0cc9cf00e8 Mar 07 14:39:19 crc kubenswrapper[4943]: I0307 14:39:19.252878 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 07 14:39:19 crc kubenswrapper[4943]: W0307 14:39:19.259745 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-fed5fdbd4ffbfdc813403d337115724541eeac22e8d82fdbb5a3d6c9a18b0b4d WatchSource:0}: Error finding container fed5fdbd4ffbfdc813403d337115724541eeac22e8d82fdbb5a3d6c9a18b0b4d: Status 404 returned error can't find the container with id fed5fdbd4ffbfdc813403d337115724541eeac22e8d82fdbb5a3d6c9a18b0b4d Mar 07 14:39:19 crc kubenswrapper[4943]: I0307 14:39:19.260601 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 07 14:39:19 crc kubenswrapper[4943]: W0307 14:39:19.276429 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-1ea5ccf799977d6d6623f418f7d8efe44ffbe7972c16394de6e0b45e28ff72ae WatchSource:0}: Error finding container 1ea5ccf799977d6d6623f418f7d8efe44ffbe7972c16394de6e0b45e28ff72ae: Status 404 returned error can't find the container with id 1ea5ccf799977d6d6623f418f7d8efe44ffbe7972c16394de6e0b45e28ff72ae Mar 07 14:39:19 crc kubenswrapper[4943]: E0307 14:39:19.278404 4943 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.75:6443: connect: connection refused" interval="800ms" Mar 07 14:39:19 crc kubenswrapper[4943]: W0307 14:39:19.278711 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-495ba70708deede937ae575d2199131bf30c288013910f114f256c55b3c6dfc5 WatchSource:0}: Error finding container 495ba70708deede937ae575d2199131bf30c288013910f114f256c55b3c6dfc5: Status 404 returned error can't find the container with id 495ba70708deede937ae575d2199131bf30c288013910f114f256c55b3c6dfc5 Mar 07 14:39:19 crc kubenswrapper[4943]: I0307 14:39:19.545263 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:19 crc kubenswrapper[4943]: I0307 14:39:19.547479 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:19 crc kubenswrapper[4943]: I0307 14:39:19.547515 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:19 crc kubenswrapper[4943]: I0307 14:39:19.547524 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:19 crc kubenswrapper[4943]: I0307 14:39:19.547549 4943 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 07 14:39:19 crc kubenswrapper[4943]: E0307 14:39:19.547912 4943 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.75:6443: connect: connection refused" node="crc" Mar 07 14:39:19 crc kubenswrapper[4943]: W0307 14:39:19.655098 4943 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.75:6443: connect: connection refused Mar 07 14:39:19 crc kubenswrapper[4943]: E0307 14:39:19.655213 4943 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.75:6443: connect: connection refused" logger="UnhandledError" Mar 07 14:39:19 crc kubenswrapper[4943]: I0307 14:39:19.676453 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.75:6443: connect: connection refused Mar 07 14:39:19 crc kubenswrapper[4943]: I0307 14:39:19.757957 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"fed5fdbd4ffbfdc813403d337115724541eeac22e8d82fdbb5a3d6c9a18b0b4d"} Mar 07 14:39:19 crc kubenswrapper[4943]: I0307 14:39:19.759309 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3eeb0e02e467e25079ebbb80d327ee626187861d5d9b53545439dc0cc9cf00e8"} Mar 07 14:39:19 crc kubenswrapper[4943]: I0307 14:39:19.760949 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"e6308e7c6f514e8caab5538828f1ad905d2f1893a4019bd8dd6b854134b7df09"} Mar 07 14:39:19 crc kubenswrapper[4943]: I0307 14:39:19.762692 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"495ba70708deede937ae575d2199131bf30c288013910f114f256c55b3c6dfc5"} Mar 07 14:39:19 crc kubenswrapper[4943]: I0307 14:39:19.764136 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"1ea5ccf799977d6d6623f418f7d8efe44ffbe7972c16394de6e0b45e28ff72ae"} Mar 07 14:39:19 crc kubenswrapper[4943]: W0307 14:39:19.822387 4943 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.75:6443: connect: connection refused Mar 07 14:39:19 crc kubenswrapper[4943]: E0307 14:39:19.822445 4943 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.75:6443: connect: connection refused" logger="UnhandledError" Mar 07 14:39:19 crc kubenswrapper[4943]: W0307 14:39:19.871481 4943 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.75:6443: connect: connection refused Mar 07 14:39:19 crc kubenswrapper[4943]: E0307 14:39:19.871736 4943 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.75:6443: connect: connection refused" logger="UnhandledError" Mar 07 14:39:20 crc kubenswrapper[4943]: E0307 14:39:20.079819 4943 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.75:6443: connect: connection refused" interval="1.6s" Mar 07 14:39:20 crc kubenswrapper[4943]: W0307 14:39:20.113629 4943 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.75:6443: connect: connection refused Mar 07 14:39:20 crc kubenswrapper[4943]: E0307 14:39:20.113737 4943 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.75:6443: connect: connection refused" logger="UnhandledError" Mar 07 14:39:20 crc kubenswrapper[4943]: I0307 14:39:20.348841 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:20 crc kubenswrapper[4943]: I0307 14:39:20.351045 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:20 crc kubenswrapper[4943]: I0307 14:39:20.351100 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:20 crc kubenswrapper[4943]: I0307 14:39:20.351118 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:20 crc kubenswrapper[4943]: I0307 14:39:20.351158 4943 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 07 14:39:20 crc kubenswrapper[4943]: E0307 14:39:20.351847 4943 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.75:6443: connect: connection refused" node="crc" Mar 07 14:39:20 crc kubenswrapper[4943]: I0307 14:39:20.676705 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.75:6443: connect: connection refused Mar 07 14:39:20 crc kubenswrapper[4943]: I0307 14:39:20.708206 4943 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 07 14:39:20 crc kubenswrapper[4943]: E0307 14:39:20.709742 4943 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.75:6443: connect: connection refused" logger="UnhandledError" Mar 07 14:39:20 crc kubenswrapper[4943]: I0307 14:39:20.769821 4943 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="ba66702b8eac6d7130faec137bc3a2485581bfe1c52ea7448f6a682cb7b95645" exitCode=0 Mar 07 14:39:20 crc kubenswrapper[4943]: I0307 14:39:20.769956 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"ba66702b8eac6d7130faec137bc3a2485581bfe1c52ea7448f6a682cb7b95645"} Mar 07 14:39:20 crc kubenswrapper[4943]: I0307 14:39:20.769979 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:20 crc kubenswrapper[4943]: I0307 14:39:20.771398 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:20 crc kubenswrapper[4943]: I0307 14:39:20.771456 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:20 crc kubenswrapper[4943]: I0307 14:39:20.771478 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:20 crc kubenswrapper[4943]: I0307 14:39:20.772775 4943 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="ecdc3c295ab5f02cbb779cfcd2d14a4f9cd92005a3115425cf034e5d30122f59" exitCode=0 Mar 07 14:39:20 crc kubenswrapper[4943]: I0307 14:39:20.772846 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"ecdc3c295ab5f02cbb779cfcd2d14a4f9cd92005a3115425cf034e5d30122f59"} Mar 07 14:39:20 crc kubenswrapper[4943]: I0307 14:39:20.772853 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:20 crc kubenswrapper[4943]: I0307 14:39:20.777490 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:20 crc kubenswrapper[4943]: I0307 14:39:20.777582 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:20 crc kubenswrapper[4943]: I0307 14:39:20.777626 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:20 crc kubenswrapper[4943]: I0307 14:39:20.782357 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"be6d5d7ea2485f235caa4752de851ab75157c1a1a03a00d0ba8c780ce38d57fe"} Mar 07 14:39:20 crc kubenswrapper[4943]: I0307 14:39:20.782406 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"9d4fc34558c7d17b0533eb3a3cec05dde5f7b0290f783e3afb130af54aefdfd6"} Mar 07 14:39:20 crc kubenswrapper[4943]: I0307 14:39:20.782421 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"0a516a1a0e24f9de9a1038d966ac4195f22bb34969d2c55081256fbacc5379ce"} Mar 07 14:39:20 crc kubenswrapper[4943]: I0307 14:39:20.782434 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"156482e9536d7a822362fe687438f9532e7eadbc2f914c05345730fc1cfd7fda"} Mar 07 14:39:20 crc kubenswrapper[4943]: I0307 14:39:20.782444 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:20 crc kubenswrapper[4943]: I0307 14:39:20.783503 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:20 crc kubenswrapper[4943]: I0307 14:39:20.783542 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:20 crc kubenswrapper[4943]: I0307 14:39:20.783556 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:20 crc kubenswrapper[4943]: I0307 14:39:20.784349 4943 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="87c2d7e4180bc1e9383182a52ba43dd30e41d4e08dd229a29f831abe304722a4" exitCode=0 Mar 07 14:39:20 crc kubenswrapper[4943]: I0307 14:39:20.784487 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"87c2d7e4180bc1e9383182a52ba43dd30e41d4e08dd229a29f831abe304722a4"} Mar 07 14:39:20 crc kubenswrapper[4943]: I0307 14:39:20.784502 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:20 crc kubenswrapper[4943]: I0307 14:39:20.785839 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:20 crc kubenswrapper[4943]: I0307 14:39:20.785881 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:20 crc kubenswrapper[4943]: I0307 14:39:20.785906 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:20 crc kubenswrapper[4943]: I0307 14:39:20.786447 4943 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="3c59fbb426b975a52163b8afcdb765ae2bcac89c8382a28cb94435ca145ca4ab" exitCode=0 Mar 07 14:39:20 crc kubenswrapper[4943]: I0307 14:39:20.786512 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"3c59fbb426b975a52163b8afcdb765ae2bcac89c8382a28cb94435ca145ca4ab"} Mar 07 14:39:20 crc kubenswrapper[4943]: I0307 14:39:20.786687 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:20 crc kubenswrapper[4943]: I0307 14:39:20.787952 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:20 crc kubenswrapper[4943]: I0307 14:39:20.788593 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:20 crc kubenswrapper[4943]: I0307 14:39:20.788659 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:20 crc kubenswrapper[4943]: I0307 14:39:20.788685 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:20 crc kubenswrapper[4943]: I0307 14:39:20.789284 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:20 crc kubenswrapper[4943]: I0307 14:39:20.789350 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:20 crc kubenswrapper[4943]: I0307 14:39:20.789374 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:21 crc kubenswrapper[4943]: W0307 14:39:21.258378 4943 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.75:6443: connect: connection refused Mar 07 14:39:21 crc kubenswrapper[4943]: E0307 14:39:21.258481 4943 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.75:6443: connect: connection refused" logger="UnhandledError" Mar 07 14:39:21 crc kubenswrapper[4943]: W0307 14:39:21.435804 4943 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.75:6443: connect: connection refused Mar 07 14:39:21 crc kubenswrapper[4943]: E0307 14:39:21.435875 4943 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.75:6443: connect: connection refused" logger="UnhandledError" Mar 07 14:39:21 crc kubenswrapper[4943]: I0307 14:39:21.660089 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 07 14:39:21 crc kubenswrapper[4943]: I0307 14:39:21.677480 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.75:6443: connect: connection refused Mar 07 14:39:21 crc kubenswrapper[4943]: E0307 14:39:21.680843 4943 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.75:6443: connect: connection refused" interval="3.2s" Mar 07 14:39:21 crc kubenswrapper[4943]: I0307 14:39:21.790226 4943 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="600cab174b5bd2db862477541412232c3b6fb71a0db7e2392717b902bd664dba" exitCode=0 Mar 07 14:39:21 crc kubenswrapper[4943]: I0307 14:39:21.790313 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"600cab174b5bd2db862477541412232c3b6fb71a0db7e2392717b902bd664dba"} Mar 07 14:39:21 crc kubenswrapper[4943]: I0307 14:39:21.790474 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:21 crc kubenswrapper[4943]: I0307 14:39:21.791994 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:21 crc kubenswrapper[4943]: I0307 14:39:21.792030 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:21 crc kubenswrapper[4943]: I0307 14:39:21.792043 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:21 crc kubenswrapper[4943]: I0307 14:39:21.794858 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"5085a4c1868049f0545bdd3f1108848a7df389f3b1949226af0f125adc7f2af3"} Mar 07 14:39:21 crc kubenswrapper[4943]: I0307 14:39:21.795018 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:21 crc kubenswrapper[4943]: I0307 14:39:21.795992 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:21 crc kubenswrapper[4943]: I0307 14:39:21.796022 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:21 crc kubenswrapper[4943]: I0307 14:39:21.796036 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:21 crc kubenswrapper[4943]: I0307 14:39:21.799423 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"ffebebb63b088a468b4bdef9ee24a09b28fcbbeee029140c00ae81a42b01cc30"} Mar 07 14:39:21 crc kubenswrapper[4943]: I0307 14:39:21.799457 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"f211719d6b694ae49827c51e11edec155de7d732e2a4b19ab43960f8def24bec"} Mar 07 14:39:21 crc kubenswrapper[4943]: I0307 14:39:21.799473 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"5eb947dc35035e1b672fd9abf7ce1505479ca7b11d6df1f1585f49a3eacfcbdf"} Mar 07 14:39:21 crc kubenswrapper[4943]: I0307 14:39:21.799560 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:21 crc kubenswrapper[4943]: I0307 14:39:21.801112 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:21 crc kubenswrapper[4943]: I0307 14:39:21.801142 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:21 crc kubenswrapper[4943]: I0307 14:39:21.801155 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:21 crc kubenswrapper[4943]: I0307 14:39:21.816228 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:21 crc kubenswrapper[4943]: I0307 14:39:21.817253 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"69d5ad495f66b4bdb2e419c05c8a8060bfc94df1a9c5cb3156bed86ba312562b"} Mar 07 14:39:21 crc kubenswrapper[4943]: I0307 14:39:21.817317 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5f9507997ae006ce9864d88b639c5c782aaed107dbd59b42ea0bb34e1606bf0c"} Mar 07 14:39:21 crc kubenswrapper[4943]: I0307 14:39:21.817339 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"46a2cc9aec0e9e4a54e8fa06e6eeaec8ba29e88d0a0672c00f9a01b439ad4331"} Mar 07 14:39:21 crc kubenswrapper[4943]: I0307 14:39:21.817360 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6084025a1e017c5e63892cdc0cc359ad0297878a6eb80cbf8d4d620bb5f2f681"} Mar 07 14:39:21 crc kubenswrapper[4943]: I0307 14:39:21.818098 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:21 crc kubenswrapper[4943]: I0307 14:39:21.818142 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:21 crc kubenswrapper[4943]: I0307 14:39:21.818153 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:21 crc kubenswrapper[4943]: I0307 14:39:21.952648 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:21 crc kubenswrapper[4943]: I0307 14:39:21.953949 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:21 crc kubenswrapper[4943]: I0307 14:39:21.954028 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:21 crc kubenswrapper[4943]: I0307 14:39:21.954039 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:21 crc kubenswrapper[4943]: I0307 14:39:21.954065 4943 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 07 14:39:21 crc kubenswrapper[4943]: E0307 14:39:21.954379 4943 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.75:6443: connect: connection refused" node="crc" Mar 07 14:39:22 crc kubenswrapper[4943]: W0307 14:39:22.280815 4943 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.75:6443: connect: connection refused Mar 07 14:39:22 crc kubenswrapper[4943]: E0307 14:39:22.280973 4943 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.75:6443: connect: connection refused" logger="UnhandledError" Mar 07 14:39:22 crc kubenswrapper[4943]: I0307 14:39:22.823904 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"73a0d1456ae6ea55e508240342d50b14f7dd3856510fc756951c504786e4f49e"} Mar 07 14:39:22 crc kubenswrapper[4943]: I0307 14:39:22.824076 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:22 crc kubenswrapper[4943]: I0307 14:39:22.826022 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:22 crc kubenswrapper[4943]: I0307 14:39:22.826086 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:22 crc kubenswrapper[4943]: I0307 14:39:22.826111 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:22 crc kubenswrapper[4943]: I0307 14:39:22.827166 4943 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="d2facab8651e65dc787f897799529be044fe51660477fb725d67bde6ed3f2cd8" exitCode=0 Mar 07 14:39:22 crc kubenswrapper[4943]: I0307 14:39:22.827328 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"d2facab8651e65dc787f897799529be044fe51660477fb725d67bde6ed3f2cd8"} Mar 07 14:39:22 crc kubenswrapper[4943]: I0307 14:39:22.827361 4943 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 07 14:39:22 crc kubenswrapper[4943]: I0307 14:39:22.827437 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:22 crc kubenswrapper[4943]: I0307 14:39:22.827372 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:22 crc kubenswrapper[4943]: I0307 14:39:22.827467 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:22 crc kubenswrapper[4943]: I0307 14:39:22.827726 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:22 crc kubenswrapper[4943]: I0307 14:39:22.829037 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:22 crc kubenswrapper[4943]: I0307 14:39:22.829084 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:22 crc kubenswrapper[4943]: I0307 14:39:22.829104 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:22 crc kubenswrapper[4943]: I0307 14:39:22.829691 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:22 crc kubenswrapper[4943]: I0307 14:39:22.829782 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:22 crc kubenswrapper[4943]: I0307 14:39:22.829842 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:22 crc kubenswrapper[4943]: I0307 14:39:22.829701 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:22 crc kubenswrapper[4943]: I0307 14:39:22.829731 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:22 crc kubenswrapper[4943]: I0307 14:39:22.830083 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:22 crc kubenswrapper[4943]: I0307 14:39:22.830116 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:22 crc kubenswrapper[4943]: I0307 14:39:22.830033 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:22 crc kubenswrapper[4943]: I0307 14:39:22.830176 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:23 crc kubenswrapper[4943]: I0307 14:39:23.498581 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 07 14:39:23 crc kubenswrapper[4943]: I0307 14:39:23.841868 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"579fcf8ef1f78bb179041a10b152a22463bc5dff174315e1d9d7c53f907d7491"} Mar 07 14:39:23 crc kubenswrapper[4943]: I0307 14:39:23.841946 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3770eb958ca7da4c3504ad491ba6eadd56dbc07ed245662f5a3c8d963889b04b"} Mar 07 14:39:23 crc kubenswrapper[4943]: I0307 14:39:23.841981 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"aa2925025d0d17e5aec74529053f6abbe1286f2a5045c05242bd2c49e2946c0d"} Mar 07 14:39:23 crc kubenswrapper[4943]: I0307 14:39:23.842005 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:23 crc kubenswrapper[4943]: I0307 14:39:23.842117 4943 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 07 14:39:23 crc kubenswrapper[4943]: I0307 14:39:23.842202 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:23 crc kubenswrapper[4943]: I0307 14:39:23.843351 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:23 crc kubenswrapper[4943]: I0307 14:39:23.843392 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:23 crc kubenswrapper[4943]: I0307 14:39:23.843403 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:23 crc kubenswrapper[4943]: I0307 14:39:23.843828 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:23 crc kubenswrapper[4943]: I0307 14:39:23.843863 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:23 crc kubenswrapper[4943]: I0307 14:39:23.843880 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:24 crc kubenswrapper[4943]: I0307 14:39:24.032908 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 07 14:39:24 crc kubenswrapper[4943]: I0307 14:39:24.101668 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 07 14:39:24 crc kubenswrapper[4943]: I0307 14:39:24.102221 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:24 crc kubenswrapper[4943]: I0307 14:39:24.104866 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:24 crc kubenswrapper[4943]: I0307 14:39:24.105587 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:24 crc kubenswrapper[4943]: I0307 14:39:24.105660 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:24 crc kubenswrapper[4943]: I0307 14:39:24.113757 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 07 14:39:24 crc kubenswrapper[4943]: I0307 14:39:24.854877 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:24 crc kubenswrapper[4943]: I0307 14:39:24.855062 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:24 crc kubenswrapper[4943]: I0307 14:39:24.855179 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d8b094b454acb299a0f6cd839c82e79c6dda198cd0dc19f72f67937682c595fa"} Mar 07 14:39:24 crc kubenswrapper[4943]: I0307 14:39:24.855389 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"5c637b6116e2142315c2d4fbdb04d8d1688c8d8ecbc2b8d8aa594addbd0c4bd3"} Mar 07 14:39:24 crc kubenswrapper[4943]: I0307 14:39:24.855503 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:24 crc kubenswrapper[4943]: I0307 14:39:24.856608 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:24 crc kubenswrapper[4943]: I0307 14:39:24.856672 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:24 crc kubenswrapper[4943]: I0307 14:39:24.856697 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:24 crc kubenswrapper[4943]: I0307 14:39:24.857266 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:24 crc kubenswrapper[4943]: I0307 14:39:24.857322 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:24 crc kubenswrapper[4943]: I0307 14:39:24.857343 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:24 crc kubenswrapper[4943]: I0307 14:39:24.857316 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:24 crc kubenswrapper[4943]: I0307 14:39:24.857478 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:24 crc kubenswrapper[4943]: I0307 14:39:24.857518 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:24 crc kubenswrapper[4943]: I0307 14:39:24.907141 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 07 14:39:24 crc kubenswrapper[4943]: I0307 14:39:24.939013 4943 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 07 14:39:25 crc kubenswrapper[4943]: I0307 14:39:25.155354 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:25 crc kubenswrapper[4943]: I0307 14:39:25.157437 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:25 crc kubenswrapper[4943]: I0307 14:39:25.157510 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:25 crc kubenswrapper[4943]: I0307 14:39:25.157525 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:25 crc kubenswrapper[4943]: I0307 14:39:25.157603 4943 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 07 14:39:25 crc kubenswrapper[4943]: I0307 14:39:25.426824 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 07 14:39:25 crc kubenswrapper[4943]: I0307 14:39:25.860154 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:25 crc kubenswrapper[4943]: I0307 14:39:25.861105 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:25 crc kubenswrapper[4943]: I0307 14:39:25.862663 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:25 crc kubenswrapper[4943]: I0307 14:39:25.862696 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:25 crc kubenswrapper[4943]: I0307 14:39:25.862737 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:25 crc kubenswrapper[4943]: I0307 14:39:25.862757 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:25 crc kubenswrapper[4943]: I0307 14:39:25.862704 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:25 crc kubenswrapper[4943]: I0307 14:39:25.862838 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:26 crc kubenswrapper[4943]: I0307 14:39:26.863590 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:26 crc kubenswrapper[4943]: I0307 14:39:26.865193 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:26 crc kubenswrapper[4943]: I0307 14:39:26.865251 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:26 crc kubenswrapper[4943]: I0307 14:39:26.865269 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:27 crc kubenswrapper[4943]: I0307 14:39:27.310626 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 07 14:39:27 crc kubenswrapper[4943]: I0307 14:39:27.310893 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:27 crc kubenswrapper[4943]: I0307 14:39:27.313042 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:27 crc kubenswrapper[4943]: I0307 14:39:27.313080 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:27 crc kubenswrapper[4943]: I0307 14:39:27.313096 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:27 crc kubenswrapper[4943]: I0307 14:39:27.335614 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 07 14:39:27 crc kubenswrapper[4943]: I0307 14:39:27.616971 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Mar 07 14:39:27 crc kubenswrapper[4943]: I0307 14:39:27.617583 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:27 crc kubenswrapper[4943]: I0307 14:39:27.619334 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:27 crc kubenswrapper[4943]: I0307 14:39:27.619522 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:27 crc kubenswrapper[4943]: I0307 14:39:27.619644 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:27 crc kubenswrapper[4943]: I0307 14:39:27.866065 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:27 crc kubenswrapper[4943]: I0307 14:39:27.868077 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:27 crc kubenswrapper[4943]: I0307 14:39:27.868140 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:27 crc kubenswrapper[4943]: I0307 14:39:27.868160 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:28 crc kubenswrapper[4943]: I0307 14:39:28.427797 4943 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 07 14:39:28 crc kubenswrapper[4943]: I0307 14:39:28.427913 4943 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 07 14:39:28 crc kubenswrapper[4943]: E0307 14:39:28.856443 4943 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 07 14:39:30 crc kubenswrapper[4943]: I0307 14:39:30.591246 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Mar 07 14:39:30 crc kubenswrapper[4943]: I0307 14:39:30.592483 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:30 crc kubenswrapper[4943]: I0307 14:39:30.594218 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:30 crc kubenswrapper[4943]: I0307 14:39:30.594278 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:30 crc kubenswrapper[4943]: I0307 14:39:30.594297 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:32 crc kubenswrapper[4943]: W0307 14:39:32.610376 4943 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Mar 07 14:39:32 crc kubenswrapper[4943]: I0307 14:39:32.610492 4943 trace.go:236] Trace[633720381]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (07-Mar-2026 14:39:22.608) (total time: 10002ms): Mar 07 14:39:32 crc kubenswrapper[4943]: Trace[633720381]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (14:39:32.610) Mar 07 14:39:32 crc kubenswrapper[4943]: Trace[633720381]: [10.002013307s] [10.002013307s] END Mar 07 14:39:32 crc kubenswrapper[4943]: E0307 14:39:32.610522 4943 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Mar 07 14:39:32 crc kubenswrapper[4943]: I0307 14:39:32.677006 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Mar 07 14:39:32 crc kubenswrapper[4943]: E0307 14:39:32.705667 4943 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": net/http: TLS handshake timeout" event="&Event{ObjectMeta:{crc.189a9606a8eeed5d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:18.671912285 +0000 UTC m=+0.624048783,LastTimestamp:2026-03-07 14:39:18.671912285 +0000 UTC m=+0.624048783,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:39:32 crc kubenswrapper[4943]: I0307 14:39:32.850907 4943 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:35564->192.168.126.11:17697: read: connection reset by peer" start-of-body= Mar 07 14:39:32 crc kubenswrapper[4943]: I0307 14:39:32.851045 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:35564->192.168.126.11:17697: read: connection reset by peer" Mar 07 14:39:33 crc kubenswrapper[4943]: W0307 14:39:33.116472 4943 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:33Z is after 2026-02-23T05:33:13Z Mar 07 14:39:33 crc kubenswrapper[4943]: E0307 14:39:33.116588 4943 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:33Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 07 14:39:33 crc kubenswrapper[4943]: E0307 14:39:33.118218 4943 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:33Z is after 2026-02-23T05:33:13Z" interval="6.4s" Mar 07 14:39:33 crc kubenswrapper[4943]: W0307 14:39:33.120519 4943 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:33Z is after 2026-02-23T05:33:13Z Mar 07 14:39:33 crc kubenswrapper[4943]: E0307 14:39:33.120588 4943 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:33Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 07 14:39:33 crc kubenswrapper[4943]: E0307 14:39:33.124855 4943 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:33Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 07 14:39:33 crc kubenswrapper[4943]: E0307 14:39:33.128272 4943 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:33Z is after 2026-02-23T05:33:13Z" node="crc" Mar 07 14:39:33 crc kubenswrapper[4943]: W0307 14:39:33.132358 4943 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:33Z is after 2026-02-23T05:33:13Z Mar 07 14:39:33 crc kubenswrapper[4943]: E0307 14:39:33.132461 4943 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:33Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 07 14:39:33 crc kubenswrapper[4943]: I0307 14:39:33.137442 4943 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 07 14:39:33 crc kubenswrapper[4943]: I0307 14:39:33.137507 4943 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Mar 07 14:39:33 crc kubenswrapper[4943]: I0307 14:39:33.144057 4943 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 07 14:39:33 crc kubenswrapper[4943]: I0307 14:39:33.144451 4943 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Mar 07 14:39:33 crc kubenswrapper[4943]: I0307 14:39:33.681259 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:33Z is after 2026-02-23T05:33:13Z Mar 07 14:39:33 crc kubenswrapper[4943]: I0307 14:39:33.885704 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 07 14:39:33 crc kubenswrapper[4943]: I0307 14:39:33.888381 4943 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="73a0d1456ae6ea55e508240342d50b14f7dd3856510fc756951c504786e4f49e" exitCode=255 Mar 07 14:39:33 crc kubenswrapper[4943]: I0307 14:39:33.888456 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"73a0d1456ae6ea55e508240342d50b14f7dd3856510fc756951c504786e4f49e"} Mar 07 14:39:33 crc kubenswrapper[4943]: I0307 14:39:33.888632 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:33 crc kubenswrapper[4943]: I0307 14:39:33.889512 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:33 crc kubenswrapper[4943]: I0307 14:39:33.889690 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:33 crc kubenswrapper[4943]: I0307 14:39:33.889831 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:33 crc kubenswrapper[4943]: I0307 14:39:33.890729 4943 scope.go:117] "RemoveContainer" containerID="73a0d1456ae6ea55e508240342d50b14f7dd3856510fc756951c504786e4f49e" Mar 07 14:39:34 crc kubenswrapper[4943]: I0307 14:39:34.680496 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:34Z is after 2026-02-23T05:33:13Z Mar 07 14:39:34 crc kubenswrapper[4943]: I0307 14:39:34.895554 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 07 14:39:34 crc kubenswrapper[4943]: I0307 14:39:34.899291 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"98c50a697e78bbda717a3e300f17572873af5a3241f3f20bb059d6737cb550fe"} Mar 07 14:39:34 crc kubenswrapper[4943]: I0307 14:39:34.899540 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:34 crc kubenswrapper[4943]: I0307 14:39:34.900801 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:34 crc kubenswrapper[4943]: I0307 14:39:34.901028 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:34 crc kubenswrapper[4943]: I0307 14:39:34.901182 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:34 crc kubenswrapper[4943]: I0307 14:39:34.916304 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 07 14:39:34 crc kubenswrapper[4943]: I0307 14:39:34.916693 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:34 crc kubenswrapper[4943]: I0307 14:39:34.922383 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:34 crc kubenswrapper[4943]: I0307 14:39:34.922456 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:34 crc kubenswrapper[4943]: I0307 14:39:34.922487 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:35 crc kubenswrapper[4943]: I0307 14:39:35.680087 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:35Z is after 2026-02-23T05:33:13Z Mar 07 14:39:35 crc kubenswrapper[4943]: I0307 14:39:35.905415 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 07 14:39:35 crc kubenswrapper[4943]: I0307 14:39:35.906375 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 07 14:39:35 crc kubenswrapper[4943]: I0307 14:39:35.909069 4943 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="98c50a697e78bbda717a3e300f17572873af5a3241f3f20bb059d6737cb550fe" exitCode=255 Mar 07 14:39:35 crc kubenswrapper[4943]: I0307 14:39:35.909146 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"98c50a697e78bbda717a3e300f17572873af5a3241f3f20bb059d6737cb550fe"} Mar 07 14:39:35 crc kubenswrapper[4943]: I0307 14:39:35.909232 4943 scope.go:117] "RemoveContainer" containerID="73a0d1456ae6ea55e508240342d50b14f7dd3856510fc756951c504786e4f49e" Mar 07 14:39:35 crc kubenswrapper[4943]: I0307 14:39:35.909382 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:35 crc kubenswrapper[4943]: I0307 14:39:35.910860 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:35 crc kubenswrapper[4943]: I0307 14:39:35.911443 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:35 crc kubenswrapper[4943]: I0307 14:39:35.911471 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:35 crc kubenswrapper[4943]: I0307 14:39:35.915263 4943 scope.go:117] "RemoveContainer" containerID="98c50a697e78bbda717a3e300f17572873af5a3241f3f20bb059d6737cb550fe" Mar 07 14:39:35 crc kubenswrapper[4943]: E0307 14:39:35.915685 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 07 14:39:36 crc kubenswrapper[4943]: I0307 14:39:36.680367 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:36Z is after 2026-02-23T05:33:13Z Mar 07 14:39:36 crc kubenswrapper[4943]: I0307 14:39:36.915855 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 07 14:39:37 crc kubenswrapper[4943]: I0307 14:39:37.317752 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 07 14:39:37 crc kubenswrapper[4943]: I0307 14:39:37.318005 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:37 crc kubenswrapper[4943]: I0307 14:39:37.320001 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:37 crc kubenswrapper[4943]: I0307 14:39:37.320077 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:37 crc kubenswrapper[4943]: I0307 14:39:37.320090 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:37 crc kubenswrapper[4943]: I0307 14:39:37.321121 4943 scope.go:117] "RemoveContainer" containerID="98c50a697e78bbda717a3e300f17572873af5a3241f3f20bb059d6737cb550fe" Mar 07 14:39:37 crc kubenswrapper[4943]: E0307 14:39:37.321419 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 07 14:39:37 crc kubenswrapper[4943]: I0307 14:39:37.324532 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 07 14:39:37 crc kubenswrapper[4943]: I0307 14:39:37.682820 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:37Z is after 2026-02-23T05:33:13Z Mar 07 14:39:37 crc kubenswrapper[4943]: I0307 14:39:37.924755 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:37 crc kubenswrapper[4943]: I0307 14:39:37.926384 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:37 crc kubenswrapper[4943]: I0307 14:39:37.926452 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:37 crc kubenswrapper[4943]: I0307 14:39:37.926472 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:37 crc kubenswrapper[4943]: I0307 14:39:37.927609 4943 scope.go:117] "RemoveContainer" containerID="98c50a697e78bbda717a3e300f17572873af5a3241f3f20bb059d6737cb550fe" Mar 07 14:39:37 crc kubenswrapper[4943]: E0307 14:39:37.927962 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 07 14:39:38 crc kubenswrapper[4943]: W0307 14:39:38.263150 4943 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:38Z is after 2026-02-23T05:33:13Z Mar 07 14:39:38 crc kubenswrapper[4943]: E0307 14:39:38.263234 4943 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:38Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 07 14:39:38 crc kubenswrapper[4943]: I0307 14:39:38.428342 4943 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 07 14:39:38 crc kubenswrapper[4943]: I0307 14:39:38.428438 4943 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 07 14:39:38 crc kubenswrapper[4943]: I0307 14:39:38.680883 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:38Z is after 2026-02-23T05:33:13Z Mar 07 14:39:38 crc kubenswrapper[4943]: E0307 14:39:38.858226 4943 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 07 14:39:39 crc kubenswrapper[4943]: E0307 14:39:39.525541 4943 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:39Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 07 14:39:39 crc kubenswrapper[4943]: I0307 14:39:39.528788 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:39 crc kubenswrapper[4943]: I0307 14:39:39.530729 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:39 crc kubenswrapper[4943]: I0307 14:39:39.531165 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:39 crc kubenswrapper[4943]: I0307 14:39:39.531320 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:39 crc kubenswrapper[4943]: I0307 14:39:39.531493 4943 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 07 14:39:39 crc kubenswrapper[4943]: E0307 14:39:39.537446 4943 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:39Z is after 2026-02-23T05:33:13Z" node="crc" Mar 07 14:39:39 crc kubenswrapper[4943]: I0307 14:39:39.680826 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:39Z is after 2026-02-23T05:33:13Z Mar 07 14:39:40 crc kubenswrapper[4943]: W0307 14:39:40.568530 4943 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:40Z is after 2026-02-23T05:33:13Z Mar 07 14:39:40 crc kubenswrapper[4943]: E0307 14:39:40.568675 4943 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:40Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 07 14:39:40 crc kubenswrapper[4943]: I0307 14:39:40.627995 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Mar 07 14:39:40 crc kubenswrapper[4943]: I0307 14:39:40.628316 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:40 crc kubenswrapper[4943]: I0307 14:39:40.630037 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:40 crc kubenswrapper[4943]: I0307 14:39:40.630096 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:40 crc kubenswrapper[4943]: I0307 14:39:40.630117 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:40 crc kubenswrapper[4943]: I0307 14:39:40.649021 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Mar 07 14:39:40 crc kubenswrapper[4943]: I0307 14:39:40.681049 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:40Z is after 2026-02-23T05:33:13Z Mar 07 14:39:40 crc kubenswrapper[4943]: I0307 14:39:40.933035 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:40 crc kubenswrapper[4943]: I0307 14:39:40.934546 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:40 crc kubenswrapper[4943]: I0307 14:39:40.934611 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:40 crc kubenswrapper[4943]: I0307 14:39:40.934632 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:41 crc kubenswrapper[4943]: W0307 14:39:41.179464 4943 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:41Z is after 2026-02-23T05:33:13Z Mar 07 14:39:41 crc kubenswrapper[4943]: E0307 14:39:41.179569 4943 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:41Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 07 14:39:41 crc kubenswrapper[4943]: I0307 14:39:41.360278 4943 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 07 14:39:41 crc kubenswrapper[4943]: E0307 14:39:41.366237 4943 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:41Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 07 14:39:41 crc kubenswrapper[4943]: I0307 14:39:41.681220 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:41Z is after 2026-02-23T05:33:13Z Mar 07 14:39:41 crc kubenswrapper[4943]: I0307 14:39:41.941126 4943 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 07 14:39:41 crc kubenswrapper[4943]: I0307 14:39:41.941370 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:41 crc kubenswrapper[4943]: I0307 14:39:41.942622 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:41 crc kubenswrapper[4943]: I0307 14:39:41.942675 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:41 crc kubenswrapper[4943]: I0307 14:39:41.942694 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:41 crc kubenswrapper[4943]: I0307 14:39:41.943579 4943 scope.go:117] "RemoveContainer" containerID="98c50a697e78bbda717a3e300f17572873af5a3241f3f20bb059d6737cb550fe" Mar 07 14:39:41 crc kubenswrapper[4943]: E0307 14:39:41.943867 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 07 14:39:42 crc kubenswrapper[4943]: I0307 14:39:42.681057 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:42Z is after 2026-02-23T05:33:13Z Mar 07 14:39:42 crc kubenswrapper[4943]: E0307 14:39:42.711466 4943 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:42Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189a9606a8eeed5d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:18.671912285 +0000 UTC m=+0.624048783,LastTimestamp:2026-03-07 14:39:18.671912285 +0000 UTC m=+0.624048783,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:39:42 crc kubenswrapper[4943]: W0307 14:39:42.746448 4943 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:42Z is after 2026-02-23T05:33:13Z Mar 07 14:39:42 crc kubenswrapper[4943]: E0307 14:39:42.746895 4943 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:42Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 07 14:39:43 crc kubenswrapper[4943]: I0307 14:39:43.678550 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:43Z is after 2026-02-23T05:33:13Z Mar 07 14:39:44 crc kubenswrapper[4943]: I0307 14:39:44.033525 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 07 14:39:44 crc kubenswrapper[4943]: I0307 14:39:44.033784 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:44 crc kubenswrapper[4943]: I0307 14:39:44.035407 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:44 crc kubenswrapper[4943]: I0307 14:39:44.035477 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:44 crc kubenswrapper[4943]: I0307 14:39:44.035503 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:44 crc kubenswrapper[4943]: I0307 14:39:44.036424 4943 scope.go:117] "RemoveContainer" containerID="98c50a697e78bbda717a3e300f17572873af5a3241f3f20bb059d6737cb550fe" Mar 07 14:39:44 crc kubenswrapper[4943]: E0307 14:39:44.036739 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 07 14:39:44 crc kubenswrapper[4943]: I0307 14:39:44.680607 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:44Z is after 2026-02-23T05:33:13Z Mar 07 14:39:45 crc kubenswrapper[4943]: I0307 14:39:45.680698 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:45Z is after 2026-02-23T05:33:13Z Mar 07 14:39:46 crc kubenswrapper[4943]: E0307 14:39:46.532579 4943 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:46Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 07 14:39:46 crc kubenswrapper[4943]: I0307 14:39:46.537911 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:46 crc kubenswrapper[4943]: I0307 14:39:46.539846 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:46 crc kubenswrapper[4943]: I0307 14:39:46.539976 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:46 crc kubenswrapper[4943]: I0307 14:39:46.540005 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:46 crc kubenswrapper[4943]: I0307 14:39:46.540056 4943 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 07 14:39:46 crc kubenswrapper[4943]: E0307 14:39:46.545151 4943 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:46Z is after 2026-02-23T05:33:13Z" node="crc" Mar 07 14:39:46 crc kubenswrapper[4943]: I0307 14:39:46.681246 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:46Z is after 2026-02-23T05:33:13Z Mar 07 14:39:47 crc kubenswrapper[4943]: I0307 14:39:47.680498 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:47Z is after 2026-02-23T05:33:13Z Mar 07 14:39:47 crc kubenswrapper[4943]: W0307 14:39:47.732796 4943 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:47Z is after 2026-02-23T05:33:13Z Mar 07 14:39:47 crc kubenswrapper[4943]: E0307 14:39:47.732917 4943 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:47Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 07 14:39:48 crc kubenswrapper[4943]: I0307 14:39:48.427956 4943 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 07 14:39:48 crc kubenswrapper[4943]: I0307 14:39:48.428043 4943 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 07 14:39:48 crc kubenswrapper[4943]: I0307 14:39:48.428132 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 07 14:39:48 crc kubenswrapper[4943]: I0307 14:39:48.428342 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:48 crc kubenswrapper[4943]: I0307 14:39:48.429848 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:48 crc kubenswrapper[4943]: I0307 14:39:48.429922 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:48 crc kubenswrapper[4943]: I0307 14:39:48.429969 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:48 crc kubenswrapper[4943]: I0307 14:39:48.430736 4943 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cluster-policy-controller" containerStatusID={"Type":"cri-o","ID":"0a516a1a0e24f9de9a1038d966ac4195f22bb34969d2c55081256fbacc5379ce"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container cluster-policy-controller failed startup probe, will be restarted" Mar 07 14:39:48 crc kubenswrapper[4943]: I0307 14:39:48.431034 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" containerID="cri-o://0a516a1a0e24f9de9a1038d966ac4195f22bb34969d2c55081256fbacc5379ce" gracePeriod=30 Mar 07 14:39:48 crc kubenswrapper[4943]: I0307 14:39:48.680148 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:48Z is after 2026-02-23T05:33:13Z Mar 07 14:39:48 crc kubenswrapper[4943]: E0307 14:39:48.859758 4943 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 07 14:39:48 crc kubenswrapper[4943]: I0307 14:39:48.964218 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 07 14:39:48 crc kubenswrapper[4943]: I0307 14:39:48.965579 4943 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="0a516a1a0e24f9de9a1038d966ac4195f22bb34969d2c55081256fbacc5379ce" exitCode=255 Mar 07 14:39:48 crc kubenswrapper[4943]: I0307 14:39:48.965656 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"0a516a1a0e24f9de9a1038d966ac4195f22bb34969d2c55081256fbacc5379ce"} Mar 07 14:39:48 crc kubenswrapper[4943]: I0307 14:39:48.965706 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"8073fd51bfd3dc6747801540643a50f2314fd71db33698d0d1293c8cd5d7ce2d"} Mar 07 14:39:48 crc kubenswrapper[4943]: I0307 14:39:48.965866 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:48 crc kubenswrapper[4943]: I0307 14:39:48.967248 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:48 crc kubenswrapper[4943]: I0307 14:39:48.967296 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:48 crc kubenswrapper[4943]: I0307 14:39:48.967317 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:49 crc kubenswrapper[4943]: I0307 14:39:49.678339 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:49Z is after 2026-02-23T05:33:13Z Mar 07 14:39:50 crc kubenswrapper[4943]: I0307 14:39:50.679818 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:50Z is after 2026-02-23T05:33:13Z Mar 07 14:39:51 crc kubenswrapper[4943]: I0307 14:39:51.660611 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 07 14:39:51 crc kubenswrapper[4943]: I0307 14:39:51.660846 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:51 crc kubenswrapper[4943]: I0307 14:39:51.662360 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:51 crc kubenswrapper[4943]: I0307 14:39:51.662465 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:51 crc kubenswrapper[4943]: I0307 14:39:51.662486 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:51 crc kubenswrapper[4943]: I0307 14:39:51.681007 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:51Z is after 2026-02-23T05:33:13Z Mar 07 14:39:52 crc kubenswrapper[4943]: I0307 14:39:52.680082 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:52Z is after 2026-02-23T05:33:13Z Mar 07 14:39:52 crc kubenswrapper[4943]: E0307 14:39:52.716874 4943 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:52Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189a9606a8eeed5d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:18.671912285 +0000 UTC m=+0.624048783,LastTimestamp:2026-03-07 14:39:18.671912285 +0000 UTC m=+0.624048783,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:39:53 crc kubenswrapper[4943]: E0307 14:39:53.537866 4943 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:53Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 07 14:39:53 crc kubenswrapper[4943]: I0307 14:39:53.546055 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:53 crc kubenswrapper[4943]: I0307 14:39:53.547825 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:53 crc kubenswrapper[4943]: I0307 14:39:53.547875 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:53 crc kubenswrapper[4943]: I0307 14:39:53.547896 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:53 crc kubenswrapper[4943]: I0307 14:39:53.547966 4943 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 07 14:39:53 crc kubenswrapper[4943]: E0307 14:39:53.552792 4943 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:53Z is after 2026-02-23T05:33:13Z" node="crc" Mar 07 14:39:53 crc kubenswrapper[4943]: I0307 14:39:53.680254 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:53Z is after 2026-02-23T05:33:13Z Mar 07 14:39:54 crc kubenswrapper[4943]: W0307 14:39:54.169099 4943 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:54Z is after 2026-02-23T05:33:13Z Mar 07 14:39:54 crc kubenswrapper[4943]: E0307 14:39:54.169203 4943 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:54Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 07 14:39:54 crc kubenswrapper[4943]: I0307 14:39:54.680836 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:54Z is after 2026-02-23T05:33:13Z Mar 07 14:39:55 crc kubenswrapper[4943]: I0307 14:39:55.427055 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 07 14:39:55 crc kubenswrapper[4943]: I0307 14:39:55.427539 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:55 crc kubenswrapper[4943]: I0307 14:39:55.429126 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:55 crc kubenswrapper[4943]: I0307 14:39:55.429190 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:55 crc kubenswrapper[4943]: I0307 14:39:55.429210 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:55 crc kubenswrapper[4943]: I0307 14:39:55.681544 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:55Z is after 2026-02-23T05:33:13Z Mar 07 14:39:56 crc kubenswrapper[4943]: I0307 14:39:56.680645 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:56Z is after 2026-02-23T05:33:13Z Mar 07 14:39:57 crc kubenswrapper[4943]: I0307 14:39:57.426815 4943 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 07 14:39:57 crc kubenswrapper[4943]: E0307 14:39:57.432575 4943 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:57Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 07 14:39:57 crc kubenswrapper[4943]: E0307 14:39:57.433857 4943 certificate_manager.go:440] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Reached backoff limit, still unable to rotate certs: timed out waiting for the condition" logger="UnhandledError" Mar 07 14:39:57 crc kubenswrapper[4943]: I0307 14:39:57.680969 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:57Z is after 2026-02-23T05:33:13Z Mar 07 14:39:57 crc kubenswrapper[4943]: I0307 14:39:57.754578 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:57 crc kubenswrapper[4943]: I0307 14:39:57.756111 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:57 crc kubenswrapper[4943]: I0307 14:39:57.756163 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:57 crc kubenswrapper[4943]: I0307 14:39:57.756182 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:57 crc kubenswrapper[4943]: I0307 14:39:57.757013 4943 scope.go:117] "RemoveContainer" containerID="98c50a697e78bbda717a3e300f17572873af5a3241f3f20bb059d6737cb550fe" Mar 07 14:39:58 crc kubenswrapper[4943]: I0307 14:39:58.427503 4943 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 07 14:39:58 crc kubenswrapper[4943]: I0307 14:39:58.427578 4943 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 07 14:39:58 crc kubenswrapper[4943]: I0307 14:39:58.680564 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:58Z is after 2026-02-23T05:33:13Z Mar 07 14:39:58 crc kubenswrapper[4943]: E0307 14:39:58.860718 4943 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 07 14:39:58 crc kubenswrapper[4943]: I0307 14:39:58.998683 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 07 14:39:58 crc kubenswrapper[4943]: I0307 14:39:58.999774 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 07 14:39:59 crc kubenswrapper[4943]: I0307 14:39:59.003328 4943 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="a25d75089d6ba494eeab9670dc43f7d2ed9a201acee535fe805e39e80ceb0cb5" exitCode=255 Mar 07 14:39:59 crc kubenswrapper[4943]: I0307 14:39:59.003390 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"a25d75089d6ba494eeab9670dc43f7d2ed9a201acee535fe805e39e80ceb0cb5"} Mar 07 14:39:59 crc kubenswrapper[4943]: I0307 14:39:59.003455 4943 scope.go:117] "RemoveContainer" containerID="98c50a697e78bbda717a3e300f17572873af5a3241f3f20bb059d6737cb550fe" Mar 07 14:39:59 crc kubenswrapper[4943]: I0307 14:39:59.003620 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:39:59 crc kubenswrapper[4943]: I0307 14:39:59.004973 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:39:59 crc kubenswrapper[4943]: I0307 14:39:59.005039 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:39:59 crc kubenswrapper[4943]: I0307 14:39:59.005059 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:39:59 crc kubenswrapper[4943]: I0307 14:39:59.006058 4943 scope.go:117] "RemoveContainer" containerID="a25d75089d6ba494eeab9670dc43f7d2ed9a201acee535fe805e39e80ceb0cb5" Mar 07 14:39:59 crc kubenswrapper[4943]: E0307 14:39:59.006487 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 07 14:39:59 crc kubenswrapper[4943]: I0307 14:39:59.680405 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:39:59Z is after 2026-02-23T05:33:13Z Mar 07 14:40:00 crc kubenswrapper[4943]: I0307 14:40:00.013851 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 07 14:40:00 crc kubenswrapper[4943]: E0307 14:40:00.544582 4943 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:40:00Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 07 14:40:00 crc kubenswrapper[4943]: I0307 14:40:00.553807 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:40:00 crc kubenswrapper[4943]: I0307 14:40:00.555657 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:00 crc kubenswrapper[4943]: I0307 14:40:00.555718 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:00 crc kubenswrapper[4943]: I0307 14:40:00.555738 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:00 crc kubenswrapper[4943]: I0307 14:40:00.555778 4943 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 07 14:40:00 crc kubenswrapper[4943]: E0307 14:40:00.560575 4943 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:40:00Z is after 2026-02-23T05:33:13Z" node="crc" Mar 07 14:40:00 crc kubenswrapper[4943]: I0307 14:40:00.681998 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:40:00Z is after 2026-02-23T05:33:13Z Mar 07 14:40:01 crc kubenswrapper[4943]: I0307 14:40:01.680840 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:40:01Z is after 2026-02-23T05:33:13Z Mar 07 14:40:01 crc kubenswrapper[4943]: I0307 14:40:01.941385 4943 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 07 14:40:01 crc kubenswrapper[4943]: I0307 14:40:01.941628 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:40:01 crc kubenswrapper[4943]: I0307 14:40:01.943549 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:01 crc kubenswrapper[4943]: I0307 14:40:01.943613 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:01 crc kubenswrapper[4943]: I0307 14:40:01.943633 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:01 crc kubenswrapper[4943]: I0307 14:40:01.945014 4943 scope.go:117] "RemoveContainer" containerID="a25d75089d6ba494eeab9670dc43f7d2ed9a201acee535fe805e39e80ceb0cb5" Mar 07 14:40:01 crc kubenswrapper[4943]: E0307 14:40:01.945552 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 07 14:40:02 crc kubenswrapper[4943]: I0307 14:40:02.680267 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:40:02Z is after 2026-02-23T05:33:13Z Mar 07 14:40:02 crc kubenswrapper[4943]: E0307 14:40:02.723006 4943 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:40:02Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189a9606a8eeed5d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:18.671912285 +0000 UTC m=+0.624048783,LastTimestamp:2026-03-07 14:39:18.671912285 +0000 UTC m=+0.624048783,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:03 crc kubenswrapper[4943]: I0307 14:40:03.678138 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:40:03Z is after 2026-02-23T05:33:13Z Mar 07 14:40:03 crc kubenswrapper[4943]: W0307 14:40:03.710636 4943 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:40:03Z is after 2026-02-23T05:33:13Z Mar 07 14:40:03 crc kubenswrapper[4943]: E0307 14:40:03.710741 4943 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:40:03Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 07 14:40:04 crc kubenswrapper[4943]: I0307 14:40:04.033806 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 07 14:40:04 crc kubenswrapper[4943]: I0307 14:40:04.035022 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:40:04 crc kubenswrapper[4943]: I0307 14:40:04.036438 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:04 crc kubenswrapper[4943]: I0307 14:40:04.036709 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:04 crc kubenswrapper[4943]: I0307 14:40:04.036903 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:04 crc kubenswrapper[4943]: I0307 14:40:04.038003 4943 scope.go:117] "RemoveContainer" containerID="a25d75089d6ba494eeab9670dc43f7d2ed9a201acee535fe805e39e80ceb0cb5" Mar 07 14:40:04 crc kubenswrapper[4943]: E0307 14:40:04.038559 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 07 14:40:04 crc kubenswrapper[4943]: W0307 14:40:04.625332 4943 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:40:04Z is after 2026-02-23T05:33:13Z Mar 07 14:40:04 crc kubenswrapper[4943]: E0307 14:40:04.625732 4943 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:40:04Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 07 14:40:04 crc kubenswrapper[4943]: I0307 14:40:04.680418 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:40:04Z is after 2026-02-23T05:33:13Z Mar 07 14:40:05 crc kubenswrapper[4943]: I0307 14:40:05.678864 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:40:05Z is after 2026-02-23T05:33:13Z Mar 07 14:40:06 crc kubenswrapper[4943]: I0307 14:40:06.680665 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:40:06Z is after 2026-02-23T05:33:13Z Mar 07 14:40:07 crc kubenswrapper[4943]: E0307 14:40:07.550767 4943 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:40:07Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 07 14:40:07 crc kubenswrapper[4943]: I0307 14:40:07.560964 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:40:07 crc kubenswrapper[4943]: I0307 14:40:07.562888 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:07 crc kubenswrapper[4943]: I0307 14:40:07.562986 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:07 crc kubenswrapper[4943]: I0307 14:40:07.563007 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:07 crc kubenswrapper[4943]: I0307 14:40:07.563048 4943 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 07 14:40:07 crc kubenswrapper[4943]: E0307 14:40:07.568103 4943 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:40:07Z is after 2026-02-23T05:33:13Z" node="crc" Mar 07 14:40:07 crc kubenswrapper[4943]: I0307 14:40:07.680423 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:40:07Z is after 2026-02-23T05:33:13Z Mar 07 14:40:08 crc kubenswrapper[4943]: I0307 14:40:08.426822 4943 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 07 14:40:08 crc kubenswrapper[4943]: I0307 14:40:08.426920 4943 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 07 14:40:08 crc kubenswrapper[4943]: I0307 14:40:08.681985 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:40:08Z is after 2026-02-23T05:33:13Z Mar 07 14:40:08 crc kubenswrapper[4943]: E0307 14:40:08.862826 4943 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 07 14:40:09 crc kubenswrapper[4943]: I0307 14:40:09.679449 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:40:09Z is after 2026-02-23T05:33:13Z Mar 07 14:40:10 crc kubenswrapper[4943]: I0307 14:40:10.681421 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:40:10Z is after 2026-02-23T05:33:13Z Mar 07 14:40:11 crc kubenswrapper[4943]: I0307 14:40:11.680317 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:40:11Z is after 2026-02-23T05:33:13Z Mar 07 14:40:12 crc kubenswrapper[4943]: I0307 14:40:12.680823 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:40:12Z is after 2026-02-23T05:33:13Z Mar 07 14:40:12 crc kubenswrapper[4943]: E0307 14:40:12.728732 4943 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:40:12Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189a9606a8eeed5d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:18.671912285 +0000 UTC m=+0.624048783,LastTimestamp:2026-03-07 14:39:18.671912285 +0000 UTC m=+0.624048783,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:12 crc kubenswrapper[4943]: W0307 14:40:12.789678 4943 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:40:12Z is after 2026-02-23T05:33:13Z Mar 07 14:40:12 crc kubenswrapper[4943]: E0307 14:40:12.790032 4943 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:40:12Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 07 14:40:13 crc kubenswrapper[4943]: I0307 14:40:13.506494 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 07 14:40:13 crc kubenswrapper[4943]: I0307 14:40:13.506685 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:40:13 crc kubenswrapper[4943]: I0307 14:40:13.508985 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:13 crc kubenswrapper[4943]: I0307 14:40:13.509232 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:13 crc kubenswrapper[4943]: I0307 14:40:13.509398 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:13 crc kubenswrapper[4943]: I0307 14:40:13.681254 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:40:13Z is after 2026-02-23T05:33:13Z Mar 07 14:40:14 crc kubenswrapper[4943]: E0307 14:40:14.557783 4943 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:40:14Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 07 14:40:14 crc kubenswrapper[4943]: I0307 14:40:14.568781 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:40:14 crc kubenswrapper[4943]: I0307 14:40:14.570652 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:14 crc kubenswrapper[4943]: I0307 14:40:14.570864 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:14 crc kubenswrapper[4943]: I0307 14:40:14.571053 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:14 crc kubenswrapper[4943]: I0307 14:40:14.571224 4943 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 07 14:40:14 crc kubenswrapper[4943]: E0307 14:40:14.576140 4943 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:40:14Z is after 2026-02-23T05:33:13Z" node="crc" Mar 07 14:40:14 crc kubenswrapper[4943]: I0307 14:40:14.680223 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:40:14Z is after 2026-02-23T05:33:13Z Mar 07 14:40:15 crc kubenswrapper[4943]: I0307 14:40:15.680392 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:40:15Z is after 2026-02-23T05:33:13Z Mar 07 14:40:16 crc kubenswrapper[4943]: I0307 14:40:16.680083 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:40:16Z is after 2026-02-23T05:33:13Z Mar 07 14:40:16 crc kubenswrapper[4943]: I0307 14:40:16.755006 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:40:16 crc kubenswrapper[4943]: I0307 14:40:16.756505 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:16 crc kubenswrapper[4943]: I0307 14:40:16.756576 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:16 crc kubenswrapper[4943]: I0307 14:40:16.756593 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:16 crc kubenswrapper[4943]: I0307 14:40:16.757539 4943 scope.go:117] "RemoveContainer" containerID="a25d75089d6ba494eeab9670dc43f7d2ed9a201acee535fe805e39e80ceb0cb5" Mar 07 14:40:16 crc kubenswrapper[4943]: E0307 14:40:16.757860 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 07 14:40:17 crc kubenswrapper[4943]: I0307 14:40:17.680557 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:40:17Z is after 2026-02-23T05:33:13Z Mar 07 14:40:18 crc kubenswrapper[4943]: I0307 14:40:18.427411 4943 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 07 14:40:18 crc kubenswrapper[4943]: I0307 14:40:18.427561 4943 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 07 14:40:18 crc kubenswrapper[4943]: I0307 14:40:18.427667 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 07 14:40:18 crc kubenswrapper[4943]: I0307 14:40:18.427970 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:40:18 crc kubenswrapper[4943]: I0307 14:40:18.429691 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:18 crc kubenswrapper[4943]: I0307 14:40:18.429792 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:18 crc kubenswrapper[4943]: I0307 14:40:18.429816 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:18 crc kubenswrapper[4943]: I0307 14:40:18.430745 4943 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cluster-policy-controller" containerStatusID={"Type":"cri-o","ID":"8073fd51bfd3dc6747801540643a50f2314fd71db33698d0d1293c8cd5d7ce2d"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container cluster-policy-controller failed startup probe, will be restarted" Mar 07 14:40:18 crc kubenswrapper[4943]: I0307 14:40:18.430894 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" containerID="cri-o://8073fd51bfd3dc6747801540643a50f2314fd71db33698d0d1293c8cd5d7ce2d" gracePeriod=30 Mar 07 14:40:18 crc kubenswrapper[4943]: I0307 14:40:18.681756 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 07 14:40:18 crc kubenswrapper[4943]: E0307 14:40:18.863364 4943 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 07 14:40:19 crc kubenswrapper[4943]: I0307 14:40:19.080289 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/1.log" Mar 07 14:40:19 crc kubenswrapper[4943]: I0307 14:40:19.082227 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 07 14:40:19 crc kubenswrapper[4943]: I0307 14:40:19.082843 4943 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="8073fd51bfd3dc6747801540643a50f2314fd71db33698d0d1293c8cd5d7ce2d" exitCode=255 Mar 07 14:40:19 crc kubenswrapper[4943]: I0307 14:40:19.082983 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"8073fd51bfd3dc6747801540643a50f2314fd71db33698d0d1293c8cd5d7ce2d"} Mar 07 14:40:19 crc kubenswrapper[4943]: I0307 14:40:19.083098 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a0041d3ab720ecf5656e6380ea4e6145afa51663a438e1dd17d031bb4162f540"} Mar 07 14:40:19 crc kubenswrapper[4943]: I0307 14:40:19.083187 4943 scope.go:117] "RemoveContainer" containerID="0a516a1a0e24f9de9a1038d966ac4195f22bb34969d2c55081256fbacc5379ce" Mar 07 14:40:19 crc kubenswrapper[4943]: I0307 14:40:19.083419 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:40:19 crc kubenswrapper[4943]: I0307 14:40:19.084889 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:19 crc kubenswrapper[4943]: I0307 14:40:19.085010 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:19 crc kubenswrapper[4943]: I0307 14:40:19.085087 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:19 crc kubenswrapper[4943]: I0307 14:40:19.682466 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 07 14:40:20 crc kubenswrapper[4943]: I0307 14:40:20.089468 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/1.log" Mar 07 14:40:20 crc kubenswrapper[4943]: I0307 14:40:20.683642 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 07 14:40:21 crc kubenswrapper[4943]: E0307 14:40:21.567403 4943 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 07 14:40:21 crc kubenswrapper[4943]: I0307 14:40:21.576599 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:40:21 crc kubenswrapper[4943]: I0307 14:40:21.578405 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:21 crc kubenswrapper[4943]: I0307 14:40:21.578549 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:21 crc kubenswrapper[4943]: I0307 14:40:21.578636 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:21 crc kubenswrapper[4943]: I0307 14:40:21.578742 4943 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 07 14:40:21 crc kubenswrapper[4943]: E0307 14:40:21.586476 4943 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 07 14:40:21 crc kubenswrapper[4943]: I0307 14:40:21.660515 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 07 14:40:21 crc kubenswrapper[4943]: I0307 14:40:21.660682 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:40:21 crc kubenswrapper[4943]: I0307 14:40:21.661872 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:21 crc kubenswrapper[4943]: I0307 14:40:21.661921 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:21 crc kubenswrapper[4943]: I0307 14:40:21.661964 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:21 crc kubenswrapper[4943]: I0307 14:40:21.682022 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 07 14:40:22 crc kubenswrapper[4943]: I0307 14:40:22.681336 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 07 14:40:22 crc kubenswrapper[4943]: E0307 14:40:22.737708 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189a9606a8eeed5d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:18.671912285 +0000 UTC m=+0.624048783,LastTimestamp:2026-03-07 14:39:18.671912285 +0000 UTC m=+0.624048783,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:22 crc kubenswrapper[4943]: E0307 14:40:22.743502 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189a9606ace0181e default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:18.738049054 +0000 UTC m=+0.690185562,LastTimestamp:2026-03-07 14:39:18.738049054 +0000 UTC m=+0.690185562,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:22 crc kubenswrapper[4943]: E0307 14:40:22.751119 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189a9606ace07e4f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:18.738075215 +0000 UTC m=+0.690211723,LastTimestamp:2026-03-07 14:39:18.738075215 +0000 UTC m=+0.690211723,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:22 crc kubenswrapper[4943]: E0307 14:40:22.757544 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189a9606ace0a5af default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:18.738085295 +0000 UTC m=+0.690221803,LastTimestamp:2026-03-07 14:39:18.738085295 +0000 UTC m=+0.690221803,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:22 crc kubenswrapper[4943]: E0307 14:40:22.764491 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189a9606b3341381 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeAllocatableEnforced,Message:Updated Node Allocatable limit across pods,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:18.844216193 +0000 UTC m=+0.796352721,LastTimestamp:2026-03-07 14:39:18.844216193 +0000 UTC m=+0.796352721,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:22 crc kubenswrapper[4943]: E0307 14:40:22.770279 4943 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189a9606ace0181e\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189a9606ace0181e default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:18.738049054 +0000 UTC m=+0.690185562,LastTimestamp:2026-03-07 14:39:18.856261307 +0000 UTC m=+0.808397825,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:22 crc kubenswrapper[4943]: E0307 14:40:22.777296 4943 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189a9606ace07e4f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189a9606ace07e4f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:18.738075215 +0000 UTC m=+0.690211723,LastTimestamp:2026-03-07 14:39:18.856293418 +0000 UTC m=+0.808429936,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:22 crc kubenswrapper[4943]: E0307 14:40:22.783304 4943 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189a9606ace0a5af\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189a9606ace0a5af default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:18.738085295 +0000 UTC m=+0.690221803,LastTimestamp:2026-03-07 14:39:18.856308138 +0000 UTC m=+0.808444656,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:22 crc kubenswrapper[4943]: E0307 14:40:22.787848 4943 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189a9606ace0181e\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189a9606ace0181e default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:18.738049054 +0000 UTC m=+0.690185562,LastTimestamp:2026-03-07 14:39:18.85830053 +0000 UTC m=+0.810437038,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:22 crc kubenswrapper[4943]: E0307 14:40:22.791870 4943 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189a9606ace07e4f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189a9606ace07e4f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:18.738075215 +0000 UTC m=+0.690211723,LastTimestamp:2026-03-07 14:39:18.85831535 +0000 UTC m=+0.810451858,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:22 crc kubenswrapper[4943]: E0307 14:40:22.796575 4943 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189a9606ace0a5af\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189a9606ace0a5af default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:18.738085295 +0000 UTC m=+0.690221803,LastTimestamp:2026-03-07 14:39:18.85832665 +0000 UTC m=+0.810463158,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:22 crc kubenswrapper[4943]: E0307 14:40:22.803284 4943 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189a9606ace0181e\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189a9606ace0181e default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:18.738049054 +0000 UTC m=+0.690185562,LastTimestamp:2026-03-07 14:39:18.858345101 +0000 UTC m=+0.810481619,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:22 crc kubenswrapper[4943]: E0307 14:40:22.808648 4943 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189a9606ace07e4f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189a9606ace07e4f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:18.738075215 +0000 UTC m=+0.690211723,LastTimestamp:2026-03-07 14:39:18.858417972 +0000 UTC m=+0.810554480,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:22 crc kubenswrapper[4943]: E0307 14:40:22.814349 4943 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189a9606ace0a5af\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189a9606ace0a5af default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:18.738085295 +0000 UTC m=+0.690221803,LastTimestamp:2026-03-07 14:39:18.858436273 +0000 UTC m=+0.810572781,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:22 crc kubenswrapper[4943]: E0307 14:40:22.820591 4943 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189a9606ace0181e\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189a9606ace0181e default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:18.738049054 +0000 UTC m=+0.690185562,LastTimestamp:2026-03-07 14:39:18.859399223 +0000 UTC m=+0.811535731,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:22 crc kubenswrapper[4943]: E0307 14:40:22.826260 4943 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189a9606ace07e4f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189a9606ace07e4f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:18.738075215 +0000 UTC m=+0.690211723,LastTimestamp:2026-03-07 14:39:18.859428654 +0000 UTC m=+0.811565172,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:22 crc kubenswrapper[4943]: E0307 14:40:22.830965 4943 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189a9606ace0a5af\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189a9606ace0a5af default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:18.738085295 +0000 UTC m=+0.690221803,LastTimestamp:2026-03-07 14:39:18.859443144 +0000 UTC m=+0.811579652,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:22 crc kubenswrapper[4943]: E0307 14:40:22.835889 4943 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189a9606ace0181e\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189a9606ace0181e default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:18.738049054 +0000 UTC m=+0.690185562,LastTimestamp:2026-03-07 14:39:18.86067149 +0000 UTC m=+0.812808008,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:22 crc kubenswrapper[4943]: E0307 14:40:22.841034 4943 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189a9606ace07e4f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189a9606ace07e4f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:18.738075215 +0000 UTC m=+0.690211723,LastTimestamp:2026-03-07 14:39:18.86068603 +0000 UTC m=+0.812822538,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:22 crc kubenswrapper[4943]: E0307 14:40:22.845260 4943 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189a9606ace0a5af\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189a9606ace0a5af default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:18.738085295 +0000 UTC m=+0.690221803,LastTimestamp:2026-03-07 14:39:18.86069693 +0000 UTC m=+0.812833438,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:22 crc kubenswrapper[4943]: E0307 14:40:22.853918 4943 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189a9606ace0181e\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189a9606ace0181e default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:18.738049054 +0000 UTC m=+0.690185562,LastTimestamp:2026-03-07 14:39:18.860997737 +0000 UTC m=+0.813134275,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:22 crc kubenswrapper[4943]: E0307 14:40:22.859250 4943 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189a9606ace07e4f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189a9606ace07e4f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:18.738075215 +0000 UTC m=+0.690211723,LastTimestamp:2026-03-07 14:39:18.861075478 +0000 UTC m=+0.813212016,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:22 crc kubenswrapper[4943]: E0307 14:40:22.863958 4943 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189a9606ace0a5af\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189a9606ace0a5af default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:18.738085295 +0000 UTC m=+0.690221803,LastTimestamp:2026-03-07 14:39:18.861097269 +0000 UTC m=+0.813233807,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:22 crc kubenswrapper[4943]: E0307 14:40:22.869847 4943 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189a9606ace0181e\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189a9606ace0181e default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:18.738049054 +0000 UTC m=+0.690185562,LastTimestamp:2026-03-07 14:39:18.861241082 +0000 UTC m=+0.813377600,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:22 crc kubenswrapper[4943]: E0307 14:40:22.874294 4943 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189a9606ace07e4f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189a9606ace07e4f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:18.738075215 +0000 UTC m=+0.690211723,LastTimestamp:2026-03-07 14:39:18.861252852 +0000 UTC m=+0.813389360,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:22 crc kubenswrapper[4943]: E0307 14:40:22.881225 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189a9606cc281608 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:19.262860808 +0000 UTC m=+1.214997336,LastTimestamp:2026-03-07 14:39:19.262860808 +0000 UTC m=+1.214997336,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:22 crc kubenswrapper[4943]: E0307 14:40:22.885580 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189a9606cc2eec43 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:19.263308867 +0000 UTC m=+1.215445405,LastTimestamp:2026-03-07 14:39:19.263308867 +0000 UTC m=+1.215445405,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:22 crc kubenswrapper[4943]: E0307 14:40:22.887970 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189a9606cc2fb534 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:19.263360308 +0000 UTC m=+1.215496846,LastTimestamp:2026-03-07 14:39:19.263360308 +0000 UTC m=+1.215496846,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:22 crc kubenswrapper[4943]: E0307 14:40:22.890805 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189a9606cd4457ed openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:19.281489901 +0000 UTC m=+1.233626429,LastTimestamp:2026-03-07 14:39:19.281489901 +0000 UTC m=+1.233626429,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:22 crc kubenswrapper[4943]: E0307 14:40:22.892496 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189a9606cd75177c openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:19.284684668 +0000 UTC m=+1.236821156,LastTimestamp:2026-03-07 14:39:19.284684668 +0000 UTC m=+1.236821156,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:22 crc kubenswrapper[4943]: E0307 14:40:22.896695 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189a9606ef10c620 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Created,Message:Created container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:19.848535584 +0000 UTC m=+1.800672122,LastTimestamp:2026-03-07 14:39:19.848535584 +0000 UTC m=+1.800672122,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:22 crc kubenswrapper[4943]: E0307 14:40:22.900791 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189a9606ef92fb7c openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:19.857068924 +0000 UTC m=+1.809205422,LastTimestamp:2026-03-07 14:39:19.857068924 +0000 UTC m=+1.809205422,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:22 crc kubenswrapper[4943]: E0307 14:40:22.904644 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189a9606efa0e6f5 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Created,Message:Created container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:19.857981173 +0000 UTC m=+1.810117671,LastTimestamp:2026-03-07 14:39:19.857981173 +0000 UTC m=+1.810117671,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:22 crc kubenswrapper[4943]: E0307 14:40:22.908393 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189a9606efdf73f3 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:19.862080499 +0000 UTC m=+1.814216997,LastTimestamp:2026-03-07 14:39:19.862080499 +0000 UTC m=+1.814216997,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:22 crc kubenswrapper[4943]: E0307 14:40:22.912498 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189a9606efecf5b8 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:19.862965688 +0000 UTC m=+1.815102186,LastTimestamp:2026-03-07 14:39:19.862965688 +0000 UTC m=+1.815102186,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:22 crc kubenswrapper[4943]: E0307 14:40:22.916117 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189a9606f0637d4a openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Started,Message:Started container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:19.870733642 +0000 UTC m=+1.822870130,LastTimestamp:2026-03-07 14:39:19.870733642 +0000 UTC m=+1.822870130,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:22 crc kubenswrapper[4943]: E0307 14:40:22.919886 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189a9606f076ecf1 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Started,Message:Started container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:19.872007409 +0000 UTC m=+1.824143917,LastTimestamp:2026-03-07 14:39:19.872007409 +0000 UTC m=+1.824143917,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:22 crc kubenswrapper[4943]: E0307 14:40:22.923852 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189a9606f0778819 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:19.872047129 +0000 UTC m=+1.824183667,LastTimestamp:2026-03-07 14:39:19.872047129 +0000 UTC m=+1.824183667,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:22 crc kubenswrapper[4943]: E0307 14:40:22.931999 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189a9606f089c79b openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:19.873243035 +0000 UTC m=+1.825379573,LastTimestamp:2026-03-07 14:39:19.873243035 +0000 UTC m=+1.825379573,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:22 crc kubenswrapper[4943]: E0307 14:40:22.936541 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189a9606f0cc59a5 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:19.877605797 +0000 UTC m=+1.829742345,LastTimestamp:2026-03-07 14:39:19.877605797 +0000 UTC m=+1.829742345,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:22 crc kubenswrapper[4943]: E0307 14:40:22.940525 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189a9606f14459dc openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:19.885470172 +0000 UTC m=+1.837606710,LastTimestamp:2026-03-07 14:39:19.885470172 +0000 UTC m=+1.837606710,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:22 crc kubenswrapper[4943]: E0307 14:40:22.944373 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189a960702f91f7a openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:20.182529914 +0000 UTC m=+2.134666442,LastTimestamp:2026-03-07 14:39:20.182529914 +0000 UTC m=+2.134666442,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:22 crc kubenswrapper[4943]: E0307 14:40:22.947998 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189a960703ad53ef openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:20.194339823 +0000 UTC m=+2.146476361,LastTimestamp:2026-03-07 14:39:20.194339823 +0000 UTC m=+2.146476361,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:22 crc kubenswrapper[4943]: E0307 14:40:22.952135 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189a960703c649a4 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:20.195975588 +0000 UTC m=+2.148112116,LastTimestamp:2026-03-07 14:39:20.195975588 +0000 UTC m=+2.148112116,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:22 crc kubenswrapper[4943]: E0307 14:40:22.959158 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189a9607104037f0 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Created,Message:Created container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:20.40529304 +0000 UTC m=+2.357429568,LastTimestamp:2026-03-07 14:39:20.40529304 +0000 UTC m=+2.357429568,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:22 crc kubenswrapper[4943]: E0307 14:40:22.963379 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189a96071106e4f9 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Started,Message:Started container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:20.418313465 +0000 UTC m=+2.370450003,LastTimestamp:2026-03-07 14:39:20.418313465 +0000 UTC m=+2.370450003,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:22 crc kubenswrapper[4943]: E0307 14:40:22.967848 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189a960711217af9 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:20.420055801 +0000 UTC m=+2.372192389,LastTimestamp:2026-03-07 14:39:20.420055801 +0000 UTC m=+2.372192389,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:22 crc kubenswrapper[4943]: E0307 14:40:22.971817 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189a96071f87d550 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Created,Message:Created container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:20.661644624 +0000 UTC m=+2.613781162,LastTimestamp:2026-03-07 14:39:20.661644624 +0000 UTC m=+2.613781162,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:22 crc kubenswrapper[4943]: E0307 14:40:22.976759 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189a960720664ce7 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Started,Message:Started container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:20.676224231 +0000 UTC m=+2.628360769,LastTimestamp:2026-03-07 14:39:20.676224231 +0000 UTC m=+2.628360769,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:22 crc kubenswrapper[4943]: E0307 14:40:22.983627 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189a9607266cb280 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:20.777306752 +0000 UTC m=+2.729443260,LastTimestamp:2026-03-07 14:39:20.777306752 +0000 UTC m=+2.729443260,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:22 crc kubenswrapper[4943]: E0307 14:40:22.990402 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189a960726afd45f openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:20.781706335 +0000 UTC m=+2.733842883,LastTimestamp:2026-03-07 14:39:20.781706335 +0000 UTC m=+2.733842883,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:22 crc kubenswrapper[4943]: E0307 14:40:22.997050 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189a9607270b3de1 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:20.787697121 +0000 UTC m=+2.739833649,LastTimestamp:2026-03-07 14:39:20.787697121 +0000 UTC m=+2.739833649,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.001644 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189a9607274cd644 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:20.791995972 +0000 UTC m=+2.744132510,LastTimestamp:2026-03-07 14:39:20.791995972 +0000 UTC m=+2.744132510,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.005817 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189a9607359a1cef openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Created,Message:Created container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:21.031941359 +0000 UTC m=+2.984077857,LastTimestamp:2026-03-07 14:39:21.031941359 +0000 UTC m=+2.984077857,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.010387 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189a9607359e4934 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Created,Message:Created container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:21.032214836 +0000 UTC m=+2.984351334,LastTimestamp:2026-03-07 14:39:21.032214836 +0000 UTC m=+2.984351334,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.015212 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189a9607363a9727 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Started,Message:Started container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:21.042458407 +0000 UTC m=+2.994594905,LastTimestamp:2026-03-07 14:39:21.042458407 +0000 UTC m=+2.994594905,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.020939 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189a9607364eb584 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:21.0437769 +0000 UTC m=+2.995913398,LastTimestamp:2026-03-07 14:39:21.0437769 +0000 UTC m=+2.995913398,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.026440 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189a96073650aa28 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Created,Message:Created container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:21.043905064 +0000 UTC m=+2.996041562,LastTimestamp:2026-03-07 14:39:21.043905064 +0000 UTC m=+2.996041562,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.031680 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189a9607367b88f7 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Started,Message:Started container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:21.046714615 +0000 UTC m=+2.998851113,LastTimestamp:2026-03-07 14:39:21.046714615 +0000 UTC m=+2.998851113,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.036119 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189a960736828ba9 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Created,Message:Created container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:21.047174057 +0000 UTC m=+2.999310555,LastTimestamp:2026-03-07 14:39:21.047174057 +0000 UTC m=+2.999310555,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.041343 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189a9607379bb1e0 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Started,Message:Started container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:21.065599456 +0000 UTC m=+3.017735954,LastTimestamp:2026-03-07 14:39:21.065599456 +0000 UTC m=+3.017735954,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.045650 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189a960737ae5f03 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:21.066823427 +0000 UTC m=+3.018959925,LastTimestamp:2026-03-07 14:39:21.066823427 +0000 UTC m=+3.018959925,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.052091 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189a960737ce55a5 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Started,Message:Started container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:21.068918181 +0000 UTC m=+3.021054679,LastTimestamp:2026-03-07 14:39:21.068918181 +0000 UTC m=+3.021054679,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.058888 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189a960743713d06 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Created,Message:Created container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:21.264143622 +0000 UTC m=+3.216280140,LastTimestamp:2026-03-07 14:39:21.264143622 +0000 UTC m=+3.216280140,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.064374 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189a960744470702 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Started,Message:Started container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:21.278154498 +0000 UTC m=+3.230291006,LastTimestamp:2026-03-07 14:39:21.278154498 +0000 UTC m=+3.230291006,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.068300 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189a9607445857bd openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:21.279289277 +0000 UTC m=+3.231425785,LastTimestamp:2026-03-07 14:39:21.279289277 +0000 UTC m=+3.231425785,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.072185 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189a9607462fbbd2 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Created,Message:Created container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:21.310182354 +0000 UTC m=+3.262318852,LastTimestamp:2026-03-07 14:39:21.310182354 +0000 UTC m=+3.262318852,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.077888 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189a9607470a6147 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Started,Message:Started container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:21.324511559 +0000 UTC m=+3.276648057,LastTimestamp:2026-03-07 14:39:21.324511559 +0000 UTC m=+3.276648057,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.084164 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189a9607471febad openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:21.325923245 +0000 UTC m=+3.278059753,LastTimestamp:2026-03-07 14:39:21.325923245 +0000 UTC m=+3.278059753,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.088299 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189a9607518cc367 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Created,Message:Created container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:21.500828519 +0000 UTC m=+3.452965017,LastTimestamp:2026-03-07 14:39:21.500828519 +0000 UTC m=+3.452965017,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.091750 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189a9607528c8c18 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Started,Message:Started container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:21.517591576 +0000 UTC m=+3.469728074,LastTimestamp:2026-03-07 14:39:21.517591576 +0000 UTC m=+3.469728074,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.095812 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189a960752eb6a9c openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Created,Message:Created container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:21.523808924 +0000 UTC m=+3.475945422,LastTimestamp:2026-03-07 14:39:21.523808924 +0000 UTC m=+3.475945422,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.099683 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189a960753e75a51 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Started,Message:Started container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:21.540319825 +0000 UTC m=+3.492456323,LastTimestamp:2026-03-07 14:39:21.540319825 +0000 UTC m=+3.492456323,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.103236 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189a960753f8a79c openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:21.541453724 +0000 UTC m=+3.493590232,LastTimestamp:2026-03-07 14:39:21.541453724 +0000 UTC m=+3.493590232,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.109228 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189a96075ecdb55c openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Created,Message:Created container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:21.723188572 +0000 UTC m=+3.675325080,LastTimestamp:2026-03-07 14:39:21.723188572 +0000 UTC m=+3.675325080,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.115124 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189a96075fa18d71 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Started,Message:Started container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:21.737071985 +0000 UTC m=+3.689208523,LastTimestamp:2026-03-07 14:39:21.737071985 +0000 UTC m=+3.689208523,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.118248 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189a96075fb882fa openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:21.738576634 +0000 UTC m=+3.690713142,LastTimestamp:2026-03-07 14:39:21.738576634 +0000 UTC m=+3.690713142,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.123214 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189a96076309fc16 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:21.794247702 +0000 UTC m=+3.746384210,LastTimestamp:2026-03-07 14:39:21.794247702 +0000 UTC m=+3.746384210,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.130954 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189a96076e5e3fbe openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:21.984319422 +0000 UTC m=+3.936455930,LastTimestamp:2026-03-07 14:39:21.984319422 +0000 UTC m=+3.936455930,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.134822 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189a96076ef9268b openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Created,Message:Created container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:21.994471051 +0000 UTC m=+3.946607549,LastTimestamp:2026-03-07 14:39:21.994471051 +0000 UTC m=+3.946607549,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.140805 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189a96076f837539 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:22.003535161 +0000 UTC m=+3.955671669,LastTimestamp:2026-03-07 14:39:22.003535161 +0000 UTC m=+3.955671669,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.145945 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189a96076ff9acd9 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Started,Message:Started container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:22.011282649 +0000 UTC m=+3.963419157,LastTimestamp:2026-03-07 14:39:22.011282649 +0000 UTC m=+3.963419157,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.150154 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189a9607a0ef7cd2 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:22.832698578 +0000 UTC m=+4.784835086,LastTimestamp:2026-03-07 14:39:22.832698578 +0000 UTC m=+4.784835086,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.153686 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189a9607ada975bf openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Created,Message:Created container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:23.046213055 +0000 UTC m=+4.998349593,LastTimestamp:2026-03-07 14:39:23.046213055 +0000 UTC m=+4.998349593,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.159658 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189a9607ae639c3e openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Started,Message:Started container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:23.058412606 +0000 UTC m=+5.010549144,LastTimestamp:2026-03-07 14:39:23.058412606 +0000 UTC m=+5.010549144,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.165122 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189a9607ae7eebe8 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:23.060202472 +0000 UTC m=+5.012339010,LastTimestamp:2026-03-07 14:39:23.060202472 +0000 UTC m=+5.012339010,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.169358 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189a9607bcd5f7d2 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Created,Message:Created container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:23.300788178 +0000 UTC m=+5.252924706,LastTimestamp:2026-03-07 14:39:23.300788178 +0000 UTC m=+5.252924706,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.172767 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189a9607bdfdf96b openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Started,Message:Started container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:23.320187243 +0000 UTC m=+5.272323771,LastTimestamp:2026-03-07 14:39:23.320187243 +0000 UTC m=+5.272323771,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.176710 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189a9607be1de166 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:23.322278246 +0000 UTC m=+5.274414754,LastTimestamp:2026-03-07 14:39:23.322278246 +0000 UTC m=+5.274414754,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.180791 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189a9607ccbe0169 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Created,Message:Created container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:23.567653225 +0000 UTC m=+5.519789723,LastTimestamp:2026-03-07 14:39:23.567653225 +0000 UTC m=+5.519789723,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.185538 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189a9607cde83084 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Started,Message:Started container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:23.587195012 +0000 UTC m=+5.539331510,LastTimestamp:2026-03-07 14:39:23.587195012 +0000 UTC m=+5.539331510,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.189611 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189a9607cdfe05ff openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:23.588625919 +0000 UTC m=+5.540762407,LastTimestamp:2026-03-07 14:39:23.588625919 +0000 UTC m=+5.540762407,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.193991 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189a9607dd9737dd openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Created,Message:Created container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:23.850323933 +0000 UTC m=+5.802460441,LastTimestamp:2026-03-07 14:39:23.850323933 +0000 UTC m=+5.802460441,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.200440 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189a9607de8e6baa openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Started,Message:Started container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:23.866524586 +0000 UTC m=+5.818661124,LastTimestamp:2026-03-07 14:39:23.866524586 +0000 UTC m=+5.818661124,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.204672 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189a9607deaa11e4 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:23.868336612 +0000 UTC m=+5.820473150,LastTimestamp:2026-03-07 14:39:23.868336612 +0000 UTC m=+5.820473150,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.210375 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189a9607ed7e8458 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Created,Message:Created container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:24.117140568 +0000 UTC m=+6.069277106,LastTimestamp:2026-03-07 14:39:24.117140568 +0000 UTC m=+6.069277106,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.215013 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189a9607ee7fe55e openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Started,Message:Started container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:24.134008158 +0000 UTC m=+6.086144696,LastTimestamp:2026-03-07 14:39:24.134008158 +0000 UTC m=+6.086144696,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.225568 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 07 14:40:23 crc kubenswrapper[4943]: &Event{ObjectMeta:{kube-controller-manager-crc.189a9608ee6f2483 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": context deadline exceeded (Client.Timeout exceeded while awaiting headers) Mar 07 14:40:23 crc kubenswrapper[4943]: body: Mar 07 14:40:23 crc kubenswrapper[4943]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:28.427877507 +0000 UTC m=+10.380014035,LastTimestamp:2026-03-07 14:39:28.427877507 +0000 UTC m=+10.380014035,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 07 14:40:23 crc kubenswrapper[4943]: > Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.232762 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189a9608ee70f1f8 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:28.42799564 +0000 UTC m=+10.380132178,LastTimestamp:2026-03-07 14:39:28.42799564 +0000 UTC m=+10.380132178,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.241123 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 07 14:40:23 crc kubenswrapper[4943]: &Event{ObjectMeta:{kube-apiserver-crc.189a9609f612d7ae openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:ProbeError,Message:Liveness probe error: Get "https://192.168.126.11:17697/healthz": read tcp 192.168.126.11:35564->192.168.126.11:17697: read: connection reset by peer Mar 07 14:40:23 crc kubenswrapper[4943]: body: Mar 07 14:40:23 crc kubenswrapper[4943]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:32.85101355 +0000 UTC m=+14.803150088,LastTimestamp:2026-03-07 14:39:32.85101355 +0000 UTC m=+14.803150088,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 07 14:40:23 crc kubenswrapper[4943]: > Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.248549 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189a9609f6140db0 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Unhealthy,Message:Liveness probe failed: Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:35564->192.168.126.11:17697: read: connection reset by peer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:32.851092912 +0000 UTC m=+14.803229450,LastTimestamp:2026-03-07 14:39:32.851092912 +0000 UTC m=+14.803229450,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.255704 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 07 14:40:23 crc kubenswrapper[4943]: &Event{ObjectMeta:{kube-apiserver-crc.189a960a07261c96 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Mar 07 14:40:23 crc kubenswrapper[4943]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 07 14:40:23 crc kubenswrapper[4943]: Mar 07 14:40:23 crc kubenswrapper[4943]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:33.137489046 +0000 UTC m=+15.089625584,LastTimestamp:2026-03-07 14:39:33.137489046 +0000 UTC m=+15.089625584,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 07 14:40:23 crc kubenswrapper[4943]: > Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.261568 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189a960a0726ec71 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:33.137542257 +0000 UTC m=+15.089678785,LastTimestamp:2026-03-07 14:39:33.137542257 +0000 UTC m=+15.089678785,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.265668 4943 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189a960a07261c96\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 07 14:40:23 crc kubenswrapper[4943]: &Event{ObjectMeta:{kube-apiserver-crc.189a960a07261c96 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Mar 07 14:40:23 crc kubenswrapper[4943]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 07 14:40:23 crc kubenswrapper[4943]: Mar 07 14:40:23 crc kubenswrapper[4943]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:33.137489046 +0000 UTC m=+15.089625584,LastTimestamp:2026-03-07 14:39:33.144413882 +0000 UTC m=+15.096550400,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 07 14:40:23 crc kubenswrapper[4943]: > Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.269425 4943 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189a960a0726ec71\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189a960a0726ec71 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:33.137542257 +0000 UTC m=+15.089678785,LastTimestamp:2026-03-07 14:39:33.144573036 +0000 UTC m=+15.096709544,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.275413 4943 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189a96075fb882fa\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189a96075fb882fa openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:21.738576634 +0000 UTC m=+3.690713142,LastTimestamp:2026-03-07 14:39:33.892591586 +0000 UTC m=+15.844728114,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.282040 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 07 14:40:23 crc kubenswrapper[4943]: &Event{ObjectMeta:{kube-controller-manager-crc.189a960b4283314e openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 07 14:40:23 crc kubenswrapper[4943]: body: Mar 07 14:40:23 crc kubenswrapper[4943]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:38.428412238 +0000 UTC m=+20.380548776,LastTimestamp:2026-03-07 14:39:38.428412238 +0000 UTC m=+20.380548776,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 07 14:40:23 crc kubenswrapper[4943]: > Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.286316 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189a960b428445c2 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:38.42848301 +0000 UTC m=+20.380619548,LastTimestamp:2026-03-07 14:39:38.42848301 +0000 UTC m=+20.380619548,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.294031 4943 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189a960b4283314e\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 07 14:40:23 crc kubenswrapper[4943]: &Event{ObjectMeta:{kube-controller-manager-crc.189a960b4283314e openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 07 14:40:23 crc kubenswrapper[4943]: body: Mar 07 14:40:23 crc kubenswrapper[4943]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:38.428412238 +0000 UTC m=+20.380548776,LastTimestamp:2026-03-07 14:39:48.428021124 +0000 UTC m=+30.380157652,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 07 14:40:23 crc kubenswrapper[4943]: > Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.296486 4943 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189a960b428445c2\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189a960b428445c2 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:38.42848301 +0000 UTC m=+20.380619548,LastTimestamp:2026-03-07 14:39:48.428087666 +0000 UTC m=+30.380224194,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.299423 4943 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189a960d96b6b10a openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Killing,Message:Container cluster-policy-controller failed startup probe, will be restarted,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:48.43100801 +0000 UTC m=+30.383144538,LastTimestamp:2026-03-07 14:39:48.43100801 +0000 UTC m=+30.383144538,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.303906 4943 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189a9606f089c79b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189a9606f089c79b openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:19.873243035 +0000 UTC m=+1.825379573,LastTimestamp:2026-03-07 14:39:48.554225988 +0000 UTC m=+30.506362516,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.311282 4943 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189a960702f91f7a\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189a960702f91f7a openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:20.182529914 +0000 UTC m=+2.134666442,LastTimestamp:2026-03-07 14:39:48.791379127 +0000 UTC m=+30.743515635,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.317334 4943 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189a960703ad53ef\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189a960703ad53ef openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:20.194339823 +0000 UTC m=+2.146476361,LastTimestamp:2026-03-07 14:39:48.800640063 +0000 UTC m=+30.752776571,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.326196 4943 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189a960b4283314e\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 07 14:40:23 crc kubenswrapper[4943]: &Event{ObjectMeta:{kube-controller-manager-crc.189a960b4283314e openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 07 14:40:23 crc kubenswrapper[4943]: body: Mar 07 14:40:23 crc kubenswrapper[4943]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:38.428412238 +0000 UTC m=+20.380548776,LastTimestamp:2026-03-07 14:39:58.427555486 +0000 UTC m=+40.379692024,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 07 14:40:23 crc kubenswrapper[4943]: > Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.333823 4943 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189a960b428445c2\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189a960b428445c2 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:38.42848301 +0000 UTC m=+20.380619548,LastTimestamp:2026-03-07 14:39:58.427618068 +0000 UTC m=+40.379754596,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:40:23 crc kubenswrapper[4943]: E0307 14:40:23.342915 4943 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189a960b4283314e\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 07 14:40:23 crc kubenswrapper[4943]: &Event{ObjectMeta:{kube-controller-manager-crc.189a960b4283314e openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 07 14:40:23 crc kubenswrapper[4943]: body: Mar 07 14:40:23 crc kubenswrapper[4943]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:39:38.428412238 +0000 UTC m=+20.380548776,LastTimestamp:2026-03-07 14:40:08.426894715 +0000 UTC m=+50.379031253,Count:4,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 07 14:40:23 crc kubenswrapper[4943]: > Mar 07 14:40:23 crc kubenswrapper[4943]: I0307 14:40:23.681146 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 07 14:40:24 crc kubenswrapper[4943]: I0307 14:40:24.681891 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 07 14:40:25 crc kubenswrapper[4943]: I0307 14:40:25.427543 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 07 14:40:25 crc kubenswrapper[4943]: I0307 14:40:25.427755 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:40:25 crc kubenswrapper[4943]: I0307 14:40:25.429096 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:25 crc kubenswrapper[4943]: I0307 14:40:25.429135 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:25 crc kubenswrapper[4943]: I0307 14:40:25.429145 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:25 crc kubenswrapper[4943]: I0307 14:40:25.679988 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 07 14:40:26 crc kubenswrapper[4943]: I0307 14:40:26.338482 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 07 14:40:26 crc kubenswrapper[4943]: I0307 14:40:26.338683 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:40:26 crc kubenswrapper[4943]: I0307 14:40:26.340137 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:26 crc kubenswrapper[4943]: I0307 14:40:26.340184 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:26 crc kubenswrapper[4943]: I0307 14:40:26.340195 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:26 crc kubenswrapper[4943]: I0307 14:40:26.680774 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 07 14:40:27 crc kubenswrapper[4943]: I0307 14:40:27.685254 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 07 14:40:28 crc kubenswrapper[4943]: E0307 14:40:28.573380 4943 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 07 14:40:28 crc kubenswrapper[4943]: I0307 14:40:28.586627 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:40:28 crc kubenswrapper[4943]: I0307 14:40:28.587911 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:28 crc kubenswrapper[4943]: I0307 14:40:28.587964 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:28 crc kubenswrapper[4943]: I0307 14:40:28.587976 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:28 crc kubenswrapper[4943]: I0307 14:40:28.588000 4943 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 07 14:40:28 crc kubenswrapper[4943]: E0307 14:40:28.592523 4943 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 07 14:40:28 crc kubenswrapper[4943]: I0307 14:40:28.680065 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 07 14:40:28 crc kubenswrapper[4943]: E0307 14:40:28.864167 4943 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 07 14:40:29 crc kubenswrapper[4943]: I0307 14:40:29.435177 4943 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 07 14:40:29 crc kubenswrapper[4943]: I0307 14:40:29.451150 4943 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Mar 07 14:40:29 crc kubenswrapper[4943]: I0307 14:40:29.684057 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 07 14:40:30 crc kubenswrapper[4943]: I0307 14:40:30.680789 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 07 14:40:31 crc kubenswrapper[4943]: I0307 14:40:31.665401 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 07 14:40:31 crc kubenswrapper[4943]: I0307 14:40:31.665637 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:40:31 crc kubenswrapper[4943]: I0307 14:40:31.667229 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:31 crc kubenswrapper[4943]: I0307 14:40:31.667262 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:31 crc kubenswrapper[4943]: I0307 14:40:31.667273 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:31 crc kubenswrapper[4943]: I0307 14:40:31.681172 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 07 14:40:31 crc kubenswrapper[4943]: I0307 14:40:31.755490 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:40:31 crc kubenswrapper[4943]: I0307 14:40:31.757026 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:31 crc kubenswrapper[4943]: I0307 14:40:31.757074 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:31 crc kubenswrapper[4943]: I0307 14:40:31.757087 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:31 crc kubenswrapper[4943]: I0307 14:40:31.757779 4943 scope.go:117] "RemoveContainer" containerID="a25d75089d6ba494eeab9670dc43f7d2ed9a201acee535fe805e39e80ceb0cb5" Mar 07 14:40:32 crc kubenswrapper[4943]: I0307 14:40:32.121055 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 07 14:40:32 crc kubenswrapper[4943]: I0307 14:40:32.123066 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ed8d27f5a369a691493bda9dbd6e03c098931b8dab637bbd9bacac6fe7979c80"} Mar 07 14:40:32 crc kubenswrapper[4943]: I0307 14:40:32.123223 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:40:32 crc kubenswrapper[4943]: I0307 14:40:32.124015 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:32 crc kubenswrapper[4943]: I0307 14:40:32.124069 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:32 crc kubenswrapper[4943]: I0307 14:40:32.124088 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:32 crc kubenswrapper[4943]: I0307 14:40:32.681063 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 07 14:40:33 crc kubenswrapper[4943]: I0307 14:40:33.127405 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 07 14:40:33 crc kubenswrapper[4943]: I0307 14:40:33.128095 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 07 14:40:33 crc kubenswrapper[4943]: I0307 14:40:33.130492 4943 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="ed8d27f5a369a691493bda9dbd6e03c098931b8dab637bbd9bacac6fe7979c80" exitCode=255 Mar 07 14:40:33 crc kubenswrapper[4943]: I0307 14:40:33.130546 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"ed8d27f5a369a691493bda9dbd6e03c098931b8dab637bbd9bacac6fe7979c80"} Mar 07 14:40:33 crc kubenswrapper[4943]: I0307 14:40:33.130599 4943 scope.go:117] "RemoveContainer" containerID="a25d75089d6ba494eeab9670dc43f7d2ed9a201acee535fe805e39e80ceb0cb5" Mar 07 14:40:33 crc kubenswrapper[4943]: I0307 14:40:33.130804 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:40:33 crc kubenswrapper[4943]: I0307 14:40:33.132405 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:33 crc kubenswrapper[4943]: I0307 14:40:33.132451 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:33 crc kubenswrapper[4943]: I0307 14:40:33.132474 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:33 crc kubenswrapper[4943]: I0307 14:40:33.133440 4943 scope.go:117] "RemoveContainer" containerID="ed8d27f5a369a691493bda9dbd6e03c098931b8dab637bbd9bacac6fe7979c80" Mar 07 14:40:33 crc kubenswrapper[4943]: E0307 14:40:33.133743 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 07 14:40:33 crc kubenswrapper[4943]: I0307 14:40:33.684175 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 07 14:40:33 crc kubenswrapper[4943]: I0307 14:40:33.779651 4943 csr.go:261] certificate signing request csr-ngrmc is approved, waiting to be issued Mar 07 14:40:33 crc kubenswrapper[4943]: I0307 14:40:33.790355 4943 csr.go:257] certificate signing request csr-ngrmc is issued Mar 07 14:40:33 crc kubenswrapper[4943]: I0307 14:40:33.824162 4943 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Mar 07 14:40:34 crc kubenswrapper[4943]: I0307 14:40:34.033711 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 07 14:40:34 crc kubenswrapper[4943]: I0307 14:40:34.135000 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 07 14:40:34 crc kubenswrapper[4943]: I0307 14:40:34.137503 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:40:34 crc kubenswrapper[4943]: I0307 14:40:34.138500 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:34 crc kubenswrapper[4943]: I0307 14:40:34.138580 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:34 crc kubenswrapper[4943]: I0307 14:40:34.138607 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:34 crc kubenswrapper[4943]: I0307 14:40:34.139853 4943 scope.go:117] "RemoveContainer" containerID="ed8d27f5a369a691493bda9dbd6e03c098931b8dab637bbd9bacac6fe7979c80" Mar 07 14:40:34 crc kubenswrapper[4943]: E0307 14:40:34.140268 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 07 14:40:34 crc kubenswrapper[4943]: I0307 14:40:34.521988 4943 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Mar 07 14:40:34 crc kubenswrapper[4943]: I0307 14:40:34.792761 4943 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2027-01-09 20:02:18.4124536 +0000 UTC Mar 07 14:40:34 crc kubenswrapper[4943]: I0307 14:40:34.792833 4943 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 7397h21m43.619626321s for next certificate rotation Mar 07 14:40:35 crc kubenswrapper[4943]: I0307 14:40:35.592955 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:40:35 crc kubenswrapper[4943]: I0307 14:40:35.594469 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:35 crc kubenswrapper[4943]: I0307 14:40:35.594529 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:35 crc kubenswrapper[4943]: I0307 14:40:35.594543 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:35 crc kubenswrapper[4943]: I0307 14:40:35.594642 4943 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 07 14:40:35 crc kubenswrapper[4943]: I0307 14:40:35.603682 4943 kubelet_node_status.go:115] "Node was previously registered" node="crc" Mar 07 14:40:35 crc kubenswrapper[4943]: I0307 14:40:35.604228 4943 kubelet_node_status.go:79] "Successfully registered node" node="crc" Mar 07 14:40:35 crc kubenswrapper[4943]: E0307 14:40:35.604357 4943 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 07 14:40:35 crc kubenswrapper[4943]: I0307 14:40:35.608451 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:35 crc kubenswrapper[4943]: I0307 14:40:35.608485 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:35 crc kubenswrapper[4943]: I0307 14:40:35.608495 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:35 crc kubenswrapper[4943]: I0307 14:40:35.608532 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:35 crc kubenswrapper[4943]: I0307 14:40:35.608544 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:35Z","lastTransitionTime":"2026-03-07T14:40:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:35 crc kubenswrapper[4943]: E0307 14:40:35.627335 4943 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-07T14:40:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-07T14:40:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-07T14:40:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-07T14:40:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7a039d39-2e22-491c-a9ee-09ff4d277e8b\\\",\\\"systemUUID\\\":\\\"bf22d61b-d7ce-44b5-9dd8-1a46c0f78104\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 07 14:40:35 crc kubenswrapper[4943]: I0307 14:40:35.638095 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:35 crc kubenswrapper[4943]: I0307 14:40:35.638158 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:35 crc kubenswrapper[4943]: I0307 14:40:35.638177 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:35 crc kubenswrapper[4943]: I0307 14:40:35.638204 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:35 crc kubenswrapper[4943]: I0307 14:40:35.638223 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:35Z","lastTransitionTime":"2026-03-07T14:40:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:35 crc kubenswrapper[4943]: E0307 14:40:35.660005 4943 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-07T14:40:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-07T14:40:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-07T14:40:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-07T14:40:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7a039d39-2e22-491c-a9ee-09ff4d277e8b\\\",\\\"systemUUID\\\":\\\"bf22d61b-d7ce-44b5-9dd8-1a46c0f78104\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 07 14:40:35 crc kubenswrapper[4943]: I0307 14:40:35.671687 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:35 crc kubenswrapper[4943]: I0307 14:40:35.671765 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:35 crc kubenswrapper[4943]: I0307 14:40:35.671789 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:35 crc kubenswrapper[4943]: I0307 14:40:35.671823 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:35 crc kubenswrapper[4943]: I0307 14:40:35.671847 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:35Z","lastTransitionTime":"2026-03-07T14:40:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:35 crc kubenswrapper[4943]: E0307 14:40:35.691821 4943 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-07T14:40:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-07T14:40:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-07T14:40:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-07T14:40:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7a039d39-2e22-491c-a9ee-09ff4d277e8b\\\",\\\"systemUUID\\\":\\\"bf22d61b-d7ce-44b5-9dd8-1a46c0f78104\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 07 14:40:35 crc kubenswrapper[4943]: I0307 14:40:35.702041 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:35 crc kubenswrapper[4943]: I0307 14:40:35.702111 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:35 crc kubenswrapper[4943]: I0307 14:40:35.702124 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:35 crc kubenswrapper[4943]: I0307 14:40:35.702142 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:35 crc kubenswrapper[4943]: I0307 14:40:35.702155 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:35Z","lastTransitionTime":"2026-03-07T14:40:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:35 crc kubenswrapper[4943]: E0307 14:40:35.717242 4943 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-07T14:40:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-07T14:40:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-07T14:40:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-07T14:40:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7a039d39-2e22-491c-a9ee-09ff4d277e8b\\\",\\\"systemUUID\\\":\\\"bf22d61b-d7ce-44b5-9dd8-1a46c0f78104\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 07 14:40:35 crc kubenswrapper[4943]: E0307 14:40:35.717530 4943 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 07 14:40:35 crc kubenswrapper[4943]: E0307 14:40:35.717632 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:35 crc kubenswrapper[4943]: E0307 14:40:35.818314 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:35 crc kubenswrapper[4943]: E0307 14:40:35.919333 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:36 crc kubenswrapper[4943]: E0307 14:40:36.020239 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:36 crc kubenswrapper[4943]: E0307 14:40:36.120693 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:36 crc kubenswrapper[4943]: E0307 14:40:36.221213 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:36 crc kubenswrapper[4943]: E0307 14:40:36.321792 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:36 crc kubenswrapper[4943]: E0307 14:40:36.422384 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:36 crc kubenswrapper[4943]: E0307 14:40:36.523189 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:36 crc kubenswrapper[4943]: E0307 14:40:36.623991 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:36 crc kubenswrapper[4943]: E0307 14:40:36.724143 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:36 crc kubenswrapper[4943]: E0307 14:40:36.824679 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:36 crc kubenswrapper[4943]: E0307 14:40:36.924810 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:37 crc kubenswrapper[4943]: E0307 14:40:37.025705 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:37 crc kubenswrapper[4943]: E0307 14:40:37.126218 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:37 crc kubenswrapper[4943]: E0307 14:40:37.226486 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:37 crc kubenswrapper[4943]: E0307 14:40:37.327237 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:37 crc kubenswrapper[4943]: E0307 14:40:37.427687 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:37 crc kubenswrapper[4943]: E0307 14:40:37.528405 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:37 crc kubenswrapper[4943]: E0307 14:40:37.628772 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:37 crc kubenswrapper[4943]: E0307 14:40:37.729179 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:37 crc kubenswrapper[4943]: E0307 14:40:37.829657 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:37 crc kubenswrapper[4943]: E0307 14:40:37.930187 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:38 crc kubenswrapper[4943]: E0307 14:40:38.030383 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:38 crc kubenswrapper[4943]: E0307 14:40:38.131534 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:38 crc kubenswrapper[4943]: E0307 14:40:38.232115 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:38 crc kubenswrapper[4943]: E0307 14:40:38.332180 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:38 crc kubenswrapper[4943]: E0307 14:40:38.432393 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:38 crc kubenswrapper[4943]: E0307 14:40:38.532913 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:38 crc kubenswrapper[4943]: E0307 14:40:38.633294 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:38 crc kubenswrapper[4943]: E0307 14:40:38.734269 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:38 crc kubenswrapper[4943]: E0307 14:40:38.835399 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:38 crc kubenswrapper[4943]: E0307 14:40:38.864340 4943 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 07 14:40:38 crc kubenswrapper[4943]: E0307 14:40:38.935687 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:39 crc kubenswrapper[4943]: E0307 14:40:39.035979 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:39 crc kubenswrapper[4943]: E0307 14:40:39.136244 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:39 crc kubenswrapper[4943]: E0307 14:40:39.236733 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:39 crc kubenswrapper[4943]: E0307 14:40:39.337708 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:39 crc kubenswrapper[4943]: E0307 14:40:39.438117 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:39 crc kubenswrapper[4943]: E0307 14:40:39.538971 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:39 crc kubenswrapper[4943]: E0307 14:40:39.639467 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:39 crc kubenswrapper[4943]: E0307 14:40:39.740100 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:39 crc kubenswrapper[4943]: E0307 14:40:39.840256 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:39 crc kubenswrapper[4943]: E0307 14:40:39.941483 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:40 crc kubenswrapper[4943]: E0307 14:40:40.042219 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:40 crc kubenswrapper[4943]: E0307 14:40:40.142408 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:40 crc kubenswrapper[4943]: E0307 14:40:40.242671 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:40 crc kubenswrapper[4943]: E0307 14:40:40.342997 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:40 crc kubenswrapper[4943]: E0307 14:40:40.443539 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:40 crc kubenswrapper[4943]: E0307 14:40:40.544725 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:40 crc kubenswrapper[4943]: E0307 14:40:40.645605 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:40 crc kubenswrapper[4943]: E0307 14:40:40.746328 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:40 crc kubenswrapper[4943]: E0307 14:40:40.846705 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:40 crc kubenswrapper[4943]: E0307 14:40:40.947015 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:41 crc kubenswrapper[4943]: E0307 14:40:41.048009 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:41 crc kubenswrapper[4943]: E0307 14:40:41.148843 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:41 crc kubenswrapper[4943]: E0307 14:40:41.249404 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:41 crc kubenswrapper[4943]: I0307 14:40:41.299206 4943 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Mar 07 14:40:41 crc kubenswrapper[4943]: E0307 14:40:41.349750 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:41 crc kubenswrapper[4943]: E0307 14:40:41.450883 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:41 crc kubenswrapper[4943]: E0307 14:40:41.551958 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:41 crc kubenswrapper[4943]: E0307 14:40:41.653021 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:41 crc kubenswrapper[4943]: E0307 14:40:41.754191 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:41 crc kubenswrapper[4943]: I0307 14:40:41.754405 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:40:41 crc kubenswrapper[4943]: I0307 14:40:41.756329 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:41 crc kubenswrapper[4943]: I0307 14:40:41.756381 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:41 crc kubenswrapper[4943]: I0307 14:40:41.756395 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:41 crc kubenswrapper[4943]: E0307 14:40:41.855402 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:41 crc kubenswrapper[4943]: I0307 14:40:41.941063 4943 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 07 14:40:41 crc kubenswrapper[4943]: I0307 14:40:41.941371 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 07 14:40:41 crc kubenswrapper[4943]: I0307 14:40:41.943093 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:41 crc kubenswrapper[4943]: I0307 14:40:41.943176 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:41 crc kubenswrapper[4943]: I0307 14:40:41.943204 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:41 crc kubenswrapper[4943]: I0307 14:40:41.944200 4943 scope.go:117] "RemoveContainer" containerID="ed8d27f5a369a691493bda9dbd6e03c098931b8dab637bbd9bacac6fe7979c80" Mar 07 14:40:41 crc kubenswrapper[4943]: E0307 14:40:41.944494 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 07 14:40:41 crc kubenswrapper[4943]: E0307 14:40:41.956276 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:42 crc kubenswrapper[4943]: E0307 14:40:42.056720 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:42 crc kubenswrapper[4943]: E0307 14:40:42.157846 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:42 crc kubenswrapper[4943]: E0307 14:40:42.258644 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:42 crc kubenswrapper[4943]: E0307 14:40:42.359487 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:42 crc kubenswrapper[4943]: E0307 14:40:42.460664 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:42 crc kubenswrapper[4943]: E0307 14:40:42.561414 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:42 crc kubenswrapper[4943]: E0307 14:40:42.662555 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:42 crc kubenswrapper[4943]: E0307 14:40:42.763139 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:42 crc kubenswrapper[4943]: E0307 14:40:42.864339 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:42 crc kubenswrapper[4943]: E0307 14:40:42.965528 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:43 crc kubenswrapper[4943]: E0307 14:40:43.066158 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.108794 4943 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.168409 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.168492 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.168509 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.168535 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.168554 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:43Z","lastTransitionTime":"2026-03-07T14:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.271628 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.272035 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.272253 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.272453 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.272640 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:43Z","lastTransitionTime":"2026-03-07T14:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.376160 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.376235 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.376257 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.376286 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.376311 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:43Z","lastTransitionTime":"2026-03-07T14:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.479768 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.479836 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.479855 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.479881 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.479902 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:43Z","lastTransitionTime":"2026-03-07T14:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.583146 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.583201 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.583219 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.583244 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.583262 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:43Z","lastTransitionTime":"2026-03-07T14:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.686886 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.688166 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.688346 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.688507 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.688672 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:43Z","lastTransitionTime":"2026-03-07T14:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.728772 4943 apiserver.go:52] "Watching apiserver" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.734064 4943 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.734410 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf"] Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.734962 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.735177 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.735418 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.735514 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 07 14:40:43 crc kubenswrapper[4943]: E0307 14:40:43.735623 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 07 14:40:43 crc kubenswrapper[4943]: E0307 14:40:43.735479 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.735834 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.735994 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 07 14:40:43 crc kubenswrapper[4943]: E0307 14:40:43.736097 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.738171 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.738175 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.738324 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.738429 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.738460 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.738673 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.738916 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.741814 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.743087 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.778829 4943 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.781643 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.789097 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.789167 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.789204 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.789233 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.789296 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.789323 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.789357 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.789393 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.789422 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.789449 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.789472 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.789501 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.789541 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.789565 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.789588 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.789615 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.789641 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.789666 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.789690 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.789717 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.789739 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.789762 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.789785 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.789814 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.789891 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.789914 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.789954 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.789979 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.790002 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.790024 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.790047 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.790071 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.790043 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.790095 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.790096 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.790122 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.790320 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.790258 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.790383 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.790450 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.790491 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.790503 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.790842 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.790885 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.790968 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.791007 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.791043 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.791076 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.791111 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.791142 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.791175 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.791206 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.791238 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.791269 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.791306 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.791342 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.791381 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.791413 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.791448 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.791481 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.791514 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.791547 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.791580 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.791614 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.791649 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.791686 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.791732 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.791772 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.791813 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.791849 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.791891 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.791952 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.792071 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.792103 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.792105 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.792137 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.792150 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.792171 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.792207 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.792231 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:43Z","lastTransitionTime":"2026-03-07T14:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.792171 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.792488 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.792547 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.792598 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.792645 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.792695 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.792741 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.792793 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.792842 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.792890 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.792977 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.793027 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.793071 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.790635 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.790778 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.791112 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.791221 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.791360 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.791571 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.791663 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.792676 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.792806 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.793337 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.792837 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.793077 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.793788 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.794240 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.794390 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.794540 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.794670 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.794983 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.794855 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.795766 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.796259 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.796427 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.796515 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.797348 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.797537 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.797746 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.798076 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.798478 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.798679 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.798778 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.799134 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.799263 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.799503 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.799760 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.800007 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.800106 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.800342 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.800380 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.800531 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.800637 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.800881 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.800888 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.801352 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.801385 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.801774 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.801833 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.801954 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.802028 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.802170 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.802645 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.803820 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.804216 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.804375 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.805084 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.805227 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.805420 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.805692 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.806142 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.806158 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.806207 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.806527 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.806638 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.806834 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.807561 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.808206 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.808266 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.808608 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.808654 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.793120 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.809145 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.809205 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.809261 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.809346 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.809392 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.809453 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.809503 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.809545 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.809594 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.809646 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.809695 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.809732 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.809668 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.809775 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.809822 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.809859 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.809903 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.809993 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.810030 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.810070 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.810120 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.810166 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.810197 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.810246 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.810289 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.810320 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.810363 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.810404 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.810444 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.810453 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.810482 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.810549 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.810682 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.810747 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.810792 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.810832 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.810885 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.810913 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.810849 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.811011 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.811007 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.811053 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.811088 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.811144 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.811188 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.811223 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.811265 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.811302 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.811339 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.811382 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.811423 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.811442 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.811462 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.811598 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.811768 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.811849 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.811922 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.812090 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.812153 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.805087 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.812223 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.812291 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.812348 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.812410 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.812478 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.812541 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.812607 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.812673 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.812739 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.812792 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.812855 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.812918 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.813007 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.813078 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.813142 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.813211 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.812991 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.813307 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.813473 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.813682 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.813681 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.813858 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.813983 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.814046 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.814113 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.814183 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.814245 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.814299 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.814413 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.814627 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.815216 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.815229 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.815797 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.815797 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.815846 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.816501 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.815202 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.816886 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.817097 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.817673 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.818392 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.818449 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.818639 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.818765 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.819136 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.819784 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.819892 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.820259 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.820352 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.820626 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.817836 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.821360 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.821478 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.821553 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.820970 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.822395 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.822584 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.822760 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.822797 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.823073 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: E0307 14:40:43.823110 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:40:44.322820658 +0000 UTC m=+86.274957186 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.823466 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.823606 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.822704 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.824067 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.824199 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.824213 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.824418 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.824464 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.823697 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.824869 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.824888 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.825158 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.825234 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.825349 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.825433 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.825480 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.825487 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.825490 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.825602 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.825650 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.825721 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.826479 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.826497 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.826534 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.826555 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.826581 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.826598 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.826620 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.826719 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.826776 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.826813 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.826851 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.826890 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.826954 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.826968 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.827084 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.827153 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.827164 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.827072 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.827208 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.827495 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.827528 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.827557 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.827585 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.827609 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.827638 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.827667 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.827693 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.827715 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.827738 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.827762 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.827784 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.827806 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.827828 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.827893 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.827921 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.827967 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.827993 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.828020 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.828046 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.828080 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.828104 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.828130 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.828162 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.828176 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.828188 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.828355 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.828401 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.828436 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.828454 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: E0307 14:40:43.828600 4943 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 07 14:40:43 crc kubenswrapper[4943]: E0307 14:40:43.828805 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-07 14:40:44.328782732 +0000 UTC m=+86.280919270 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.828962 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.829150 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.829330 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.829720 4943 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.829909 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.829967 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.830245 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.829958 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.830328 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.831827 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: E0307 14:40:43.832101 4943 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 07 14:40:43 crc kubenswrapper[4943]: E0307 14:40:43.832207 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-07 14:40:44.33218151 +0000 UTC m=+86.284318018 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.832530 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.832587 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.833176 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.833225 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.828443 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.833289 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.833488 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.833652 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.833642 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.833554 4943 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.834036 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.834088 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.834117 4943 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.834146 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.834177 4943 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.834205 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.834233 4943 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.834264 4943 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.834293 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.834321 4943 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.834350 4943 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.834380 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.834397 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.834412 4943 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.834462 4943 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.834487 4943 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.834510 4943 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.834532 4943 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.834553 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.834574 4943 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.834593 4943 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.834614 4943 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.834684 4943 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.834704 4943 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.834732 4943 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.834759 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.834785 4943 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.834584 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.834813 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.834843 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.834869 4943 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.834891 4943 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.834912 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.834963 4943 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.834982 4943 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.835002 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.835022 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.835040 4943 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.835059 4943 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.835078 4943 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.835095 4943 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.835115 4943 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.835133 4943 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.835153 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.835172 4943 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.835191 4943 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.835210 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.835231 4943 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.835251 4943 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.835271 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.835290 4943 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.835309 4943 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.835328 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.835346 4943 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.835365 4943 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.835384 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.835404 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.835423 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.835441 4943 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.835460 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.835482 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.835502 4943 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.835521 4943 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.835541 4943 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.835561 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.835583 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.835603 4943 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.835621 4943 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.835639 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.835658 4943 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.835679 4943 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.835699 4943 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.835456 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.835720 4943 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.835820 4943 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.835849 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.835856 4943 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.835950 4943 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.835982 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.836005 4943 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.836025 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.836044 4943 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.836062 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.836081 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.836103 4943 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.836124 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.836146 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.836165 4943 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.836184 4943 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.836202 4943 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.836221 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.836242 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.836261 4943 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.836281 4943 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.836299 4943 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.836319 4943 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.836341 4943 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.836361 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.836380 4943 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.836400 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.836419 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.836440 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.836458 4943 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.836477 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.836495 4943 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.836515 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.836534 4943 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.836553 4943 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.836573 4943 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.836591 4943 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.836609 4943 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.836628 4943 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.836645 4943 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.836663 4943 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.836681 4943 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.836699 4943 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.836716 4943 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.836736 4943 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.836753 4943 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.836772 4943 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.836791 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.836810 4943 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.836829 4943 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.836848 4943 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.836869 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.836888 4943 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.836906 4943 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.837000 4943 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.837021 4943 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.837039 4943 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.837059 4943 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.837078 4943 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.837098 4943 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.837116 4943 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.837135 4943 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.837156 4943 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.837180 4943 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.837200 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.837218 4943 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.837238 4943 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.837267 4943 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.837286 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.837334 4943 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.837355 4943 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.837376 4943 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.837398 4943 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.837420 4943 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.837442 4943 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.837465 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.837488 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.837510 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.846876 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.849730 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 07 14:40:43 crc kubenswrapper[4943]: E0307 14:40:43.856597 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 07 14:40:43 crc kubenswrapper[4943]: E0307 14:40:43.856646 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 07 14:40:43 crc kubenswrapper[4943]: E0307 14:40:43.856677 4943 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 07 14:40:43 crc kubenswrapper[4943]: E0307 14:40:43.856792 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-07 14:40:44.356762188 +0000 UTC m=+86.308898726 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.858093 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.858378 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: E0307 14:40:43.861821 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 07 14:40:43 crc kubenswrapper[4943]: E0307 14:40:43.861992 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 07 14:40:43 crc kubenswrapper[4943]: E0307 14:40:43.862112 4943 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.862126 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 07 14:40:43 crc kubenswrapper[4943]: E0307 14:40:43.862397 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-07 14:40:44.362249531 +0000 UTC m=+86.314386039 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.862115 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.862183 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.862989 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.864033 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.867461 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.867478 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.867710 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.867786 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.868238 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.868585 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.868738 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.868855 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.869142 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.869302 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.869396 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.869963 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.870097 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.870263 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.870359 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.871082 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.871148 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.871627 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.871909 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.872038 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.872477 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.872633 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.872848 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.872878 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.873287 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.873384 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.873463 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.873913 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.874171 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.874451 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.874563 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.876505 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.877073 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.885724 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.897201 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.897244 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.897281 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.897305 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.897317 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:43Z","lastTransitionTime":"2026-03-07T14:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.897652 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.898730 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.906451 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.915017 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.920522 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.921616 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.938512 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.938573 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.938673 4943 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.938710 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.938747 4943 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.938753 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.938767 4943 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.938817 4943 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.938827 4943 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.938838 4943 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.938848 4943 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.938856 4943 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.938867 4943 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.938876 4943 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.938885 4943 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.938894 4943 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.938903 4943 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.938913 4943 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.938942 4943 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.938952 4943 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.938962 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.938970 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.938979 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.938988 4943 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.939000 4943 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.939009 4943 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.939021 4943 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.939029 4943 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.939038 4943 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.939047 4943 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.939061 4943 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.939071 4943 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.939080 4943 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.939088 4943 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.939097 4943 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.939107 4943 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.939118 4943 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.939127 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.939138 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.939150 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.939160 4943 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.939169 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.939177 4943 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.939187 4943 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.939196 4943 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.939205 4943 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.939213 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.939222 4943 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.939231 4943 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:43 crc kubenswrapper[4943]: I0307 14:40:43.939240 4943 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.000020 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.000061 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.000075 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.000094 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.000106 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:44Z","lastTransitionTime":"2026-03-07T14:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.052676 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.072512 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.077684 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 07 14:40:44 crc kubenswrapper[4943]: W0307 14:40:44.092239 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-4050696d16c78b3b17b81e710b47a5d435b848be8c3b864bdb5ba31ca8b61a51 WatchSource:0}: Error finding container 4050696d16c78b3b17b81e710b47a5d435b848be8c3b864bdb5ba31ca8b61a51: Status 404 returned error can't find the container with id 4050696d16c78b3b17b81e710b47a5d435b848be8c3b864bdb5ba31ca8b61a51 Mar 07 14:40:44 crc kubenswrapper[4943]: W0307 14:40:44.100203 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-cd11de66f478c2497c66463b604445ba3fcb3384c558f833d77f662e6520d874 WatchSource:0}: Error finding container cd11de66f478c2497c66463b604445ba3fcb3384c558f833d77f662e6520d874: Status 404 returned error can't find the container with id cd11de66f478c2497c66463b604445ba3fcb3384c558f833d77f662e6520d874 Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.102081 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.102137 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.102156 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.102185 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.102203 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:44Z","lastTransitionTime":"2026-03-07T14:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.168506 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"5912dd9cbf079ce78858dfb2e7835fd662baa5620505f28913389401508e4ec7"} Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.170417 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"cd11de66f478c2497c66463b604445ba3fcb3384c558f833d77f662e6520d874"} Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.174305 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"4050696d16c78b3b17b81e710b47a5d435b848be8c3b864bdb5ba31ca8b61a51"} Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.207225 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.207268 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.207281 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.207299 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.207312 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:44Z","lastTransitionTime":"2026-03-07T14:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.309733 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.310131 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.310144 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.310177 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.310187 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:44Z","lastTransitionTime":"2026-03-07T14:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.344418 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.344519 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.344583 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 07 14:40:44 crc kubenswrapper[4943]: E0307 14:40:44.344697 4943 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 07 14:40:44 crc kubenswrapper[4943]: E0307 14:40:44.344819 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-07 14:40:45.344758739 +0000 UTC m=+87.296895247 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 07 14:40:44 crc kubenswrapper[4943]: E0307 14:40:44.344910 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:40:45.344898862 +0000 UTC m=+87.297035380 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:40:44 crc kubenswrapper[4943]: E0307 14:40:44.345077 4943 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 07 14:40:44 crc kubenswrapper[4943]: E0307 14:40:44.345156 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-07 14:40:45.345115958 +0000 UTC m=+87.297252466 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.413186 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.413245 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.413260 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.413283 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.413301 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:44Z","lastTransitionTime":"2026-03-07T14:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.445699 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.445756 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 07 14:40:44 crc kubenswrapper[4943]: E0307 14:40:44.445965 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 07 14:40:44 crc kubenswrapper[4943]: E0307 14:40:44.445995 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 07 14:40:44 crc kubenswrapper[4943]: E0307 14:40:44.446013 4943 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 07 14:40:44 crc kubenswrapper[4943]: E0307 14:40:44.446112 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-07 14:40:45.446061827 +0000 UTC m=+87.398198335 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 07 14:40:44 crc kubenswrapper[4943]: E0307 14:40:44.446074 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 07 14:40:44 crc kubenswrapper[4943]: E0307 14:40:44.446167 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 07 14:40:44 crc kubenswrapper[4943]: E0307 14:40:44.446190 4943 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 07 14:40:44 crc kubenswrapper[4943]: E0307 14:40:44.446297 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-07 14:40:45.446268642 +0000 UTC m=+87.398405180 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.517428 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.517510 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.517539 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.517578 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.517602 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:44Z","lastTransitionTime":"2026-03-07T14:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.620137 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.620177 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.620186 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.620201 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.620209 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:44Z","lastTransitionTime":"2026-03-07T14:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.723757 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.723807 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.723830 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.723854 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.723874 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:44Z","lastTransitionTime":"2026-03-07T14:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.762572 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.764169 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.766385 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.767966 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.770698 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.771841 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.773111 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.775132 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.776398 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.778449 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.779913 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.782085 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.784128 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.785189 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.787031 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.788161 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.789277 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.790811 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.791560 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.792414 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.793594 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.794324 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.795445 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.796322 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.797436 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.798260 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.799650 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.800269 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.800996 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.802090 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.802771 4943 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.802907 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.805742 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.806456 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.807010 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.809556 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.810443 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.811179 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.812463 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.813800 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.814483 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.815298 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.816637 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.817905 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.818522 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.819860 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.820581 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.822105 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.822769 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.824146 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.824823 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.825644 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.827174 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.827390 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.827427 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.827439 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.827458 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.827469 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:44Z","lastTransitionTime":"2026-03-07T14:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.828296 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.930982 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.931039 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.931058 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.931085 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:44 crc kubenswrapper[4943]: I0307 14:40:44.931111 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:44Z","lastTransitionTime":"2026-03-07T14:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.021298 4943 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.034017 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.034067 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.034078 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.034100 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.034113 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:45Z","lastTransitionTime":"2026-03-07T14:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.136120 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.136181 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.136198 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.136226 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.136247 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:45Z","lastTransitionTime":"2026-03-07T14:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.179204 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"d2d49d38a953556b474792c0b80484611c006f46559703b1efbe6f46f1c25a85"} Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.179266 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"76f6bdfcc76046424db35a1dbc48c159d3fecefe53c7c795f384b62b14a90455"} Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.181121 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"017f9a3784e27af2c0edd259774ae6debb29257a356552f7db32bf53d864fcab"} Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.192335 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:40:45Z is after 2025-08-24T17:21:41Z" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.211455 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:40:45Z is after 2025-08-24T17:21:41Z" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.227333 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:40:45Z is after 2025-08-24T17:21:41Z" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.239030 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:40:45Z is after 2025-08-24T17:21:41Z" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.240272 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.240308 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.240356 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.240376 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.240426 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:45Z","lastTransitionTime":"2026-03-07T14:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.254264 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d49d38a953556b474792c0b80484611c006f46559703b1efbe6f46f1c25a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-07T14:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76f6bdfcc76046424db35a1dbc48c159d3fecefe53c7c795f384b62b14a90455\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-07T14:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:40:45Z is after 2025-08-24T17:21:41Z" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.268836 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:40:45Z is after 2025-08-24T17:21:41Z" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.287717 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:40:45Z is after 2025-08-24T17:21:41Z" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.303051 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:40:45Z is after 2025-08-24T17:21:41Z" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.316652 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:40:45Z is after 2025-08-24T17:21:41Z" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.328259 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:40:45Z is after 2025-08-24T17:21:41Z" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.343572 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.343600 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.343609 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.343625 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.343634 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:45Z","lastTransitionTime":"2026-03-07T14:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.343626 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d49d38a953556b474792c0b80484611c006f46559703b1efbe6f46f1c25a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-07T14:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76f6bdfcc76046424db35a1dbc48c159d3fecefe53c7c795f384b62b14a90455\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-07T14:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:40:45Z is after 2025-08-24T17:21:41Z" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.356438 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.356488 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.356521 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 07 14:40:45 crc kubenswrapper[4943]: E0307 14:40:45.356612 4943 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 07 14:40:45 crc kubenswrapper[4943]: E0307 14:40:45.356625 4943 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 07 14:40:45 crc kubenswrapper[4943]: E0307 14:40:45.356659 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-07 14:40:47.356644401 +0000 UTC m=+89.308780899 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 07 14:40:45 crc kubenswrapper[4943]: E0307 14:40:45.356702 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:40:47.356665861 +0000 UTC m=+89.308802379 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:40:45 crc kubenswrapper[4943]: E0307 14:40:45.356741 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-07 14:40:47.356731353 +0000 UTC m=+89.308867861 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.360013 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://017f9a3784e27af2c0edd259774ae6debb29257a356552f7db32bf53d864fcab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-07T14:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:40:45Z is after 2025-08-24T17:21:41Z" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.447206 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.447235 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.447243 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.447259 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.447268 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:45Z","lastTransitionTime":"2026-03-07T14:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.457809 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.457915 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 07 14:40:45 crc kubenswrapper[4943]: E0307 14:40:45.457996 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 07 14:40:45 crc kubenswrapper[4943]: E0307 14:40:45.458021 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 07 14:40:45 crc kubenswrapper[4943]: E0307 14:40:45.458033 4943 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 07 14:40:45 crc kubenswrapper[4943]: E0307 14:40:45.458079 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-07 14:40:47.458062782 +0000 UTC m=+89.410199280 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 07 14:40:45 crc kubenswrapper[4943]: E0307 14:40:45.458148 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 07 14:40:45 crc kubenswrapper[4943]: E0307 14:40:45.458184 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 07 14:40:45 crc kubenswrapper[4943]: E0307 14:40:45.458210 4943 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 07 14:40:45 crc kubenswrapper[4943]: E0307 14:40:45.458313 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-07 14:40:47.458280788 +0000 UTC m=+89.410417326 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.549287 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.549320 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.549331 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.549345 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.549353 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:45Z","lastTransitionTime":"2026-03-07T14:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.651764 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.651801 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.651811 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.651827 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.651839 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:45Z","lastTransitionTime":"2026-03-07T14:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.754157 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.754247 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.754264 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.754275 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.754292 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.754292 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.754304 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:45Z","lastTransitionTime":"2026-03-07T14:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:45 crc kubenswrapper[4943]: E0307 14:40:45.754528 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.754556 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 07 14:40:45 crc kubenswrapper[4943]: E0307 14:40:45.754640 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 07 14:40:45 crc kubenswrapper[4943]: E0307 14:40:45.754772 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.856888 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.856917 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.856941 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.856956 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.856964 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:45Z","lastTransitionTime":"2026-03-07T14:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.959848 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.959943 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.959955 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.959972 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:45 crc kubenswrapper[4943]: I0307 14:40:45.959982 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:45Z","lastTransitionTime":"2026-03-07T14:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.043944 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.044098 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.044123 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.044179 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.044211 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:46Z","lastTransitionTime":"2026-03-07T14:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:46 crc kubenswrapper[4943]: E0307 14:40:46.066126 4943 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-07T14:40:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-07T14:40:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-07T14:40:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-07T14:40:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7a039d39-2e22-491c-a9ee-09ff4d277e8b\\\",\\\"systemUUID\\\":\\\"bf22d61b-d7ce-44b5-9dd8-1a46c0f78104\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:40:46Z is after 2025-08-24T17:21:41Z" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.070964 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.071005 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.071019 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.071040 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.071056 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:46Z","lastTransitionTime":"2026-03-07T14:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:46 crc kubenswrapper[4943]: E0307 14:40:46.092470 4943 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-07T14:40:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-07T14:40:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-07T14:40:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-07T14:40:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7a039d39-2e22-491c-a9ee-09ff4d277e8b\\\",\\\"systemUUID\\\":\\\"bf22d61b-d7ce-44b5-9dd8-1a46c0f78104\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:40:46Z is after 2025-08-24T17:21:41Z" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.097240 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.097329 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.097352 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.097384 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.097408 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:46Z","lastTransitionTime":"2026-03-07T14:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:46 crc kubenswrapper[4943]: E0307 14:40:46.117400 4943 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-07T14:40:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-07T14:40:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-07T14:40:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-07T14:40:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7a039d39-2e22-491c-a9ee-09ff4d277e8b\\\",\\\"systemUUID\\\":\\\"bf22d61b-d7ce-44b5-9dd8-1a46c0f78104\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:40:46Z is after 2025-08-24T17:21:41Z" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.122798 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.122864 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.122875 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.122899 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.122912 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:46Z","lastTransitionTime":"2026-03-07T14:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:46 crc kubenswrapper[4943]: E0307 14:40:46.142147 4943 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-07T14:40:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-07T14:40:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-07T14:40:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-07T14:40:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7a039d39-2e22-491c-a9ee-09ff4d277e8b\\\",\\\"systemUUID\\\":\\\"bf22d61b-d7ce-44b5-9dd8-1a46c0f78104\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:40:46Z is after 2025-08-24T17:21:41Z" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.146765 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.146820 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.146832 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.146857 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.146870 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:46Z","lastTransitionTime":"2026-03-07T14:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:46 crc kubenswrapper[4943]: E0307 14:40:46.165789 4943 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-07T14:40:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-07T14:40:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-07T14:40:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-07T14:40:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-07T14:40:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7a039d39-2e22-491c-a9ee-09ff4d277e8b\\\",\\\"systemUUID\\\":\\\"bf22d61b-d7ce-44b5-9dd8-1a46c0f78104\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-07T14:40:46Z is after 2025-08-24T17:21:41Z" Mar 07 14:40:46 crc kubenswrapper[4943]: E0307 14:40:46.166089 4943 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.168241 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.168292 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.168315 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.168347 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.168373 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:46Z","lastTransitionTime":"2026-03-07T14:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.271150 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.271210 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.271227 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.271549 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.271593 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:46Z","lastTransitionTime":"2026-03-07T14:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.374592 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.374659 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.374685 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.374719 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.374747 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:46Z","lastTransitionTime":"2026-03-07T14:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.477636 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.477718 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.477741 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.477775 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.477798 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:46Z","lastTransitionTime":"2026-03-07T14:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.580245 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.580303 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.580320 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.580345 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.580362 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:46Z","lastTransitionTime":"2026-03-07T14:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.683746 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.683793 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.683805 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.683824 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.683836 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:46Z","lastTransitionTime":"2026-03-07T14:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.787059 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.787109 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.787120 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.787138 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.787150 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:46Z","lastTransitionTime":"2026-03-07T14:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.889693 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.889754 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.889777 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.889810 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.889835 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:46Z","lastTransitionTime":"2026-03-07T14:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.993709 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.993799 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.993819 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.993846 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:46 crc kubenswrapper[4943]: I0307 14:40:46.993864 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:46Z","lastTransitionTime":"2026-03-07T14:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:47 crc kubenswrapper[4943]: I0307 14:40:47.096631 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:47 crc kubenswrapper[4943]: I0307 14:40:47.096708 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:47 crc kubenswrapper[4943]: I0307 14:40:47.096726 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:47 crc kubenswrapper[4943]: I0307 14:40:47.096752 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:47 crc kubenswrapper[4943]: I0307 14:40:47.096769 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:47Z","lastTransitionTime":"2026-03-07T14:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:47 crc kubenswrapper[4943]: I0307 14:40:47.199539 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:47 crc kubenswrapper[4943]: I0307 14:40:47.199588 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:47 crc kubenswrapper[4943]: I0307 14:40:47.199600 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:47 crc kubenswrapper[4943]: I0307 14:40:47.199618 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:47 crc kubenswrapper[4943]: I0307 14:40:47.199631 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:47Z","lastTransitionTime":"2026-03-07T14:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:47 crc kubenswrapper[4943]: I0307 14:40:47.302222 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:47 crc kubenswrapper[4943]: I0307 14:40:47.302296 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:47 crc kubenswrapper[4943]: I0307 14:40:47.302316 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:47 crc kubenswrapper[4943]: I0307 14:40:47.302344 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:47 crc kubenswrapper[4943]: I0307 14:40:47.302362 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:47Z","lastTransitionTime":"2026-03-07T14:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:47 crc kubenswrapper[4943]: I0307 14:40:47.374763 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:40:47 crc kubenswrapper[4943]: I0307 14:40:47.374883 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 07 14:40:47 crc kubenswrapper[4943]: E0307 14:40:47.375122 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:40:51.375085576 +0000 UTC m=+93.327222084 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:40:47 crc kubenswrapper[4943]: E0307 14:40:47.375167 4943 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 07 14:40:47 crc kubenswrapper[4943]: I0307 14:40:47.375235 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 07 14:40:47 crc kubenswrapper[4943]: E0307 14:40:47.375249 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-07 14:40:51.37523195 +0000 UTC m=+93.327368508 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 07 14:40:47 crc kubenswrapper[4943]: E0307 14:40:47.375387 4943 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 07 14:40:47 crc kubenswrapper[4943]: E0307 14:40:47.375461 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-07 14:40:51.375446375 +0000 UTC m=+93.327582893 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 07 14:40:47 crc kubenswrapper[4943]: I0307 14:40:47.405460 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:47 crc kubenswrapper[4943]: I0307 14:40:47.405509 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:47 crc kubenswrapper[4943]: I0307 14:40:47.405526 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:47 crc kubenswrapper[4943]: I0307 14:40:47.405550 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:47 crc kubenswrapper[4943]: I0307 14:40:47.405567 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:47Z","lastTransitionTime":"2026-03-07T14:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:47 crc kubenswrapper[4943]: I0307 14:40:47.476387 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 07 14:40:47 crc kubenswrapper[4943]: I0307 14:40:47.476469 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 07 14:40:47 crc kubenswrapper[4943]: E0307 14:40:47.476712 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 07 14:40:47 crc kubenswrapper[4943]: E0307 14:40:47.476741 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 07 14:40:47 crc kubenswrapper[4943]: E0307 14:40:47.476764 4943 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 07 14:40:47 crc kubenswrapper[4943]: E0307 14:40:47.476842 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-07 14:40:51.476816225 +0000 UTC m=+93.428952763 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 07 14:40:47 crc kubenswrapper[4943]: E0307 14:40:47.477018 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 07 14:40:47 crc kubenswrapper[4943]: E0307 14:40:47.477230 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 07 14:40:47 crc kubenswrapper[4943]: E0307 14:40:47.477246 4943 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 07 14:40:47 crc kubenswrapper[4943]: E0307 14:40:47.477294 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-07 14:40:51.477280477 +0000 UTC m=+93.429417005 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 07 14:40:47 crc kubenswrapper[4943]: I0307 14:40:47.508053 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:47 crc kubenswrapper[4943]: I0307 14:40:47.508123 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:47 crc kubenswrapper[4943]: I0307 14:40:47.508332 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:47 crc kubenswrapper[4943]: I0307 14:40:47.508364 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:47 crc kubenswrapper[4943]: I0307 14:40:47.508388 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:47Z","lastTransitionTime":"2026-03-07T14:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:47 crc kubenswrapper[4943]: I0307 14:40:47.611368 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:47 crc kubenswrapper[4943]: I0307 14:40:47.611419 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:47 crc kubenswrapper[4943]: I0307 14:40:47.611433 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:47 crc kubenswrapper[4943]: I0307 14:40:47.611454 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:47 crc kubenswrapper[4943]: I0307 14:40:47.611469 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:47Z","lastTransitionTime":"2026-03-07T14:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:47 crc kubenswrapper[4943]: I0307 14:40:47.715113 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:47 crc kubenswrapper[4943]: I0307 14:40:47.715178 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:47 crc kubenswrapper[4943]: I0307 14:40:47.715197 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:47 crc kubenswrapper[4943]: I0307 14:40:47.715222 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:47 crc kubenswrapper[4943]: I0307 14:40:47.715242 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:47Z","lastTransitionTime":"2026-03-07T14:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:47 crc kubenswrapper[4943]: I0307 14:40:47.754637 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 07 14:40:47 crc kubenswrapper[4943]: I0307 14:40:47.754637 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 07 14:40:47 crc kubenswrapper[4943]: E0307 14:40:47.754812 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 07 14:40:47 crc kubenswrapper[4943]: I0307 14:40:47.754644 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 07 14:40:47 crc kubenswrapper[4943]: E0307 14:40:47.755059 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 07 14:40:47 crc kubenswrapper[4943]: E0307 14:40:47.755181 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 07 14:40:47 crc kubenswrapper[4943]: I0307 14:40:47.818294 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:47 crc kubenswrapper[4943]: I0307 14:40:47.818364 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:47 crc kubenswrapper[4943]: I0307 14:40:47.818383 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:47 crc kubenswrapper[4943]: I0307 14:40:47.818410 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:47 crc kubenswrapper[4943]: I0307 14:40:47.818428 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:47Z","lastTransitionTime":"2026-03-07T14:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:47 crc kubenswrapper[4943]: I0307 14:40:47.921236 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:47 crc kubenswrapper[4943]: I0307 14:40:47.921293 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:47 crc kubenswrapper[4943]: I0307 14:40:47.921310 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:47 crc kubenswrapper[4943]: I0307 14:40:47.921338 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:47 crc kubenswrapper[4943]: I0307 14:40:47.921354 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:47Z","lastTransitionTime":"2026-03-07T14:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:48 crc kubenswrapper[4943]: I0307 14:40:48.024697 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:48 crc kubenswrapper[4943]: I0307 14:40:48.024743 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:48 crc kubenswrapper[4943]: I0307 14:40:48.024752 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:48 crc kubenswrapper[4943]: I0307 14:40:48.024770 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:48 crc kubenswrapper[4943]: I0307 14:40:48.024786 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:48Z","lastTransitionTime":"2026-03-07T14:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:48 crc kubenswrapper[4943]: I0307 14:40:48.127533 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:48 crc kubenswrapper[4943]: I0307 14:40:48.127581 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:48 crc kubenswrapper[4943]: I0307 14:40:48.127597 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:48 crc kubenswrapper[4943]: I0307 14:40:48.127620 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:48 crc kubenswrapper[4943]: I0307 14:40:48.127636 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:48Z","lastTransitionTime":"2026-03-07T14:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:48 crc kubenswrapper[4943]: I0307 14:40:48.195033 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"1fbe6bc080dc9a5645975fd5d902b440d987ee0c68e7ce7b452b60e526c3d415"} Mar 07 14:40:48 crc kubenswrapper[4943]: I0307 14:40:48.231250 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:48 crc kubenswrapper[4943]: I0307 14:40:48.231304 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:48 crc kubenswrapper[4943]: I0307 14:40:48.231317 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:48 crc kubenswrapper[4943]: I0307 14:40:48.231341 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:48 crc kubenswrapper[4943]: I0307 14:40:48.231355 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:48Z","lastTransitionTime":"2026-03-07T14:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:48 crc kubenswrapper[4943]: I0307 14:40:48.333988 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:48 crc kubenswrapper[4943]: I0307 14:40:48.334044 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:48 crc kubenswrapper[4943]: I0307 14:40:48.334063 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:48 crc kubenswrapper[4943]: I0307 14:40:48.334088 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:48 crc kubenswrapper[4943]: I0307 14:40:48.334108 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:48Z","lastTransitionTime":"2026-03-07T14:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:48 crc kubenswrapper[4943]: I0307 14:40:48.436562 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:48 crc kubenswrapper[4943]: I0307 14:40:48.436601 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:48 crc kubenswrapper[4943]: I0307 14:40:48.436609 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:48 crc kubenswrapper[4943]: I0307 14:40:48.436625 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:48 crc kubenswrapper[4943]: I0307 14:40:48.436633 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:48Z","lastTransitionTime":"2026-03-07T14:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:48 crc kubenswrapper[4943]: I0307 14:40:48.539915 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:48 crc kubenswrapper[4943]: I0307 14:40:48.539968 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:48 crc kubenswrapper[4943]: I0307 14:40:48.539980 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:48 crc kubenswrapper[4943]: I0307 14:40:48.539996 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:48 crc kubenswrapper[4943]: I0307 14:40:48.540005 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:48Z","lastTransitionTime":"2026-03-07T14:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:48 crc kubenswrapper[4943]: I0307 14:40:48.642400 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:48 crc kubenswrapper[4943]: I0307 14:40:48.642447 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:48 crc kubenswrapper[4943]: I0307 14:40:48.642460 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:48 crc kubenswrapper[4943]: I0307 14:40:48.642477 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:48 crc kubenswrapper[4943]: I0307 14:40:48.642488 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:48Z","lastTransitionTime":"2026-03-07T14:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:48 crc kubenswrapper[4943]: I0307 14:40:48.745367 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:48 crc kubenswrapper[4943]: I0307 14:40:48.745423 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:48 crc kubenswrapper[4943]: I0307 14:40:48.745440 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:48 crc kubenswrapper[4943]: I0307 14:40:48.745464 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:48 crc kubenswrapper[4943]: I0307 14:40:48.745481 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:48Z","lastTransitionTime":"2026-03-07T14:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:48 crc kubenswrapper[4943]: I0307 14:40:48.848840 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:48 crc kubenswrapper[4943]: I0307 14:40:48.848914 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:48 crc kubenswrapper[4943]: I0307 14:40:48.848966 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:48 crc kubenswrapper[4943]: I0307 14:40:48.848999 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:48 crc kubenswrapper[4943]: I0307 14:40:48.849020 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:48Z","lastTransitionTime":"2026-03-07T14:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:48 crc kubenswrapper[4943]: I0307 14:40:48.953201 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:48 crc kubenswrapper[4943]: I0307 14:40:48.953729 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:48 crc kubenswrapper[4943]: I0307 14:40:48.953760 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:48 crc kubenswrapper[4943]: I0307 14:40:48.953800 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:48 crc kubenswrapper[4943]: I0307 14:40:48.953822 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:48Z","lastTransitionTime":"2026-03-07T14:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:49 crc kubenswrapper[4943]: I0307 14:40:49.057124 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:49 crc kubenswrapper[4943]: I0307 14:40:49.057183 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:49 crc kubenswrapper[4943]: I0307 14:40:49.057203 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:49 crc kubenswrapper[4943]: I0307 14:40:49.057231 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:49 crc kubenswrapper[4943]: I0307 14:40:49.057248 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:49Z","lastTransitionTime":"2026-03-07T14:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:49 crc kubenswrapper[4943]: I0307 14:40:49.161488 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:49 crc kubenswrapper[4943]: I0307 14:40:49.161561 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:49 crc kubenswrapper[4943]: I0307 14:40:49.161583 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:49 crc kubenswrapper[4943]: I0307 14:40:49.161662 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:49 crc kubenswrapper[4943]: I0307 14:40:49.161688 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:49Z","lastTransitionTime":"2026-03-07T14:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:49 crc kubenswrapper[4943]: I0307 14:40:49.266430 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:49 crc kubenswrapper[4943]: I0307 14:40:49.266538 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:49 crc kubenswrapper[4943]: I0307 14:40:49.266556 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:49 crc kubenswrapper[4943]: I0307 14:40:49.266626 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:49 crc kubenswrapper[4943]: I0307 14:40:49.266647 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:49Z","lastTransitionTime":"2026-03-07T14:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:49 crc kubenswrapper[4943]: I0307 14:40:49.369386 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:49 crc kubenswrapper[4943]: I0307 14:40:49.369457 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:49 crc kubenswrapper[4943]: I0307 14:40:49.369481 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:49 crc kubenswrapper[4943]: I0307 14:40:49.369518 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:49 crc kubenswrapper[4943]: I0307 14:40:49.369540 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:49Z","lastTransitionTime":"2026-03-07T14:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:49 crc kubenswrapper[4943]: I0307 14:40:49.472119 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:49 crc kubenswrapper[4943]: I0307 14:40:49.472170 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:49 crc kubenswrapper[4943]: I0307 14:40:49.472188 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:49 crc kubenswrapper[4943]: I0307 14:40:49.472212 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:49 crc kubenswrapper[4943]: I0307 14:40:49.472227 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:49Z","lastTransitionTime":"2026-03-07T14:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:49 crc kubenswrapper[4943]: I0307 14:40:49.575627 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:49 crc kubenswrapper[4943]: I0307 14:40:49.575680 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:49 crc kubenswrapper[4943]: I0307 14:40:49.575699 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:49 crc kubenswrapper[4943]: I0307 14:40:49.575723 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:49 crc kubenswrapper[4943]: I0307 14:40:49.575743 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:49Z","lastTransitionTime":"2026-03-07T14:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:49 crc kubenswrapper[4943]: I0307 14:40:49.678337 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:49 crc kubenswrapper[4943]: I0307 14:40:49.678381 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:49 crc kubenswrapper[4943]: I0307 14:40:49.678392 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:49 crc kubenswrapper[4943]: I0307 14:40:49.678411 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:49 crc kubenswrapper[4943]: I0307 14:40:49.678425 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:49Z","lastTransitionTime":"2026-03-07T14:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:49 crc kubenswrapper[4943]: I0307 14:40:49.754622 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 07 14:40:49 crc kubenswrapper[4943]: I0307 14:40:49.754653 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 07 14:40:49 crc kubenswrapper[4943]: I0307 14:40:49.754703 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 07 14:40:49 crc kubenswrapper[4943]: E0307 14:40:49.754785 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 07 14:40:49 crc kubenswrapper[4943]: E0307 14:40:49.754876 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 07 14:40:49 crc kubenswrapper[4943]: E0307 14:40:49.755064 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 07 14:40:49 crc kubenswrapper[4943]: I0307 14:40:49.781131 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:49 crc kubenswrapper[4943]: I0307 14:40:49.781170 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:49 crc kubenswrapper[4943]: I0307 14:40:49.781181 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:49 crc kubenswrapper[4943]: I0307 14:40:49.781197 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:49 crc kubenswrapper[4943]: I0307 14:40:49.781206 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:49Z","lastTransitionTime":"2026-03-07T14:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:49 crc kubenswrapper[4943]: I0307 14:40:49.884818 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:49 crc kubenswrapper[4943]: I0307 14:40:49.884873 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:49 crc kubenswrapper[4943]: I0307 14:40:49.884889 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:49 crc kubenswrapper[4943]: I0307 14:40:49.884915 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:49 crc kubenswrapper[4943]: I0307 14:40:49.884965 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:49Z","lastTransitionTime":"2026-03-07T14:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:49 crc kubenswrapper[4943]: I0307 14:40:49.988348 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:49 crc kubenswrapper[4943]: I0307 14:40:49.988448 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:49 crc kubenswrapper[4943]: I0307 14:40:49.988506 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:49 crc kubenswrapper[4943]: I0307 14:40:49.988533 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:49 crc kubenswrapper[4943]: I0307 14:40:49.988599 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:49Z","lastTransitionTime":"2026-03-07T14:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:50 crc kubenswrapper[4943]: I0307 14:40:50.091752 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:50 crc kubenswrapper[4943]: I0307 14:40:50.091809 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:50 crc kubenswrapper[4943]: I0307 14:40:50.091826 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:50 crc kubenswrapper[4943]: I0307 14:40:50.091849 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:50 crc kubenswrapper[4943]: I0307 14:40:50.091865 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:50Z","lastTransitionTime":"2026-03-07T14:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:50 crc kubenswrapper[4943]: I0307 14:40:50.195453 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:50 crc kubenswrapper[4943]: I0307 14:40:50.195513 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:50 crc kubenswrapper[4943]: I0307 14:40:50.195536 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:50 crc kubenswrapper[4943]: I0307 14:40:50.195561 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:50 crc kubenswrapper[4943]: I0307 14:40:50.195579 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:50Z","lastTransitionTime":"2026-03-07T14:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:50 crc kubenswrapper[4943]: I0307 14:40:50.298328 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:50 crc kubenswrapper[4943]: I0307 14:40:50.298401 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:50 crc kubenswrapper[4943]: I0307 14:40:50.298419 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:50 crc kubenswrapper[4943]: I0307 14:40:50.298455 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:50 crc kubenswrapper[4943]: I0307 14:40:50.298473 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:50Z","lastTransitionTime":"2026-03-07T14:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:50 crc kubenswrapper[4943]: I0307 14:40:50.400772 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:50 crc kubenswrapper[4943]: I0307 14:40:50.400801 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:50 crc kubenswrapper[4943]: I0307 14:40:50.400809 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:50 crc kubenswrapper[4943]: I0307 14:40:50.400822 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:50 crc kubenswrapper[4943]: I0307 14:40:50.400831 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:50Z","lastTransitionTime":"2026-03-07T14:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:50 crc kubenswrapper[4943]: I0307 14:40:50.503762 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:50 crc kubenswrapper[4943]: I0307 14:40:50.503848 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:50 crc kubenswrapper[4943]: I0307 14:40:50.503878 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:50 crc kubenswrapper[4943]: I0307 14:40:50.503908 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:50 crc kubenswrapper[4943]: I0307 14:40:50.503966 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:50Z","lastTransitionTime":"2026-03-07T14:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:50 crc kubenswrapper[4943]: I0307 14:40:50.606385 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:50 crc kubenswrapper[4943]: I0307 14:40:50.606427 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:50 crc kubenswrapper[4943]: I0307 14:40:50.606440 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:50 crc kubenswrapper[4943]: I0307 14:40:50.606458 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:50 crc kubenswrapper[4943]: I0307 14:40:50.606469 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:50Z","lastTransitionTime":"2026-03-07T14:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:50 crc kubenswrapper[4943]: I0307 14:40:50.709540 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:50 crc kubenswrapper[4943]: I0307 14:40:50.709618 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:50 crc kubenswrapper[4943]: I0307 14:40:50.709635 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:50 crc kubenswrapper[4943]: I0307 14:40:50.709663 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:50 crc kubenswrapper[4943]: I0307 14:40:50.709681 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:50Z","lastTransitionTime":"2026-03-07T14:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:50 crc kubenswrapper[4943]: I0307 14:40:50.811756 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:50 crc kubenswrapper[4943]: I0307 14:40:50.811801 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:50 crc kubenswrapper[4943]: I0307 14:40:50.811813 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:50 crc kubenswrapper[4943]: I0307 14:40:50.811831 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:50 crc kubenswrapper[4943]: I0307 14:40:50.811845 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:50Z","lastTransitionTime":"2026-03-07T14:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:50 crc kubenswrapper[4943]: I0307 14:40:50.915061 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:50 crc kubenswrapper[4943]: I0307 14:40:50.915129 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:50 crc kubenswrapper[4943]: I0307 14:40:50.915168 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:50 crc kubenswrapper[4943]: I0307 14:40:50.915201 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:50 crc kubenswrapper[4943]: I0307 14:40:50.915229 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:50Z","lastTransitionTime":"2026-03-07T14:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.018681 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.018739 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.018756 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.018782 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.018799 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:51Z","lastTransitionTime":"2026-03-07T14:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.122458 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.122510 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.122523 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.122544 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.122555 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:51Z","lastTransitionTime":"2026-03-07T14:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.225221 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.225280 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.225298 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.225321 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.225339 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:51Z","lastTransitionTime":"2026-03-07T14:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.327962 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.328008 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.328018 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.328040 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.328054 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:51Z","lastTransitionTime":"2026-03-07T14:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.415227 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.415336 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.415410 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 07 14:40:51 crc kubenswrapper[4943]: E0307 14:40:51.415553 4943 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 07 14:40:51 crc kubenswrapper[4943]: E0307 14:40:51.415560 4943 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 07 14:40:51 crc kubenswrapper[4943]: E0307 14:40:51.415607 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:40:59.4155636 +0000 UTC m=+101.367700138 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:40:51 crc kubenswrapper[4943]: E0307 14:40:51.415662 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-07 14:40:59.415647482 +0000 UTC m=+101.367784010 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 07 14:40:51 crc kubenswrapper[4943]: E0307 14:40:51.415692 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-07 14:40:59.415677653 +0000 UTC m=+101.367814181 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.431153 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.431214 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.431231 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.431258 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.431274 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:51Z","lastTransitionTime":"2026-03-07T14:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.516306 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.516382 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 07 14:40:51 crc kubenswrapper[4943]: E0307 14:40:51.516560 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 07 14:40:51 crc kubenswrapper[4943]: E0307 14:40:51.516588 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 07 14:40:51 crc kubenswrapper[4943]: E0307 14:40:51.516609 4943 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 07 14:40:51 crc kubenswrapper[4943]: E0307 14:40:51.516636 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 07 14:40:51 crc kubenswrapper[4943]: E0307 14:40:51.516681 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 07 14:40:51 crc kubenswrapper[4943]: E0307 14:40:51.516693 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-07 14:40:59.516669913 +0000 UTC m=+101.468806441 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 07 14:40:51 crc kubenswrapper[4943]: E0307 14:40:51.516702 4943 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 07 14:40:51 crc kubenswrapper[4943]: E0307 14:40:51.516789 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-07 14:40:59.516762816 +0000 UTC m=+101.468899354 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.534702 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.534810 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.534832 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.534865 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.534885 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:51Z","lastTransitionTime":"2026-03-07T14:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.637709 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.637771 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.637788 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.637818 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.637836 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:51Z","lastTransitionTime":"2026-03-07T14:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.741137 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.741199 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.741219 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.741246 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.741266 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:51Z","lastTransitionTime":"2026-03-07T14:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.754636 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.754703 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.754702 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 07 14:40:51 crc kubenswrapper[4943]: E0307 14:40:51.754817 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 07 14:40:51 crc kubenswrapper[4943]: E0307 14:40:51.754910 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 07 14:40:51 crc kubenswrapper[4943]: E0307 14:40:51.755090 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.844905 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.845021 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.845044 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.845080 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.845118 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:51Z","lastTransitionTime":"2026-03-07T14:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.948551 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.948617 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.948637 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.948804 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:51 crc kubenswrapper[4943]: I0307 14:40:51.948826 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:51Z","lastTransitionTime":"2026-03-07T14:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:52 crc kubenswrapper[4943]: I0307 14:40:52.052451 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:52 crc kubenswrapper[4943]: I0307 14:40:52.052528 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:52 crc kubenswrapper[4943]: I0307 14:40:52.052551 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:52 crc kubenswrapper[4943]: I0307 14:40:52.052583 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:52 crc kubenswrapper[4943]: I0307 14:40:52.052601 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:52Z","lastTransitionTime":"2026-03-07T14:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:52 crc kubenswrapper[4943]: I0307 14:40:52.156796 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:52 crc kubenswrapper[4943]: I0307 14:40:52.156849 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:52 crc kubenswrapper[4943]: I0307 14:40:52.156862 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:52 crc kubenswrapper[4943]: I0307 14:40:52.156887 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:52 crc kubenswrapper[4943]: I0307 14:40:52.156902 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:52Z","lastTransitionTime":"2026-03-07T14:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:52 crc kubenswrapper[4943]: I0307 14:40:52.260317 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:52 crc kubenswrapper[4943]: I0307 14:40:52.260388 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:52 crc kubenswrapper[4943]: I0307 14:40:52.260411 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:52 crc kubenswrapper[4943]: I0307 14:40:52.260534 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:52 crc kubenswrapper[4943]: I0307 14:40:52.260569 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:52Z","lastTransitionTime":"2026-03-07T14:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:52 crc kubenswrapper[4943]: I0307 14:40:52.363573 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:52 crc kubenswrapper[4943]: I0307 14:40:52.363614 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:52 crc kubenswrapper[4943]: I0307 14:40:52.363624 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:52 crc kubenswrapper[4943]: I0307 14:40:52.363640 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:52 crc kubenswrapper[4943]: I0307 14:40:52.363651 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:52Z","lastTransitionTime":"2026-03-07T14:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:52 crc kubenswrapper[4943]: I0307 14:40:52.466660 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:52 crc kubenswrapper[4943]: I0307 14:40:52.466717 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:52 crc kubenswrapper[4943]: I0307 14:40:52.466733 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:52 crc kubenswrapper[4943]: I0307 14:40:52.466758 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:52 crc kubenswrapper[4943]: I0307 14:40:52.466774 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:52Z","lastTransitionTime":"2026-03-07T14:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:52 crc kubenswrapper[4943]: I0307 14:40:52.569626 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:52 crc kubenswrapper[4943]: I0307 14:40:52.569704 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:52 crc kubenswrapper[4943]: I0307 14:40:52.569728 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:52 crc kubenswrapper[4943]: I0307 14:40:52.569762 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:52 crc kubenswrapper[4943]: I0307 14:40:52.569789 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:52Z","lastTransitionTime":"2026-03-07T14:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:52 crc kubenswrapper[4943]: I0307 14:40:52.673013 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:52 crc kubenswrapper[4943]: I0307 14:40:52.673089 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:52 crc kubenswrapper[4943]: I0307 14:40:52.673112 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:52 crc kubenswrapper[4943]: I0307 14:40:52.673143 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:52 crc kubenswrapper[4943]: I0307 14:40:52.673183 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:52Z","lastTransitionTime":"2026-03-07T14:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:52 crc kubenswrapper[4943]: I0307 14:40:52.775827 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:52 crc kubenswrapper[4943]: I0307 14:40:52.775880 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:52 crc kubenswrapper[4943]: I0307 14:40:52.775892 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:52 crc kubenswrapper[4943]: I0307 14:40:52.775915 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:52 crc kubenswrapper[4943]: I0307 14:40:52.775949 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:52Z","lastTransitionTime":"2026-03-07T14:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:52 crc kubenswrapper[4943]: I0307 14:40:52.879016 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:52 crc kubenswrapper[4943]: I0307 14:40:52.879075 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:52 crc kubenswrapper[4943]: I0307 14:40:52.879088 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:52 crc kubenswrapper[4943]: I0307 14:40:52.879110 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:52 crc kubenswrapper[4943]: I0307 14:40:52.879124 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:52Z","lastTransitionTime":"2026-03-07T14:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:52 crc kubenswrapper[4943]: I0307 14:40:52.982558 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:52 crc kubenswrapper[4943]: I0307 14:40:52.982610 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:52 crc kubenswrapper[4943]: I0307 14:40:52.982620 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:52 crc kubenswrapper[4943]: I0307 14:40:52.982638 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:52 crc kubenswrapper[4943]: I0307 14:40:52.982653 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:52Z","lastTransitionTime":"2026-03-07T14:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:53 crc kubenswrapper[4943]: I0307 14:40:53.084941 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:53 crc kubenswrapper[4943]: I0307 14:40:53.084983 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:53 crc kubenswrapper[4943]: I0307 14:40:53.084995 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:53 crc kubenswrapper[4943]: I0307 14:40:53.085023 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:53 crc kubenswrapper[4943]: I0307 14:40:53.085034 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:53Z","lastTransitionTime":"2026-03-07T14:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:53 crc kubenswrapper[4943]: I0307 14:40:53.194471 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:53 crc kubenswrapper[4943]: I0307 14:40:53.194885 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:53 crc kubenswrapper[4943]: I0307 14:40:53.194976 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:53 crc kubenswrapper[4943]: I0307 14:40:53.195054 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:53 crc kubenswrapper[4943]: I0307 14:40:53.195142 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:53Z","lastTransitionTime":"2026-03-07T14:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:53 crc kubenswrapper[4943]: I0307 14:40:53.299308 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:53 crc kubenswrapper[4943]: I0307 14:40:53.299380 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:53 crc kubenswrapper[4943]: I0307 14:40:53.299402 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:53 crc kubenswrapper[4943]: I0307 14:40:53.299431 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:53 crc kubenswrapper[4943]: I0307 14:40:53.299452 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:53Z","lastTransitionTime":"2026-03-07T14:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:53 crc kubenswrapper[4943]: I0307 14:40:53.402780 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:53 crc kubenswrapper[4943]: I0307 14:40:53.403264 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:53 crc kubenswrapper[4943]: I0307 14:40:53.403433 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:53 crc kubenswrapper[4943]: I0307 14:40:53.403580 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:53 crc kubenswrapper[4943]: I0307 14:40:53.403716 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:53Z","lastTransitionTime":"2026-03-07T14:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:53 crc kubenswrapper[4943]: I0307 14:40:53.506535 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:53 crc kubenswrapper[4943]: I0307 14:40:53.506594 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:53 crc kubenswrapper[4943]: I0307 14:40:53.506618 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:53 crc kubenswrapper[4943]: I0307 14:40:53.506678 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:53 crc kubenswrapper[4943]: I0307 14:40:53.506700 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:53Z","lastTransitionTime":"2026-03-07T14:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:53 crc kubenswrapper[4943]: I0307 14:40:53.609882 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:53 crc kubenswrapper[4943]: I0307 14:40:53.610005 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:53 crc kubenswrapper[4943]: I0307 14:40:53.610032 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:53 crc kubenswrapper[4943]: I0307 14:40:53.610064 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:53 crc kubenswrapper[4943]: I0307 14:40:53.610088 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:53Z","lastTransitionTime":"2026-03-07T14:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:53 crc kubenswrapper[4943]: I0307 14:40:53.712688 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:53 crc kubenswrapper[4943]: I0307 14:40:53.712736 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:53 crc kubenswrapper[4943]: I0307 14:40:53.712750 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:53 crc kubenswrapper[4943]: I0307 14:40:53.712769 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:53 crc kubenswrapper[4943]: I0307 14:40:53.712783 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:53Z","lastTransitionTime":"2026-03-07T14:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:53 crc kubenswrapper[4943]: I0307 14:40:53.754504 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 07 14:40:53 crc kubenswrapper[4943]: I0307 14:40:53.754504 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 07 14:40:53 crc kubenswrapper[4943]: I0307 14:40:53.754674 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 07 14:40:53 crc kubenswrapper[4943]: E0307 14:40:53.754857 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 07 14:40:53 crc kubenswrapper[4943]: E0307 14:40:53.755070 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 07 14:40:53 crc kubenswrapper[4943]: E0307 14:40:53.755132 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 07 14:40:53 crc kubenswrapper[4943]: I0307 14:40:53.815467 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:53 crc kubenswrapper[4943]: I0307 14:40:53.815509 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:53 crc kubenswrapper[4943]: I0307 14:40:53.815521 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:53 crc kubenswrapper[4943]: I0307 14:40:53.815540 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:53 crc kubenswrapper[4943]: I0307 14:40:53.815550 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:53Z","lastTransitionTime":"2026-03-07T14:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:53 crc kubenswrapper[4943]: I0307 14:40:53.918433 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:53 crc kubenswrapper[4943]: I0307 14:40:53.918485 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:53 crc kubenswrapper[4943]: I0307 14:40:53.918501 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:53 crc kubenswrapper[4943]: I0307 14:40:53.918521 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:53 crc kubenswrapper[4943]: I0307 14:40:53.918538 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:53Z","lastTransitionTime":"2026-03-07T14:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:54 crc kubenswrapper[4943]: I0307 14:40:54.022782 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:54 crc kubenswrapper[4943]: I0307 14:40:54.022846 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:54 crc kubenswrapper[4943]: I0307 14:40:54.022863 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:54 crc kubenswrapper[4943]: I0307 14:40:54.022892 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:54 crc kubenswrapper[4943]: I0307 14:40:54.022910 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:54Z","lastTransitionTime":"2026-03-07T14:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:54 crc kubenswrapper[4943]: I0307 14:40:54.125530 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:54 crc kubenswrapper[4943]: I0307 14:40:54.125565 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:54 crc kubenswrapper[4943]: I0307 14:40:54.125575 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:54 crc kubenswrapper[4943]: I0307 14:40:54.125596 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:54 crc kubenswrapper[4943]: I0307 14:40:54.125608 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:54Z","lastTransitionTime":"2026-03-07T14:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:54 crc kubenswrapper[4943]: I0307 14:40:54.227622 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:54 crc kubenswrapper[4943]: I0307 14:40:54.227675 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:54 crc kubenswrapper[4943]: I0307 14:40:54.227691 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:54 crc kubenswrapper[4943]: I0307 14:40:54.230417 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:54 crc kubenswrapper[4943]: I0307 14:40:54.230455 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:54Z","lastTransitionTime":"2026-03-07T14:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:54 crc kubenswrapper[4943]: I0307 14:40:54.333026 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:54 crc kubenswrapper[4943]: I0307 14:40:54.333072 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:54 crc kubenswrapper[4943]: I0307 14:40:54.333084 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:54 crc kubenswrapper[4943]: I0307 14:40:54.333102 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:54 crc kubenswrapper[4943]: I0307 14:40:54.333117 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:54Z","lastTransitionTime":"2026-03-07T14:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:54 crc kubenswrapper[4943]: I0307 14:40:54.439281 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:54 crc kubenswrapper[4943]: I0307 14:40:54.439378 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:54 crc kubenswrapper[4943]: I0307 14:40:54.439403 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:54 crc kubenswrapper[4943]: I0307 14:40:54.439434 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:54 crc kubenswrapper[4943]: I0307 14:40:54.439456 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:54Z","lastTransitionTime":"2026-03-07T14:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:54 crc kubenswrapper[4943]: I0307 14:40:54.541668 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:54 crc kubenswrapper[4943]: I0307 14:40:54.541708 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:54 crc kubenswrapper[4943]: I0307 14:40:54.541723 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:54 crc kubenswrapper[4943]: I0307 14:40:54.541747 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:54 crc kubenswrapper[4943]: I0307 14:40:54.541764 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:54Z","lastTransitionTime":"2026-03-07T14:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:54 crc kubenswrapper[4943]: I0307 14:40:54.644542 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:54 crc kubenswrapper[4943]: I0307 14:40:54.644824 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:54 crc kubenswrapper[4943]: I0307 14:40:54.644993 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:54 crc kubenswrapper[4943]: I0307 14:40:54.645132 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:54 crc kubenswrapper[4943]: I0307 14:40:54.645217 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:54Z","lastTransitionTime":"2026-03-07T14:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:54 crc kubenswrapper[4943]: I0307 14:40:54.758040 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:54 crc kubenswrapper[4943]: I0307 14:40:54.758098 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:54 crc kubenswrapper[4943]: I0307 14:40:54.758114 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:54 crc kubenswrapper[4943]: I0307 14:40:54.758138 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:54 crc kubenswrapper[4943]: I0307 14:40:54.758153 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:54Z","lastTransitionTime":"2026-03-07T14:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:54 crc kubenswrapper[4943]: I0307 14:40:54.860661 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:54 crc kubenswrapper[4943]: I0307 14:40:54.860714 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:54 crc kubenswrapper[4943]: I0307 14:40:54.860730 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:54 crc kubenswrapper[4943]: I0307 14:40:54.860763 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:54 crc kubenswrapper[4943]: I0307 14:40:54.860782 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:54Z","lastTransitionTime":"2026-03-07T14:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:54 crc kubenswrapper[4943]: I0307 14:40:54.963424 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:54 crc kubenswrapper[4943]: I0307 14:40:54.963469 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:54 crc kubenswrapper[4943]: I0307 14:40:54.963485 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:54 crc kubenswrapper[4943]: I0307 14:40:54.963511 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:54 crc kubenswrapper[4943]: I0307 14:40:54.963530 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:54Z","lastTransitionTime":"2026-03-07T14:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:55 crc kubenswrapper[4943]: I0307 14:40:55.068863 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:55 crc kubenswrapper[4943]: I0307 14:40:55.068917 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:55 crc kubenswrapper[4943]: I0307 14:40:55.068961 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:55 crc kubenswrapper[4943]: I0307 14:40:55.068990 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:55 crc kubenswrapper[4943]: I0307 14:40:55.069014 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:55Z","lastTransitionTime":"2026-03-07T14:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:55 crc kubenswrapper[4943]: I0307 14:40:55.171652 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:55 crc kubenswrapper[4943]: I0307 14:40:55.171716 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:55 crc kubenswrapper[4943]: I0307 14:40:55.171732 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:55 crc kubenswrapper[4943]: I0307 14:40:55.171755 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:55 crc kubenswrapper[4943]: I0307 14:40:55.171772 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:55Z","lastTransitionTime":"2026-03-07T14:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:55 crc kubenswrapper[4943]: I0307 14:40:55.275257 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:55 crc kubenswrapper[4943]: I0307 14:40:55.275351 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:55 crc kubenswrapper[4943]: I0307 14:40:55.275375 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:55 crc kubenswrapper[4943]: I0307 14:40:55.275409 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:55 crc kubenswrapper[4943]: I0307 14:40:55.275428 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:55Z","lastTransitionTime":"2026-03-07T14:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:55 crc kubenswrapper[4943]: I0307 14:40:55.379280 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:55 crc kubenswrapper[4943]: I0307 14:40:55.379874 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:55 crc kubenswrapper[4943]: I0307 14:40:55.380085 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:55 crc kubenswrapper[4943]: I0307 14:40:55.380251 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:55 crc kubenswrapper[4943]: I0307 14:40:55.380401 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:55Z","lastTransitionTime":"2026-03-07T14:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:55 crc kubenswrapper[4943]: I0307 14:40:55.484242 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:55 crc kubenswrapper[4943]: I0307 14:40:55.484711 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:55 crc kubenswrapper[4943]: I0307 14:40:55.484958 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:55 crc kubenswrapper[4943]: I0307 14:40:55.485190 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:55 crc kubenswrapper[4943]: I0307 14:40:55.485382 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:55Z","lastTransitionTime":"2026-03-07T14:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:55 crc kubenswrapper[4943]: I0307 14:40:55.588606 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:55 crc kubenswrapper[4943]: I0307 14:40:55.588663 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:55 crc kubenswrapper[4943]: I0307 14:40:55.588680 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:55 crc kubenswrapper[4943]: I0307 14:40:55.588705 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:55 crc kubenswrapper[4943]: I0307 14:40:55.588723 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:55Z","lastTransitionTime":"2026-03-07T14:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:55 crc kubenswrapper[4943]: I0307 14:40:55.691428 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:55 crc kubenswrapper[4943]: I0307 14:40:55.691488 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:55 crc kubenswrapper[4943]: I0307 14:40:55.691505 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:55 crc kubenswrapper[4943]: I0307 14:40:55.691529 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:55 crc kubenswrapper[4943]: I0307 14:40:55.691547 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:55Z","lastTransitionTime":"2026-03-07T14:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:55 crc kubenswrapper[4943]: I0307 14:40:55.754502 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 07 14:40:55 crc kubenswrapper[4943]: I0307 14:40:55.754518 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 07 14:40:55 crc kubenswrapper[4943]: E0307 14:40:55.754711 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 07 14:40:55 crc kubenswrapper[4943]: I0307 14:40:55.754537 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 07 14:40:55 crc kubenswrapper[4943]: E0307 14:40:55.754815 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 07 14:40:55 crc kubenswrapper[4943]: E0307 14:40:55.754904 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 07 14:40:55 crc kubenswrapper[4943]: I0307 14:40:55.793806 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:55 crc kubenswrapper[4943]: I0307 14:40:55.793857 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:55 crc kubenswrapper[4943]: I0307 14:40:55.793875 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:55 crc kubenswrapper[4943]: I0307 14:40:55.793901 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:55 crc kubenswrapper[4943]: I0307 14:40:55.793919 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:55Z","lastTransitionTime":"2026-03-07T14:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:55 crc kubenswrapper[4943]: I0307 14:40:55.896700 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:55 crc kubenswrapper[4943]: I0307 14:40:55.896765 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:55 crc kubenswrapper[4943]: I0307 14:40:55.896782 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:55 crc kubenswrapper[4943]: I0307 14:40:55.896808 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:55 crc kubenswrapper[4943]: I0307 14:40:55.896827 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:55Z","lastTransitionTime":"2026-03-07T14:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:56 crc kubenswrapper[4943]: I0307 14:40:55.999896 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:56 crc kubenswrapper[4943]: I0307 14:40:55.999995 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:56 crc kubenswrapper[4943]: I0307 14:40:56.000014 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:56 crc kubenswrapper[4943]: I0307 14:40:56.000045 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:56 crc kubenswrapper[4943]: I0307 14:40:56.000064 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:56Z","lastTransitionTime":"2026-03-07T14:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:56 crc kubenswrapper[4943]: I0307 14:40:56.103052 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:56 crc kubenswrapper[4943]: I0307 14:40:56.103132 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:56 crc kubenswrapper[4943]: I0307 14:40:56.103158 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:56 crc kubenswrapper[4943]: I0307 14:40:56.103198 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:56 crc kubenswrapper[4943]: I0307 14:40:56.103223 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:56Z","lastTransitionTime":"2026-03-07T14:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:56 crc kubenswrapper[4943]: I0307 14:40:56.206333 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:56 crc kubenswrapper[4943]: I0307 14:40:56.206400 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:56 crc kubenswrapper[4943]: I0307 14:40:56.206417 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:56 crc kubenswrapper[4943]: I0307 14:40:56.206443 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:56 crc kubenswrapper[4943]: I0307 14:40:56.206462 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:56Z","lastTransitionTime":"2026-03-07T14:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:56 crc kubenswrapper[4943]: I0307 14:40:56.263905 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 07 14:40:56 crc kubenswrapper[4943]: I0307 14:40:56.264047 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 07 14:40:56 crc kubenswrapper[4943]: I0307 14:40:56.264074 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 07 14:40:56 crc kubenswrapper[4943]: I0307 14:40:56.264113 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 07 14:40:56 crc kubenswrapper[4943]: I0307 14:40:56.264139 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T14:40:56Z","lastTransitionTime":"2026-03-07T14:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 07 14:40:56 crc kubenswrapper[4943]: I0307 14:40:56.737557 4943 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Mar 07 14:40:56 crc kubenswrapper[4943]: I0307 14:40:56.747214 4943 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Mar 07 14:40:56 crc kubenswrapper[4943]: I0307 14:40:56.772162 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 07 14:40:56 crc kubenswrapper[4943]: I0307 14:40:56.772594 4943 scope.go:117] "RemoveContainer" containerID="ed8d27f5a369a691493bda9dbd6e03c098931b8dab637bbd9bacac6fe7979c80" Mar 07 14:40:56 crc kubenswrapper[4943]: E0307 14:40:56.773012 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 07 14:40:57 crc kubenswrapper[4943]: I0307 14:40:57.218551 4943 scope.go:117] "RemoveContainer" containerID="ed8d27f5a369a691493bda9dbd6e03c098931b8dab637bbd9bacac6fe7979c80" Mar 07 14:40:57 crc kubenswrapper[4943]: E0307 14:40:57.218802 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 07 14:40:57 crc kubenswrapper[4943]: I0307 14:40:57.754720 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 07 14:40:57 crc kubenswrapper[4943]: I0307 14:40:57.754720 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 07 14:40:57 crc kubenswrapper[4943]: E0307 14:40:57.754864 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 07 14:40:57 crc kubenswrapper[4943]: I0307 14:40:57.754998 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 07 14:40:57 crc kubenswrapper[4943]: E0307 14:40:57.755157 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 07 14:40:57 crc kubenswrapper[4943]: E0307 14:40:57.755281 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 07 14:40:59 crc kubenswrapper[4943]: I0307 14:40:59.499728 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:40:59 crc kubenswrapper[4943]: I0307 14:40:59.499857 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 07 14:40:59 crc kubenswrapper[4943]: I0307 14:40:59.499966 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 07 14:40:59 crc kubenswrapper[4943]: E0307 14:40:59.500089 4943 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 07 14:40:59 crc kubenswrapper[4943]: E0307 14:40:59.500118 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:15.500070831 +0000 UTC m=+117.452207359 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:40:59 crc kubenswrapper[4943]: E0307 14:40:59.500172 4943 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 07 14:40:59 crc kubenswrapper[4943]: E0307 14:40:59.500179 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-07 14:41:15.500155593 +0000 UTC m=+117.452292121 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 07 14:40:59 crc kubenswrapper[4943]: E0307 14:40:59.500298 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-07 14:41:15.500272636 +0000 UTC m=+117.452409204 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 07 14:40:59 crc kubenswrapper[4943]: I0307 14:40:59.601204 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 07 14:40:59 crc kubenswrapper[4943]: I0307 14:40:59.601489 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 07 14:40:59 crc kubenswrapper[4943]: E0307 14:40:59.601848 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 07 14:40:59 crc kubenswrapper[4943]: E0307 14:40:59.601897 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 07 14:40:59 crc kubenswrapper[4943]: E0307 14:40:59.601922 4943 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 07 14:40:59 crc kubenswrapper[4943]: E0307 14:40:59.602055 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-07 14:41:15.602025376 +0000 UTC m=+117.554161904 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 07 14:40:59 crc kubenswrapper[4943]: E0307 14:40:59.602061 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 07 14:40:59 crc kubenswrapper[4943]: E0307 14:40:59.602109 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 07 14:40:59 crc kubenswrapper[4943]: E0307 14:40:59.602131 4943 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 07 14:40:59 crc kubenswrapper[4943]: E0307 14:40:59.602292 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-07 14:41:15.602265772 +0000 UTC m=+117.554402310 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 07 14:40:59 crc kubenswrapper[4943]: I0307 14:40:59.754724 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 07 14:40:59 crc kubenswrapper[4943]: I0307 14:40:59.754767 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 07 14:40:59 crc kubenswrapper[4943]: I0307 14:40:59.754792 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 07 14:40:59 crc kubenswrapper[4943]: E0307 14:40:59.754892 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 07 14:40:59 crc kubenswrapper[4943]: E0307 14:40:59.755039 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 07 14:40:59 crc kubenswrapper[4943]: E0307 14:40:59.755167 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 07 14:41:00 crc kubenswrapper[4943]: I0307 14:41:00.397842 4943 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Mar 07 14:41:01 crc kubenswrapper[4943]: I0307 14:41:01.755177 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 07 14:41:01 crc kubenswrapper[4943]: I0307 14:41:01.755258 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 07 14:41:01 crc kubenswrapper[4943]: I0307 14:41:01.755269 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 07 14:41:01 crc kubenswrapper[4943]: E0307 14:41:01.755365 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 07 14:41:01 crc kubenswrapper[4943]: E0307 14:41:01.755612 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 07 14:41:01 crc kubenswrapper[4943]: E0307 14:41:01.755886 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 07 14:41:02 crc kubenswrapper[4943]: I0307 14:41:02.785345 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Mar 07 14:41:03 crc kubenswrapper[4943]: I0307 14:41:03.755223 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 07 14:41:03 crc kubenswrapper[4943]: I0307 14:41:03.755327 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 07 14:41:03 crc kubenswrapper[4943]: I0307 14:41:03.755422 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 07 14:41:03 crc kubenswrapper[4943]: E0307 14:41:03.755413 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 07 14:41:03 crc kubenswrapper[4943]: E0307 14:41:03.755555 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 07 14:41:03 crc kubenswrapper[4943]: E0307 14:41:03.755625 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.721417 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-cpdmm"] Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.721849 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.726410 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.727507 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-qrlpk"] Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.727770 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-qrlpk" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.727965 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.728098 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.728293 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.728306 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.729275 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-npf7n"] Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.729752 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-npf7n" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.731048 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-4j27g"] Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.738368 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.739099 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.743289 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.743324 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.744718 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.745020 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.745104 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.745371 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-4j27g" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.746579 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.748301 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.748358 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.755217 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.755377 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 07 14:41:05 crc kubenswrapper[4943]: E0307 14:41:05.755549 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.756790 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 07 14:41:05 crc kubenswrapper[4943]: E0307 14:41:05.759545 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 07 14:41:05 crc kubenswrapper[4943]: E0307 14:41:05.761380 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.761857 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a13f1d2e-9974-4c40-85f5-d876d5644993-os-release\") pod \"multus-qrlpk\" (UID: \"a13f1d2e-9974-4c40-85f5-d876d5644993\") " pod="openshift-multus/multus-qrlpk" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.761919 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-s8cw7"] Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.761978 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a13f1d2e-9974-4c40-85f5-d876d5644993-cni-binary-copy\") pod \"multus-qrlpk\" (UID: \"a13f1d2e-9974-4c40-85f5-d876d5644993\") " pod="openshift-multus/multus-qrlpk" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.762030 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/65707314-fccf-4ae5-8790-eb24f63ad3ad-cnibin\") pod \"multus-additional-cni-plugins-4j27g\" (UID: \"65707314-fccf-4ae5-8790-eb24f63ad3ad\") " pod="openshift-multus/multus-additional-cni-plugins-4j27g" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.762193 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a13f1d2e-9974-4c40-85f5-d876d5644993-multus-cni-dir\") pod \"multus-qrlpk\" (UID: \"a13f1d2e-9974-4c40-85f5-d876d5644993\") " pod="openshift-multus/multus-qrlpk" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.762271 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/a13f1d2e-9974-4c40-85f5-d876d5644993-multus-socket-dir-parent\") pod \"multus-qrlpk\" (UID: \"a13f1d2e-9974-4c40-85f5-d876d5644993\") " pod="openshift-multus/multus-qrlpk" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.762353 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/a13f1d2e-9974-4c40-85f5-d876d5644993-host-var-lib-kubelet\") pod \"multus-qrlpk\" (UID: \"a13f1d2e-9974-4c40-85f5-d876d5644993\") " pod="openshift-multus/multus-qrlpk" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.762387 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/a13f1d2e-9974-4c40-85f5-d876d5644993-host-var-lib-cni-multus\") pod \"multus-qrlpk\" (UID: \"a13f1d2e-9974-4c40-85f5-d876d5644993\") " pod="openshift-multus/multus-qrlpk" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.762458 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/65707314-fccf-4ae5-8790-eb24f63ad3ad-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-4j27g\" (UID: \"65707314-fccf-4ae5-8790-eb24f63ad3ad\") " pod="openshift-multus/multus-additional-cni-plugins-4j27g" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.762530 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a13f1d2e-9974-4c40-85f5-d876d5644993-multus-conf-dir\") pod \"multus-qrlpk\" (UID: \"a13f1d2e-9974-4c40-85f5-d876d5644993\") " pod="openshift-multus/multus-qrlpk" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.762608 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/a13f1d2e-9974-4c40-85f5-d876d5644993-multus-daemon-config\") pod \"multus-qrlpk\" (UID: \"a13f1d2e-9974-4c40-85f5-d876d5644993\") " pod="openshift-multus/multus-qrlpk" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.762640 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a13f1d2e-9974-4c40-85f5-d876d5644993-system-cni-dir\") pod \"multus-qrlpk\" (UID: \"a13f1d2e-9974-4c40-85f5-d876d5644993\") " pod="openshift-multus/multus-qrlpk" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.762756 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/a13f1d2e-9974-4c40-85f5-d876d5644993-hostroot\") pod \"multus-qrlpk\" (UID: \"a13f1d2e-9974-4c40-85f5-d876d5644993\") " pod="openshift-multus/multus-qrlpk" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.762801 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a13f1d2e-9974-4c40-85f5-d876d5644993-cnibin\") pod \"multus-qrlpk\" (UID: \"a13f1d2e-9974-4c40-85f5-d876d5644993\") " pod="openshift-multus/multus-qrlpk" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.762833 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a13f1d2e-9974-4c40-85f5-d876d5644993-host-run-netns\") pod \"multus-qrlpk\" (UID: \"a13f1d2e-9974-4c40-85f5-d876d5644993\") " pod="openshift-multus/multus-qrlpk" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.762863 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/65707314-fccf-4ae5-8790-eb24f63ad3ad-os-release\") pod \"multus-additional-cni-plugins-4j27g\" (UID: \"65707314-fccf-4ae5-8790-eb24f63ad3ad\") " pod="openshift-multus/multus-additional-cni-plugins-4j27g" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.762893 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/65707314-fccf-4ae5-8790-eb24f63ad3ad-cni-binary-copy\") pod \"multus-additional-cni-plugins-4j27g\" (UID: \"65707314-fccf-4ae5-8790-eb24f63ad3ad\") " pod="openshift-multus/multus-additional-cni-plugins-4j27g" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.762951 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a13f1d2e-9974-4c40-85f5-d876d5644993-host-var-lib-cni-bin\") pod \"multus-qrlpk\" (UID: \"a13f1d2e-9974-4c40-85f5-d876d5644993\") " pod="openshift-multus/multus-qrlpk" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.762985 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/a13f1d2e-9974-4c40-85f5-d876d5644993-host-run-k8s-cni-cncf-io\") pod \"multus-qrlpk\" (UID: \"a13f1d2e-9974-4c40-85f5-d876d5644993\") " pod="openshift-multus/multus-qrlpk" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.763049 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jfd6g\" (UniqueName: \"kubernetes.io/projected/4c1f18ea-e557-41fc-95dd-ce4e1722a61e-kube-api-access-jfd6g\") pod \"machine-config-daemon-cpdmm\" (UID: \"4c1f18ea-e557-41fc-95dd-ce4e1722a61e\") " pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.763097 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rn69r\" (UniqueName: \"kubernetes.io/projected/27cc8510-d4f1-476e-8d35-d7c64cec5e99-kube-api-access-rn69r\") pod \"node-resolver-npf7n\" (UID: \"27cc8510-d4f1-476e-8d35-d7c64cec5e99\") " pod="openshift-dns/node-resolver-npf7n" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.763144 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.763199 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/65707314-fccf-4ae5-8790-eb24f63ad3ad-system-cni-dir\") pod \"multus-additional-cni-plugins-4j27g\" (UID: \"65707314-fccf-4ae5-8790-eb24f63ad3ad\") " pod="openshift-multus/multus-additional-cni-plugins-4j27g" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.763408 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4c1f18ea-e557-41fc-95dd-ce4e1722a61e-mcd-auth-proxy-config\") pod \"machine-config-daemon-cpdmm\" (UID: \"4c1f18ea-e557-41fc-95dd-ce4e1722a61e\") " pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.763456 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/65707314-fccf-4ae5-8790-eb24f63ad3ad-tuning-conf-dir\") pod \"multus-additional-cni-plugins-4j27g\" (UID: \"65707314-fccf-4ae5-8790-eb24f63ad3ad\") " pod="openshift-multus/multus-additional-cni-plugins-4j27g" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.763522 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4c1f18ea-e557-41fc-95dd-ce4e1722a61e-proxy-tls\") pod \"machine-config-daemon-cpdmm\" (UID: \"4c1f18ea-e557-41fc-95dd-ce4e1722a61e\") " pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.763558 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6p27\" (UniqueName: \"kubernetes.io/projected/a13f1d2e-9974-4c40-85f5-d876d5644993-kube-api-access-r6p27\") pod \"multus-qrlpk\" (UID: \"a13f1d2e-9974-4c40-85f5-d876d5644993\") " pod="openshift-multus/multus-qrlpk" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.763603 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a13f1d2e-9974-4c40-85f5-d876d5644993-etc-kubernetes\") pod \"multus-qrlpk\" (UID: \"a13f1d2e-9974-4c40-85f5-d876d5644993\") " pod="openshift-multus/multus-qrlpk" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.763634 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/27cc8510-d4f1-476e-8d35-d7c64cec5e99-hosts-file\") pod \"node-resolver-npf7n\" (UID: \"27cc8510-d4f1-476e-8d35-d7c64cec5e99\") " pod="openshift-dns/node-resolver-npf7n" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.763667 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/a13f1d2e-9974-4c40-85f5-d876d5644993-host-run-multus-certs\") pod \"multus-qrlpk\" (UID: \"a13f1d2e-9974-4c40-85f5-d876d5644993\") " pod="openshift-multus/multus-qrlpk" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.763698 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krtgv\" (UniqueName: \"kubernetes.io/projected/65707314-fccf-4ae5-8790-eb24f63ad3ad-kube-api-access-krtgv\") pod \"multus-additional-cni-plugins-4j27g\" (UID: \"65707314-fccf-4ae5-8790-eb24f63ad3ad\") " pod="openshift-multus/multus-additional-cni-plugins-4j27g" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.763738 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/4c1f18ea-e557-41fc-95dd-ce4e1722a61e-rootfs\") pod \"machine-config-daemon-cpdmm\" (UID: \"4c1f18ea-e557-41fc-95dd-ce4e1722a61e\") " pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.770235 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.770300 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.770379 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.770431 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.770881 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.771191 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.771439 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.793964 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=3.793897323 podStartE2EDuration="3.793897323s" podCreationTimestamp="2026-03-07 14:41:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:05.79034205 +0000 UTC m=+107.742478598" watchObservedRunningTime="2026-03-07 14:41:05.793897323 +0000 UTC m=+107.746033831" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.818742 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-8fg7f"] Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.819312 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-8fg7f" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.821056 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.821882 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.822145 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.823189 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.867010 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6p27\" (UniqueName: \"kubernetes.io/projected/a13f1d2e-9974-4c40-85f5-d876d5644993-kube-api-access-r6p27\") pod \"multus-qrlpk\" (UID: \"a13f1d2e-9974-4c40-85f5-d876d5644993\") " pod="openshift-multus/multus-qrlpk" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.867075 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d340687c-faf0-48dc-a1ac-dec5067ac665-ovnkube-script-lib\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.867114 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4c1f18ea-e557-41fc-95dd-ce4e1722a61e-proxy-tls\") pod \"machine-config-daemon-cpdmm\" (UID: \"4c1f18ea-e557-41fc-95dd-ce4e1722a61e\") " pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.867143 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/27cc8510-d4f1-476e-8d35-d7c64cec5e99-hosts-file\") pod \"node-resolver-npf7n\" (UID: \"27cc8510-d4f1-476e-8d35-d7c64cec5e99\") " pod="openshift-dns/node-resolver-npf7n" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.867168 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-var-lib-openvswitch\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.867296 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d340687c-faf0-48dc-a1ac-dec5067ac665-env-overrides\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.867362 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/27cc8510-d4f1-476e-8d35-d7c64cec5e99-hosts-file\") pod \"node-resolver-npf7n\" (UID: \"27cc8510-d4f1-476e-8d35-d7c64cec5e99\") " pod="openshift-dns/node-resolver-npf7n" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.867533 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/210b9f6a-802b-442b-a940-66c003e96c41-host\") pod \"node-ca-8fg7f\" (UID: \"210b9f6a-802b-442b-a940-66c003e96c41\") " pod="openshift-image-registry/node-ca-8fg7f" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.867592 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/a13f1d2e-9974-4c40-85f5-d876d5644993-host-run-multus-certs\") pod \"multus-qrlpk\" (UID: \"a13f1d2e-9974-4c40-85f5-d876d5644993\") " pod="openshift-multus/multus-qrlpk" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.867633 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a13f1d2e-9974-4c40-85f5-d876d5644993-etc-kubernetes\") pod \"multus-qrlpk\" (UID: \"a13f1d2e-9974-4c40-85f5-d876d5644993\") " pod="openshift-multus/multus-qrlpk" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.867667 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/a13f1d2e-9974-4c40-85f5-d876d5644993-host-run-multus-certs\") pod \"multus-qrlpk\" (UID: \"a13f1d2e-9974-4c40-85f5-d876d5644993\") " pod="openshift-multus/multus-qrlpk" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.867680 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/4c1f18ea-e557-41fc-95dd-ce4e1722a61e-rootfs\") pod \"machine-config-daemon-cpdmm\" (UID: \"4c1f18ea-e557-41fc-95dd-ce4e1722a61e\") " pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.867727 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/4c1f18ea-e557-41fc-95dd-ce4e1722a61e-rootfs\") pod \"machine-config-daemon-cpdmm\" (UID: \"4c1f18ea-e557-41fc-95dd-ce4e1722a61e\") " pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.867747 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a13f1d2e-9974-4c40-85f5-d876d5644993-etc-kubernetes\") pod \"multus-qrlpk\" (UID: \"a13f1d2e-9974-4c40-85f5-d876d5644993\") " pod="openshift-multus/multus-qrlpk" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.867786 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krtgv\" (UniqueName: \"kubernetes.io/projected/65707314-fccf-4ae5-8790-eb24f63ad3ad-kube-api-access-krtgv\") pod \"multus-additional-cni-plugins-4j27g\" (UID: \"65707314-fccf-4ae5-8790-eb24f63ad3ad\") " pod="openshift-multus/multus-additional-cni-plugins-4j27g" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.867845 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-run-systemd\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.867885 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d340687c-faf0-48dc-a1ac-dec5067ac665-ovnkube-config\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.867976 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7d2m\" (UniqueName: \"kubernetes.io/projected/d340687c-faf0-48dc-a1ac-dec5067ac665-kube-api-access-p7d2m\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.868029 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a13f1d2e-9974-4c40-85f5-d876d5644993-os-release\") pod \"multus-qrlpk\" (UID: \"a13f1d2e-9974-4c40-85f5-d876d5644993\") " pod="openshift-multus/multus-qrlpk" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.868071 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/65707314-fccf-4ae5-8790-eb24f63ad3ad-cnibin\") pod \"multus-additional-cni-plugins-4j27g\" (UID: \"65707314-fccf-4ae5-8790-eb24f63ad3ad\") " pod="openshift-multus/multus-additional-cni-plugins-4j27g" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.868110 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-host-cni-bin\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.868138 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a13f1d2e-9974-4c40-85f5-d876d5644993-os-release\") pod \"multus-qrlpk\" (UID: \"a13f1d2e-9974-4c40-85f5-d876d5644993\") " pod="openshift-multus/multus-qrlpk" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.868149 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a13f1d2e-9974-4c40-85f5-d876d5644993-multus-cni-dir\") pod \"multus-qrlpk\" (UID: \"a13f1d2e-9974-4c40-85f5-d876d5644993\") " pod="openshift-multus/multus-qrlpk" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.868169 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/65707314-fccf-4ae5-8790-eb24f63ad3ad-cnibin\") pod \"multus-additional-cni-plugins-4j27g\" (UID: \"65707314-fccf-4ae5-8790-eb24f63ad3ad\") " pod="openshift-multus/multus-additional-cni-plugins-4j27g" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.868191 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a13f1d2e-9974-4c40-85f5-d876d5644993-cni-binary-copy\") pod \"multus-qrlpk\" (UID: \"a13f1d2e-9974-4c40-85f5-d876d5644993\") " pod="openshift-multus/multus-qrlpk" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.868272 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a13f1d2e-9974-4c40-85f5-d876d5644993-multus-cni-dir\") pod \"multus-qrlpk\" (UID: \"a13f1d2e-9974-4c40-85f5-d876d5644993\") " pod="openshift-multus/multus-qrlpk" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.868286 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/a13f1d2e-9974-4c40-85f5-d876d5644993-host-var-lib-kubelet\") pod \"multus-qrlpk\" (UID: \"a13f1d2e-9974-4c40-85f5-d876d5644993\") " pod="openshift-multus/multus-qrlpk" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.868326 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-etc-openvswitch\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.868360 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/a13f1d2e-9974-4c40-85f5-d876d5644993-multus-socket-dir-parent\") pod \"multus-qrlpk\" (UID: \"a13f1d2e-9974-4c40-85f5-d876d5644993\") " pod="openshift-multus/multus-qrlpk" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.868390 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/a13f1d2e-9974-4c40-85f5-d876d5644993-host-var-lib-cni-multus\") pod \"multus-qrlpk\" (UID: \"a13f1d2e-9974-4c40-85f5-d876d5644993\") " pod="openshift-multus/multus-qrlpk" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.868426 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/a13f1d2e-9974-4c40-85f5-d876d5644993-host-var-lib-kubelet\") pod \"multus-qrlpk\" (UID: \"a13f1d2e-9974-4c40-85f5-d876d5644993\") " pod="openshift-multus/multus-qrlpk" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.868487 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/65707314-fccf-4ae5-8790-eb24f63ad3ad-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-4j27g\" (UID: \"65707314-fccf-4ae5-8790-eb24f63ad3ad\") " pod="openshift-multus/multus-additional-cni-plugins-4j27g" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.868512 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/a13f1d2e-9974-4c40-85f5-d876d5644993-host-var-lib-cni-multus\") pod \"multus-qrlpk\" (UID: \"a13f1d2e-9974-4c40-85f5-d876d5644993\") " pod="openshift-multus/multus-qrlpk" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.868579 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/a13f1d2e-9974-4c40-85f5-d876d5644993-multus-socket-dir-parent\") pod \"multus-qrlpk\" (UID: \"a13f1d2e-9974-4c40-85f5-d876d5644993\") " pod="openshift-multus/multus-qrlpk" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.868649 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-host-run-netns\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.868748 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a13f1d2e-9974-4c40-85f5-d876d5644993-multus-conf-dir\") pod \"multus-qrlpk\" (UID: \"a13f1d2e-9974-4c40-85f5-d876d5644993\") " pod="openshift-multus/multus-qrlpk" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.868796 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a13f1d2e-9974-4c40-85f5-d876d5644993-multus-conf-dir\") pod \"multus-qrlpk\" (UID: \"a13f1d2e-9974-4c40-85f5-d876d5644993\") " pod="openshift-multus/multus-qrlpk" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.868828 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/a13f1d2e-9974-4c40-85f5-d876d5644993-multus-daemon-config\") pod \"multus-qrlpk\" (UID: \"a13f1d2e-9974-4c40-85f5-d876d5644993\") " pod="openshift-multus/multus-qrlpk" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.868860 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-host-slash\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.868889 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.868947 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a13f1d2e-9974-4c40-85f5-d876d5644993-system-cni-dir\") pod \"multus-qrlpk\" (UID: \"a13f1d2e-9974-4c40-85f5-d876d5644993\") " pod="openshift-multus/multus-qrlpk" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.868981 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/a13f1d2e-9974-4c40-85f5-d876d5644993-hostroot\") pod \"multus-qrlpk\" (UID: \"a13f1d2e-9974-4c40-85f5-d876d5644993\") " pod="openshift-multus/multus-qrlpk" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.869009 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-host-kubelet\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.869035 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a13f1d2e-9974-4c40-85f5-d876d5644993-cnibin\") pod \"multus-qrlpk\" (UID: \"a13f1d2e-9974-4c40-85f5-d876d5644993\") " pod="openshift-multus/multus-qrlpk" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.869070 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/65707314-fccf-4ae5-8790-eb24f63ad3ad-os-release\") pod \"multus-additional-cni-plugins-4j27g\" (UID: \"65707314-fccf-4ae5-8790-eb24f63ad3ad\") " pod="openshift-multus/multus-additional-cni-plugins-4j27g" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.869092 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-run-openvswitch\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.869119 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-log-socket\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.869149 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/a13f1d2e-9974-4c40-85f5-d876d5644993-hostroot\") pod \"multus-qrlpk\" (UID: \"a13f1d2e-9974-4c40-85f5-d876d5644993\") " pod="openshift-multus/multus-qrlpk" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.869150 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-host-run-ovn-kubernetes\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.869202 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-host-cni-netd\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.869211 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a13f1d2e-9974-4c40-85f5-d876d5644993-cnibin\") pod \"multus-qrlpk\" (UID: \"a13f1d2e-9974-4c40-85f5-d876d5644993\") " pod="openshift-multus/multus-qrlpk" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.869235 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a13f1d2e-9974-4c40-85f5-d876d5644993-host-run-netns\") pod \"multus-qrlpk\" (UID: \"a13f1d2e-9974-4c40-85f5-d876d5644993\") " pod="openshift-multus/multus-qrlpk" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.869257 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a13f1d2e-9974-4c40-85f5-d876d5644993-host-var-lib-cni-bin\") pod \"multus-qrlpk\" (UID: \"a13f1d2e-9974-4c40-85f5-d876d5644993\") " pod="openshift-multus/multus-qrlpk" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.869263 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/65707314-fccf-4ae5-8790-eb24f63ad3ad-os-release\") pod \"multus-additional-cni-plugins-4j27g\" (UID: \"65707314-fccf-4ae5-8790-eb24f63ad3ad\") " pod="openshift-multus/multus-additional-cni-plugins-4j27g" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.869279 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/65707314-fccf-4ae5-8790-eb24f63ad3ad-cni-binary-copy\") pod \"multus-additional-cni-plugins-4j27g\" (UID: \"65707314-fccf-4ae5-8790-eb24f63ad3ad\") " pod="openshift-multus/multus-additional-cni-plugins-4j27g" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.869306 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/a13f1d2e-9974-4c40-85f5-d876d5644993-host-run-k8s-cni-cncf-io\") pod \"multus-qrlpk\" (UID: \"a13f1d2e-9974-4c40-85f5-d876d5644993\") " pod="openshift-multus/multus-qrlpk" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.869339 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jfd6g\" (UniqueName: \"kubernetes.io/projected/4c1f18ea-e557-41fc-95dd-ce4e1722a61e-kube-api-access-jfd6g\") pod \"machine-config-daemon-cpdmm\" (UID: \"4c1f18ea-e557-41fc-95dd-ce4e1722a61e\") " pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.869362 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rn69r\" (UniqueName: \"kubernetes.io/projected/27cc8510-d4f1-476e-8d35-d7c64cec5e99-kube-api-access-rn69r\") pod \"node-resolver-npf7n\" (UID: \"27cc8510-d4f1-476e-8d35-d7c64cec5e99\") " pod="openshift-dns/node-resolver-npf7n" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.869381 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a13f1d2e-9974-4c40-85f5-d876d5644993-system-cni-dir\") pod \"multus-qrlpk\" (UID: \"a13f1d2e-9974-4c40-85f5-d876d5644993\") " pod="openshift-multus/multus-qrlpk" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.869412 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/65707314-fccf-4ae5-8790-eb24f63ad3ad-system-cni-dir\") pod \"multus-additional-cni-plugins-4j27g\" (UID: \"65707314-fccf-4ae5-8790-eb24f63ad3ad\") " pod="openshift-multus/multus-additional-cni-plugins-4j27g" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.869382 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/65707314-fccf-4ae5-8790-eb24f63ad3ad-system-cni-dir\") pod \"multus-additional-cni-plugins-4j27g\" (UID: \"65707314-fccf-4ae5-8790-eb24f63ad3ad\") " pod="openshift-multus/multus-additional-cni-plugins-4j27g" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.869459 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-systemd-units\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.869486 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-run-ovn\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.869532 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-node-log\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.869560 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/210b9f6a-802b-442b-a940-66c003e96c41-serviceca\") pod \"node-ca-8fg7f\" (UID: \"210b9f6a-802b-442b-a940-66c003e96c41\") " pod="openshift-image-registry/node-ca-8fg7f" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.869596 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4496\" (UniqueName: \"kubernetes.io/projected/210b9f6a-802b-442b-a940-66c003e96c41-kube-api-access-z4496\") pod \"node-ca-8fg7f\" (UID: \"210b9f6a-802b-442b-a940-66c003e96c41\") " pod="openshift-image-registry/node-ca-8fg7f" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.869618 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/65707314-fccf-4ae5-8790-eb24f63ad3ad-tuning-conf-dir\") pod \"multus-additional-cni-plugins-4j27g\" (UID: \"65707314-fccf-4ae5-8790-eb24f63ad3ad\") " pod="openshift-multus/multus-additional-cni-plugins-4j27g" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.869648 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d340687c-faf0-48dc-a1ac-dec5067ac665-ovn-node-metrics-cert\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.869699 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4c1f18ea-e557-41fc-95dd-ce4e1722a61e-mcd-auth-proxy-config\") pod \"machine-config-daemon-cpdmm\" (UID: \"4c1f18ea-e557-41fc-95dd-ce4e1722a61e\") " pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.869780 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a13f1d2e-9974-4c40-85f5-d876d5644993-host-run-netns\") pod \"multus-qrlpk\" (UID: \"a13f1d2e-9974-4c40-85f5-d876d5644993\") " pod="openshift-multus/multus-qrlpk" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.869810 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a13f1d2e-9974-4c40-85f5-d876d5644993-host-var-lib-cni-bin\") pod \"multus-qrlpk\" (UID: \"a13f1d2e-9974-4c40-85f5-d876d5644993\") " pod="openshift-multus/multus-qrlpk" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.869914 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a13f1d2e-9974-4c40-85f5-d876d5644993-cni-binary-copy\") pod \"multus-qrlpk\" (UID: \"a13f1d2e-9974-4c40-85f5-d876d5644993\") " pod="openshift-multus/multus-qrlpk" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.869913 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/65707314-fccf-4ae5-8790-eb24f63ad3ad-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-4j27g\" (UID: \"65707314-fccf-4ae5-8790-eb24f63ad3ad\") " pod="openshift-multus/multus-additional-cni-plugins-4j27g" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.870047 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/a13f1d2e-9974-4c40-85f5-d876d5644993-multus-daemon-config\") pod \"multus-qrlpk\" (UID: \"a13f1d2e-9974-4c40-85f5-d876d5644993\") " pod="openshift-multus/multus-qrlpk" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.870084 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/a13f1d2e-9974-4c40-85f5-d876d5644993-host-run-k8s-cni-cncf-io\") pod \"multus-qrlpk\" (UID: \"a13f1d2e-9974-4c40-85f5-d876d5644993\") " pod="openshift-multus/multus-qrlpk" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.870265 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/65707314-fccf-4ae5-8790-eb24f63ad3ad-cni-binary-copy\") pod \"multus-additional-cni-plugins-4j27g\" (UID: \"65707314-fccf-4ae5-8790-eb24f63ad3ad\") " pod="openshift-multus/multus-additional-cni-plugins-4j27g" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.870674 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/65707314-fccf-4ae5-8790-eb24f63ad3ad-tuning-conf-dir\") pod \"multus-additional-cni-plugins-4j27g\" (UID: \"65707314-fccf-4ae5-8790-eb24f63ad3ad\") " pod="openshift-multus/multus-additional-cni-plugins-4j27g" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.871079 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4c1f18ea-e557-41fc-95dd-ce4e1722a61e-mcd-auth-proxy-config\") pod \"machine-config-daemon-cpdmm\" (UID: \"4c1f18ea-e557-41fc-95dd-ce4e1722a61e\") " pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.882415 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4c1f18ea-e557-41fc-95dd-ce4e1722a61e-proxy-tls\") pod \"machine-config-daemon-cpdmm\" (UID: \"4c1f18ea-e557-41fc-95dd-ce4e1722a61e\") " pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.891182 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctnkz"] Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.891656 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctnkz" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.896836 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.897034 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jfd6g\" (UniqueName: \"kubernetes.io/projected/4c1f18ea-e557-41fc-95dd-ce4e1722a61e-kube-api-access-jfd6g\") pod \"machine-config-daemon-cpdmm\" (UID: \"4c1f18ea-e557-41fc-95dd-ce4e1722a61e\") " pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.897184 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.897310 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.898852 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rn69r\" (UniqueName: \"kubernetes.io/projected/27cc8510-d4f1-476e-8d35-d7c64cec5e99-kube-api-access-rn69r\") pod \"node-resolver-npf7n\" (UID: \"27cc8510-d4f1-476e-8d35-d7c64cec5e99\") " pod="openshift-dns/node-resolver-npf7n" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.900177 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.905830 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krtgv\" (UniqueName: \"kubernetes.io/projected/65707314-fccf-4ae5-8790-eb24f63ad3ad-kube-api-access-krtgv\") pod \"multus-additional-cni-plugins-4j27g\" (UID: \"65707314-fccf-4ae5-8790-eb24f63ad3ad\") " pod="openshift-multus/multus-additional-cni-plugins-4j27g" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.908682 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6p27\" (UniqueName: \"kubernetes.io/projected/a13f1d2e-9974-4c40-85f5-d876d5644993-kube-api-access-r6p27\") pod \"multus-qrlpk\" (UID: \"a13f1d2e-9974-4c40-85f5-d876d5644993\") " pod="openshift-multus/multus-qrlpk" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.970626 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-host-run-netns\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.970672 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-host-slash\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.970697 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.970726 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/9b3f2813-5e5c-4e52-bdcf-95b1b2939c1a-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-ctnkz\" (UID: \"9b3f2813-5e5c-4e52-bdcf-95b1b2939c1a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctnkz" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.970747 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-host-kubelet\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.970766 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b3f2813-5e5c-4e52-bdcf-95b1b2939c1a-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-ctnkz\" (UID: \"9b3f2813-5e5c-4e52-bdcf-95b1b2939c1a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctnkz" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.970797 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-run-openvswitch\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.970817 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-log-socket\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.970836 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-host-run-ovn-kubernetes\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.970854 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-host-cni-netd\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.970873 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/9b3f2813-5e5c-4e52-bdcf-95b1b2939c1a-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-ctnkz\" (UID: \"9b3f2813-5e5c-4e52-bdcf-95b1b2939c1a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctnkz" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.970900 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-systemd-units\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.970921 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-run-ovn\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.970962 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-node-log\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.970982 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/210b9f6a-802b-442b-a940-66c003e96c41-serviceca\") pod \"node-ca-8fg7f\" (UID: \"210b9f6a-802b-442b-a940-66c003e96c41\") " pod="openshift-image-registry/node-ca-8fg7f" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.971005 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4496\" (UniqueName: \"kubernetes.io/projected/210b9f6a-802b-442b-a940-66c003e96c41-kube-api-access-z4496\") pod \"node-ca-8fg7f\" (UID: \"210b9f6a-802b-442b-a940-66c003e96c41\") " pod="openshift-image-registry/node-ca-8fg7f" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.971026 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d340687c-faf0-48dc-a1ac-dec5067ac665-ovn-node-metrics-cert\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.971059 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d340687c-faf0-48dc-a1ac-dec5067ac665-ovnkube-script-lib\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.971089 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-var-lib-openvswitch\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.971107 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d340687c-faf0-48dc-a1ac-dec5067ac665-env-overrides\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.971127 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/210b9f6a-802b-442b-a940-66c003e96c41-host\") pod \"node-ca-8fg7f\" (UID: \"210b9f6a-802b-442b-a940-66c003e96c41\") " pod="openshift-image-registry/node-ca-8fg7f" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.971160 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-run-systemd\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.971179 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d340687c-faf0-48dc-a1ac-dec5067ac665-ovnkube-config\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.971199 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7d2m\" (UniqueName: \"kubernetes.io/projected/d340687c-faf0-48dc-a1ac-dec5067ac665-kube-api-access-p7d2m\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.971225 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9b3f2813-5e5c-4e52-bdcf-95b1b2939c1a-service-ca\") pod \"cluster-version-operator-5c965bbfc6-ctnkz\" (UID: \"9b3f2813-5e5c-4e52-bdcf-95b1b2939c1a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctnkz" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.971248 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-host-cni-bin\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.971270 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-etc-openvswitch\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.971292 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9b3f2813-5e5c-4e52-bdcf-95b1b2939c1a-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-ctnkz\" (UID: \"9b3f2813-5e5c-4e52-bdcf-95b1b2939c1a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctnkz" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.971386 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-host-run-netns\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.971425 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-host-slash\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.971456 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.971496 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-host-kubelet\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.971536 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-run-openvswitch\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.971568 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-log-socket\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.971597 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-host-run-ovn-kubernetes\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.971634 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-host-cni-netd\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.971672 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-systemd-units\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.971698 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-run-ovn\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.971727 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-node-log\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.972772 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/210b9f6a-802b-442b-a940-66c003e96c41-serviceca\") pod \"node-ca-8fg7f\" (UID: \"210b9f6a-802b-442b-a940-66c003e96c41\") " pod="openshift-image-registry/node-ca-8fg7f" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.972969 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-run-systemd\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.973135 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-host-cni-bin\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.973173 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-etc-openvswitch\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.973206 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-var-lib-openvswitch\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.973496 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d340687c-faf0-48dc-a1ac-dec5067ac665-ovnkube-config\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.973698 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/210b9f6a-802b-442b-a940-66c003e96c41-host\") pod \"node-ca-8fg7f\" (UID: \"210b9f6a-802b-442b-a940-66c003e96c41\") " pod="openshift-image-registry/node-ca-8fg7f" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.973766 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d340687c-faf0-48dc-a1ac-dec5067ac665-ovnkube-script-lib\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.974013 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d340687c-faf0-48dc-a1ac-dec5067ac665-env-overrides\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.975593 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d340687c-faf0-48dc-a1ac-dec5067ac665-ovn-node-metrics-cert\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.999583 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7d2m\" (UniqueName: \"kubernetes.io/projected/d340687c-faf0-48dc-a1ac-dec5067ac665-kube-api-access-p7d2m\") pod \"ovnkube-node-s8cw7\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:05 crc kubenswrapper[4943]: I0307 14:41:05.999646 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4496\" (UniqueName: \"kubernetes.io/projected/210b9f6a-802b-442b-a940-66c003e96c41-kube-api-access-z4496\") pod \"node-ca-8fg7f\" (UID: \"210b9f6a-802b-442b-a940-66c003e96c41\") " pod="openshift-image-registry/node-ca-8fg7f" Mar 07 14:41:06 crc kubenswrapper[4943]: I0307 14:41:06.072281 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9b3f2813-5e5c-4e52-bdcf-95b1b2939c1a-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-ctnkz\" (UID: \"9b3f2813-5e5c-4e52-bdcf-95b1b2939c1a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctnkz" Mar 07 14:41:06 crc kubenswrapper[4943]: I0307 14:41:06.072354 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/9b3f2813-5e5c-4e52-bdcf-95b1b2939c1a-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-ctnkz\" (UID: \"9b3f2813-5e5c-4e52-bdcf-95b1b2939c1a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctnkz" Mar 07 14:41:06 crc kubenswrapper[4943]: I0307 14:41:06.072397 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b3f2813-5e5c-4e52-bdcf-95b1b2939c1a-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-ctnkz\" (UID: \"9b3f2813-5e5c-4e52-bdcf-95b1b2939c1a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctnkz" Mar 07 14:41:06 crc kubenswrapper[4943]: I0307 14:41:06.072430 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/9b3f2813-5e5c-4e52-bdcf-95b1b2939c1a-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-ctnkz\" (UID: \"9b3f2813-5e5c-4e52-bdcf-95b1b2939c1a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctnkz" Mar 07 14:41:06 crc kubenswrapper[4943]: I0307 14:41:06.072506 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9b3f2813-5e5c-4e52-bdcf-95b1b2939c1a-service-ca\") pod \"cluster-version-operator-5c965bbfc6-ctnkz\" (UID: \"9b3f2813-5e5c-4e52-bdcf-95b1b2939c1a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctnkz" Mar 07 14:41:06 crc kubenswrapper[4943]: I0307 14:41:06.072884 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/9b3f2813-5e5c-4e52-bdcf-95b1b2939c1a-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-ctnkz\" (UID: \"9b3f2813-5e5c-4e52-bdcf-95b1b2939c1a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctnkz" Mar 07 14:41:06 crc kubenswrapper[4943]: I0307 14:41:06.072947 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/9b3f2813-5e5c-4e52-bdcf-95b1b2939c1a-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-ctnkz\" (UID: \"9b3f2813-5e5c-4e52-bdcf-95b1b2939c1a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctnkz" Mar 07 14:41:06 crc kubenswrapper[4943]: I0307 14:41:06.073108 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" Mar 07 14:41:06 crc kubenswrapper[4943]: I0307 14:41:06.073899 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9b3f2813-5e5c-4e52-bdcf-95b1b2939c1a-service-ca\") pod \"cluster-version-operator-5c965bbfc6-ctnkz\" (UID: \"9b3f2813-5e5c-4e52-bdcf-95b1b2939c1a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctnkz" Mar 07 14:41:06 crc kubenswrapper[4943]: I0307 14:41:06.077526 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b3f2813-5e5c-4e52-bdcf-95b1b2939c1a-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-ctnkz\" (UID: \"9b3f2813-5e5c-4e52-bdcf-95b1b2939c1a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctnkz" Mar 07 14:41:06 crc kubenswrapper[4943]: I0307 14:41:06.082624 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-qrlpk" Mar 07 14:41:06 crc kubenswrapper[4943]: I0307 14:41:06.092997 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-npf7n" Mar 07 14:41:06 crc kubenswrapper[4943]: W0307 14:41:06.095234 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4c1f18ea_e557_41fc_95dd_ce4e1722a61e.slice/crio-ce1c892a6da37a7c070c095d8e95d4f1e6b3a7cb3789fcb8b8fcb4e4cbbe518a WatchSource:0}: Error finding container ce1c892a6da37a7c070c095d8e95d4f1e6b3a7cb3789fcb8b8fcb4e4cbbe518a: Status 404 returned error can't find the container with id ce1c892a6da37a7c070c095d8e95d4f1e6b3a7cb3789fcb8b8fcb4e4cbbe518a Mar 07 14:41:06 crc kubenswrapper[4943]: W0307 14:41:06.100331 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda13f1d2e_9974_4c40_85f5_d876d5644993.slice/crio-d4bda950676e3e615aaba5f9f5a50f002bc67f767d705b3a770a062f1a4f7a1d WatchSource:0}: Error finding container d4bda950676e3e615aaba5f9f5a50f002bc67f767d705b3a770a062f1a4f7a1d: Status 404 returned error can't find the container with id d4bda950676e3e615aaba5f9f5a50f002bc67f767d705b3a770a062f1a4f7a1d Mar 07 14:41:06 crc kubenswrapper[4943]: I0307 14:41:06.100908 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9b3f2813-5e5c-4e52-bdcf-95b1b2939c1a-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-ctnkz\" (UID: \"9b3f2813-5e5c-4e52-bdcf-95b1b2939c1a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctnkz" Mar 07 14:41:06 crc kubenswrapper[4943]: W0307 14:41:06.111026 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod27cc8510_d4f1_476e_8d35_d7c64cec5e99.slice/crio-7a672d2d38417af323ef0df6465a9fbb6ccd21c23e29a5b801e64ce597761f32 WatchSource:0}: Error finding container 7a672d2d38417af323ef0df6465a9fbb6ccd21c23e29a5b801e64ce597761f32: Status 404 returned error can't find the container with id 7a672d2d38417af323ef0df6465a9fbb6ccd21c23e29a5b801e64ce597761f32 Mar 07 14:41:06 crc kubenswrapper[4943]: I0307 14:41:06.112327 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-4j27g" Mar 07 14:41:06 crc kubenswrapper[4943]: I0307 14:41:06.122072 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:06 crc kubenswrapper[4943]: I0307 14:41:06.132399 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-8fg7f" Mar 07 14:41:06 crc kubenswrapper[4943]: W0307 14:41:06.140730 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod65707314_fccf_4ae5_8790_eb24f63ad3ad.slice/crio-0ab4b22fad5f1bc0c958b38f5ffb00c0cc0fd011fdbad9453ad271d5320304fd WatchSource:0}: Error finding container 0ab4b22fad5f1bc0c958b38f5ffb00c0cc0fd011fdbad9453ad271d5320304fd: Status 404 returned error can't find the container with id 0ab4b22fad5f1bc0c958b38f5ffb00c0cc0fd011fdbad9453ad271d5320304fd Mar 07 14:41:06 crc kubenswrapper[4943]: W0307 14:41:06.159975 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd340687c_faf0_48dc_a1ac_dec5067ac665.slice/crio-15fafcc9303d61bcfb6c9516eb858fa860668edc2f434892f764a3af40ac36ed WatchSource:0}: Error finding container 15fafcc9303d61bcfb6c9516eb858fa860668edc2f434892f764a3af40ac36ed: Status 404 returned error can't find the container with id 15fafcc9303d61bcfb6c9516eb858fa860668edc2f434892f764a3af40ac36ed Mar 07 14:41:06 crc kubenswrapper[4943]: W0307 14:41:06.161917 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod210b9f6a_802b_442b_a940_66c003e96c41.slice/crio-a17ca7ac56872084b139bd42a23b4156b14c8b3b34476c21657d2c8131d2bd60 WatchSource:0}: Error finding container a17ca7ac56872084b139bd42a23b4156b14c8b3b34476c21657d2c8131d2bd60: Status 404 returned error can't find the container with id a17ca7ac56872084b139bd42a23b4156b14c8b3b34476c21657d2c8131d2bd60 Mar 07 14:41:06 crc kubenswrapper[4943]: I0307 14:41:06.189326 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p2w65"] Mar 07 14:41:06 crc kubenswrapper[4943]: I0307 14:41:06.190962 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p2w65" Mar 07 14:41:06 crc kubenswrapper[4943]: I0307 14:41:06.194617 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Mar 07 14:41:06 crc kubenswrapper[4943]: I0307 14:41:06.197671 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Mar 07 14:41:06 crc kubenswrapper[4943]: I0307 14:41:06.225040 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-ntc42"] Mar 07 14:41:06 crc kubenswrapper[4943]: I0307 14:41:06.225696 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ntc42" Mar 07 14:41:06 crc kubenswrapper[4943]: E0307 14:41:06.225773 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ntc42" podUID="e6f60f83-1b2e-4f75-9065-9971654e878a" Mar 07 14:41:06 crc kubenswrapper[4943]: I0307 14:41:06.228156 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctnkz" Mar 07 14:41:06 crc kubenswrapper[4943]: I0307 14:41:06.248479 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" event={"ID":"4c1f18ea-e557-41fc-95dd-ce4e1722a61e","Type":"ContainerStarted","Data":"ce1c892a6da37a7c070c095d8e95d4f1e6b3a7cb3789fcb8b8fcb4e4cbbe518a"} Mar 07 14:41:06 crc kubenswrapper[4943]: I0307 14:41:06.249612 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-npf7n" event={"ID":"27cc8510-d4f1-476e-8d35-d7c64cec5e99","Type":"ContainerStarted","Data":"7a672d2d38417af323ef0df6465a9fbb6ccd21c23e29a5b801e64ce597761f32"} Mar 07 14:41:06 crc kubenswrapper[4943]: I0307 14:41:06.250498 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-8fg7f" event={"ID":"210b9f6a-802b-442b-a940-66c003e96c41","Type":"ContainerStarted","Data":"a17ca7ac56872084b139bd42a23b4156b14c8b3b34476c21657d2c8131d2bd60"} Mar 07 14:41:06 crc kubenswrapper[4943]: I0307 14:41:06.251534 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" event={"ID":"d340687c-faf0-48dc-a1ac-dec5067ac665","Type":"ContainerStarted","Data":"15fafcc9303d61bcfb6c9516eb858fa860668edc2f434892f764a3af40ac36ed"} Mar 07 14:41:06 crc kubenswrapper[4943]: I0307 14:41:06.253673 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4j27g" event={"ID":"65707314-fccf-4ae5-8790-eb24f63ad3ad","Type":"ContainerStarted","Data":"0ab4b22fad5f1bc0c958b38f5ffb00c0cc0fd011fdbad9453ad271d5320304fd"} Mar 07 14:41:06 crc kubenswrapper[4943]: I0307 14:41:06.254597 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-qrlpk" event={"ID":"a13f1d2e-9974-4c40-85f5-d876d5644993","Type":"ContainerStarted","Data":"d4bda950676e3e615aaba5f9f5a50f002bc67f767d705b3a770a062f1a4f7a1d"} Mar 07 14:41:06 crc kubenswrapper[4943]: W0307 14:41:06.264652 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9b3f2813_5e5c_4e52_bdcf_95b1b2939c1a.slice/crio-a4515fae7a75570211feebe7dad29465854c87b1b5413e55f0d0a515ea640c38 WatchSource:0}: Error finding container a4515fae7a75570211feebe7dad29465854c87b1b5413e55f0d0a515ea640c38: Status 404 returned error can't find the container with id a4515fae7a75570211feebe7dad29465854c87b1b5413e55f0d0a515ea640c38 Mar 07 14:41:06 crc kubenswrapper[4943]: I0307 14:41:06.274365 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/96e7bc5a-bdce-4284-b08f-0971de4d5232-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-p2w65\" (UID: \"96e7bc5a-bdce-4284-b08f-0971de4d5232\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p2w65" Mar 07 14:41:06 crc kubenswrapper[4943]: I0307 14:41:06.274431 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/96e7bc5a-bdce-4284-b08f-0971de4d5232-env-overrides\") pod \"ovnkube-control-plane-749d76644c-p2w65\" (UID: \"96e7bc5a-bdce-4284-b08f-0971de4d5232\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p2w65" Mar 07 14:41:06 crc kubenswrapper[4943]: I0307 14:41:06.274522 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mb6bh\" (UniqueName: \"kubernetes.io/projected/96e7bc5a-bdce-4284-b08f-0971de4d5232-kube-api-access-mb6bh\") pod \"ovnkube-control-plane-749d76644c-p2w65\" (UID: \"96e7bc5a-bdce-4284-b08f-0971de4d5232\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p2w65" Mar 07 14:41:06 crc kubenswrapper[4943]: I0307 14:41:06.274571 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e6f60f83-1b2e-4f75-9065-9971654e878a-metrics-certs\") pod \"network-metrics-daemon-ntc42\" (UID: \"e6f60f83-1b2e-4f75-9065-9971654e878a\") " pod="openshift-multus/network-metrics-daemon-ntc42" Mar 07 14:41:06 crc kubenswrapper[4943]: I0307 14:41:06.274621 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nf62j\" (UniqueName: \"kubernetes.io/projected/e6f60f83-1b2e-4f75-9065-9971654e878a-kube-api-access-nf62j\") pod \"network-metrics-daemon-ntc42\" (UID: \"e6f60f83-1b2e-4f75-9065-9971654e878a\") " pod="openshift-multus/network-metrics-daemon-ntc42" Mar 07 14:41:06 crc kubenswrapper[4943]: I0307 14:41:06.274678 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/96e7bc5a-bdce-4284-b08f-0971de4d5232-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-p2w65\" (UID: \"96e7bc5a-bdce-4284-b08f-0971de4d5232\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p2w65" Mar 07 14:41:06 crc kubenswrapper[4943]: I0307 14:41:06.375596 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e6f60f83-1b2e-4f75-9065-9971654e878a-metrics-certs\") pod \"network-metrics-daemon-ntc42\" (UID: \"e6f60f83-1b2e-4f75-9065-9971654e878a\") " pod="openshift-multus/network-metrics-daemon-ntc42" Mar 07 14:41:06 crc kubenswrapper[4943]: I0307 14:41:06.375641 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nf62j\" (UniqueName: \"kubernetes.io/projected/e6f60f83-1b2e-4f75-9065-9971654e878a-kube-api-access-nf62j\") pod \"network-metrics-daemon-ntc42\" (UID: \"e6f60f83-1b2e-4f75-9065-9971654e878a\") " pod="openshift-multus/network-metrics-daemon-ntc42" Mar 07 14:41:06 crc kubenswrapper[4943]: I0307 14:41:06.375684 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/96e7bc5a-bdce-4284-b08f-0971de4d5232-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-p2w65\" (UID: \"96e7bc5a-bdce-4284-b08f-0971de4d5232\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p2w65" Mar 07 14:41:06 crc kubenswrapper[4943]: I0307 14:41:06.375710 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/96e7bc5a-bdce-4284-b08f-0971de4d5232-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-p2w65\" (UID: \"96e7bc5a-bdce-4284-b08f-0971de4d5232\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p2w65" Mar 07 14:41:06 crc kubenswrapper[4943]: I0307 14:41:06.375757 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/96e7bc5a-bdce-4284-b08f-0971de4d5232-env-overrides\") pod \"ovnkube-control-plane-749d76644c-p2w65\" (UID: \"96e7bc5a-bdce-4284-b08f-0971de4d5232\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p2w65" Mar 07 14:41:06 crc kubenswrapper[4943]: I0307 14:41:06.375789 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mb6bh\" (UniqueName: \"kubernetes.io/projected/96e7bc5a-bdce-4284-b08f-0971de4d5232-kube-api-access-mb6bh\") pod \"ovnkube-control-plane-749d76644c-p2w65\" (UID: \"96e7bc5a-bdce-4284-b08f-0971de4d5232\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p2w65" Mar 07 14:41:06 crc kubenswrapper[4943]: E0307 14:41:06.375838 4943 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 07 14:41:06 crc kubenswrapper[4943]: E0307 14:41:06.375997 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e6f60f83-1b2e-4f75-9065-9971654e878a-metrics-certs podName:e6f60f83-1b2e-4f75-9065-9971654e878a nodeName:}" failed. No retries permitted until 2026-03-07 14:41:06.875968424 +0000 UTC m=+108.828104952 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e6f60f83-1b2e-4f75-9065-9971654e878a-metrics-certs") pod "network-metrics-daemon-ntc42" (UID: "e6f60f83-1b2e-4f75-9065-9971654e878a") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 07 14:41:06 crc kubenswrapper[4943]: I0307 14:41:06.376571 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/96e7bc5a-bdce-4284-b08f-0971de4d5232-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-p2w65\" (UID: \"96e7bc5a-bdce-4284-b08f-0971de4d5232\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p2w65" Mar 07 14:41:06 crc kubenswrapper[4943]: I0307 14:41:06.377004 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/96e7bc5a-bdce-4284-b08f-0971de4d5232-env-overrides\") pod \"ovnkube-control-plane-749d76644c-p2w65\" (UID: \"96e7bc5a-bdce-4284-b08f-0971de4d5232\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p2w65" Mar 07 14:41:06 crc kubenswrapper[4943]: I0307 14:41:06.382177 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/96e7bc5a-bdce-4284-b08f-0971de4d5232-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-p2w65\" (UID: \"96e7bc5a-bdce-4284-b08f-0971de4d5232\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p2w65" Mar 07 14:41:06 crc kubenswrapper[4943]: I0307 14:41:06.394386 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nf62j\" (UniqueName: \"kubernetes.io/projected/e6f60f83-1b2e-4f75-9065-9971654e878a-kube-api-access-nf62j\") pod \"network-metrics-daemon-ntc42\" (UID: \"e6f60f83-1b2e-4f75-9065-9971654e878a\") " pod="openshift-multus/network-metrics-daemon-ntc42" Mar 07 14:41:06 crc kubenswrapper[4943]: I0307 14:41:06.396729 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mb6bh\" (UniqueName: \"kubernetes.io/projected/96e7bc5a-bdce-4284-b08f-0971de4d5232-kube-api-access-mb6bh\") pod \"ovnkube-control-plane-749d76644c-p2w65\" (UID: \"96e7bc5a-bdce-4284-b08f-0971de4d5232\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p2w65" Mar 07 14:41:06 crc kubenswrapper[4943]: I0307 14:41:06.510082 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p2w65" Mar 07 14:41:06 crc kubenswrapper[4943]: W0307 14:41:06.526328 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod96e7bc5a_bdce_4284_b08f_0971de4d5232.slice/crio-50fe7bcb1bd2c7b05f3bcb4cdf1a4df13a967d7e7f7ea5734ff9e8d39bd71ff2 WatchSource:0}: Error finding container 50fe7bcb1bd2c7b05f3bcb4cdf1a4df13a967d7e7f7ea5734ff9e8d39bd71ff2: Status 404 returned error can't find the container with id 50fe7bcb1bd2c7b05f3bcb4cdf1a4df13a967d7e7f7ea5734ff9e8d39bd71ff2 Mar 07 14:41:06 crc kubenswrapper[4943]: I0307 14:41:06.882723 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e6f60f83-1b2e-4f75-9065-9971654e878a-metrics-certs\") pod \"network-metrics-daemon-ntc42\" (UID: \"e6f60f83-1b2e-4f75-9065-9971654e878a\") " pod="openshift-multus/network-metrics-daemon-ntc42" Mar 07 14:41:06 crc kubenswrapper[4943]: E0307 14:41:06.882986 4943 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 07 14:41:06 crc kubenswrapper[4943]: E0307 14:41:06.883094 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e6f60f83-1b2e-4f75-9065-9971654e878a-metrics-certs podName:e6f60f83-1b2e-4f75-9065-9971654e878a nodeName:}" failed. No retries permitted until 2026-03-07 14:41:07.88306067 +0000 UTC m=+109.835197198 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e6f60f83-1b2e-4f75-9065-9971654e878a-metrics-certs") pod "network-metrics-daemon-ntc42" (UID: "e6f60f83-1b2e-4f75-9065-9971654e878a") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 07 14:41:07 crc kubenswrapper[4943]: I0307 14:41:07.261139 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" event={"ID":"4c1f18ea-e557-41fc-95dd-ce4e1722a61e","Type":"ContainerStarted","Data":"ccef5f155520b8fd91afa6d9d5bb268ac15244573eccd1e9d89c5841bc00f721"} Mar 07 14:41:07 crc kubenswrapper[4943]: I0307 14:41:07.261552 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" event={"ID":"4c1f18ea-e557-41fc-95dd-ce4e1722a61e","Type":"ContainerStarted","Data":"fca9947e0c4ec525f0c6b799660a1244b1d37a4616b55a03cd6f059674d81757"} Mar 07 14:41:07 crc kubenswrapper[4943]: I0307 14:41:07.263824 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-npf7n" event={"ID":"27cc8510-d4f1-476e-8d35-d7c64cec5e99","Type":"ContainerStarted","Data":"15bf318695774f2bec62fc76d0dc77bca1e58de8838f9fa8592fbe61712f77ad"} Mar 07 14:41:07 crc kubenswrapper[4943]: I0307 14:41:07.266439 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-8fg7f" event={"ID":"210b9f6a-802b-442b-a940-66c003e96c41","Type":"ContainerStarted","Data":"2679a3133dcd6979a95b61e16055756e4002af202b6d29a49a0f6d9058321400"} Mar 07 14:41:07 crc kubenswrapper[4943]: I0307 14:41:07.268592 4943 generic.go:334] "Generic (PLEG): container finished" podID="d340687c-faf0-48dc-a1ac-dec5067ac665" containerID="ce64b91eda434340bdacc6a953afe291459d3b18a9cbf842dc8d81723fcdd70e" exitCode=0 Mar 07 14:41:07 crc kubenswrapper[4943]: I0307 14:41:07.268683 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" event={"ID":"d340687c-faf0-48dc-a1ac-dec5067ac665","Type":"ContainerDied","Data":"ce64b91eda434340bdacc6a953afe291459d3b18a9cbf842dc8d81723fcdd70e"} Mar 07 14:41:07 crc kubenswrapper[4943]: I0307 14:41:07.271534 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctnkz" event={"ID":"9b3f2813-5e5c-4e52-bdcf-95b1b2939c1a","Type":"ContainerStarted","Data":"41ff02c06375463dde159904936844bd24e10344bcdcb58e2d2644e5e50556b3"} Mar 07 14:41:07 crc kubenswrapper[4943]: I0307 14:41:07.271572 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctnkz" event={"ID":"9b3f2813-5e5c-4e52-bdcf-95b1b2939c1a","Type":"ContainerStarted","Data":"a4515fae7a75570211feebe7dad29465854c87b1b5413e55f0d0a515ea640c38"} Mar 07 14:41:07 crc kubenswrapper[4943]: I0307 14:41:07.274072 4943 generic.go:334] "Generic (PLEG): container finished" podID="65707314-fccf-4ae5-8790-eb24f63ad3ad" containerID="936e731a7a610db38ce677ed985b228c2520318431ef8ed264b6e2db181c1253" exitCode=0 Mar 07 14:41:07 crc kubenswrapper[4943]: I0307 14:41:07.274142 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4j27g" event={"ID":"65707314-fccf-4ae5-8790-eb24f63ad3ad","Type":"ContainerDied","Data":"936e731a7a610db38ce677ed985b228c2520318431ef8ed264b6e2db181c1253"} Mar 07 14:41:07 crc kubenswrapper[4943]: I0307 14:41:07.278218 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p2w65" event={"ID":"96e7bc5a-bdce-4284-b08f-0971de4d5232","Type":"ContainerStarted","Data":"391e940e0189e6a1d3d3faf1accaa7de3f1f72c62f5a1e276c64f7ece2c269b5"} Mar 07 14:41:07 crc kubenswrapper[4943]: I0307 14:41:07.278334 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p2w65" event={"ID":"96e7bc5a-bdce-4284-b08f-0971de4d5232","Type":"ContainerStarted","Data":"00646ddf5b1ae175fe0b50943475052bb5bed3dd0517f7b9a9d0c003735985ea"} Mar 07 14:41:07 crc kubenswrapper[4943]: I0307 14:41:07.278365 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p2w65" event={"ID":"96e7bc5a-bdce-4284-b08f-0971de4d5232","Type":"ContainerStarted","Data":"50fe7bcb1bd2c7b05f3bcb4cdf1a4df13a967d7e7f7ea5734ff9e8d39bd71ff2"} Mar 07 14:41:07 crc kubenswrapper[4943]: I0307 14:41:07.282598 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-qrlpk" event={"ID":"a13f1d2e-9974-4c40-85f5-d876d5644993","Type":"ContainerStarted","Data":"ea59bca2daa4e410ff69f16a6b5d802dd7375344ebb97d3d5eda572e70bb2db1"} Mar 07 14:41:07 crc kubenswrapper[4943]: I0307 14:41:07.296005 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podStartSLOduration=35.295970382 podStartE2EDuration="35.295970382s" podCreationTimestamp="2026-03-07 14:40:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:07.295684105 +0000 UTC m=+109.247820643" watchObservedRunningTime="2026-03-07 14:41:07.295970382 +0000 UTC m=+109.248106930" Mar 07 14:41:07 crc kubenswrapper[4943]: I0307 14:41:07.338378 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-qrlpk" podStartSLOduration=35.338351852 podStartE2EDuration="35.338351852s" podCreationTimestamp="2026-03-07 14:40:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:07.324662806 +0000 UTC m=+109.276799344" watchObservedRunningTime="2026-03-07 14:41:07.338351852 +0000 UTC m=+109.290488380" Mar 07 14:41:07 crc kubenswrapper[4943]: I0307 14:41:07.339802 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-npf7n" podStartSLOduration=36.339792769 podStartE2EDuration="36.339792769s" podCreationTimestamp="2026-03-07 14:40:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:07.339138512 +0000 UTC m=+109.291275050" watchObservedRunningTime="2026-03-07 14:41:07.339792769 +0000 UTC m=+109.291929307" Mar 07 14:41:07 crc kubenswrapper[4943]: I0307 14:41:07.356809 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p2w65" podStartSLOduration=35.35679526 podStartE2EDuration="35.35679526s" podCreationTimestamp="2026-03-07 14:40:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:07.355833835 +0000 UTC m=+109.307970363" watchObservedRunningTime="2026-03-07 14:41:07.35679526 +0000 UTC m=+109.308931788" Mar 07 14:41:07 crc kubenswrapper[4943]: I0307 14:41:07.377819 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctnkz" podStartSLOduration=35.377792815 podStartE2EDuration="35.377792815s" podCreationTimestamp="2026-03-07 14:40:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:07.37682266 +0000 UTC m=+109.328959198" watchObservedRunningTime="2026-03-07 14:41:07.377792815 +0000 UTC m=+109.329929343" Mar 07 14:41:07 crc kubenswrapper[4943]: I0307 14:41:07.462215 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-8fg7f" podStartSLOduration=36.462193694 podStartE2EDuration="36.462193694s" podCreationTimestamp="2026-03-07 14:40:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:07.461851166 +0000 UTC m=+109.413987674" watchObservedRunningTime="2026-03-07 14:41:07.462193694 +0000 UTC m=+109.414330192" Mar 07 14:41:07 crc kubenswrapper[4943]: I0307 14:41:07.754462 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 07 14:41:07 crc kubenswrapper[4943]: E0307 14:41:07.754864 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 07 14:41:07 crc kubenswrapper[4943]: I0307 14:41:07.755426 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 07 14:41:07 crc kubenswrapper[4943]: I0307 14:41:07.755834 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 07 14:41:07 crc kubenswrapper[4943]: E0307 14:41:07.755921 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 07 14:41:07 crc kubenswrapper[4943]: E0307 14:41:07.757245 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 07 14:41:07 crc kubenswrapper[4943]: I0307 14:41:07.757910 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ntc42" Mar 07 14:41:07 crc kubenswrapper[4943]: E0307 14:41:07.758710 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ntc42" podUID="e6f60f83-1b2e-4f75-9065-9971654e878a" Mar 07 14:41:07 crc kubenswrapper[4943]: I0307 14:41:07.896879 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e6f60f83-1b2e-4f75-9065-9971654e878a-metrics-certs\") pod \"network-metrics-daemon-ntc42\" (UID: \"e6f60f83-1b2e-4f75-9065-9971654e878a\") " pod="openshift-multus/network-metrics-daemon-ntc42" Mar 07 14:41:07 crc kubenswrapper[4943]: E0307 14:41:07.897108 4943 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 07 14:41:07 crc kubenswrapper[4943]: E0307 14:41:07.897228 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e6f60f83-1b2e-4f75-9065-9971654e878a-metrics-certs podName:e6f60f83-1b2e-4f75-9065-9971654e878a nodeName:}" failed. No retries permitted until 2026-03-07 14:41:09.89719391 +0000 UTC m=+111.849330448 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e6f60f83-1b2e-4f75-9065-9971654e878a-metrics-certs") pod "network-metrics-daemon-ntc42" (UID: "e6f60f83-1b2e-4f75-9065-9971654e878a") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 07 14:41:08 crc kubenswrapper[4943]: I0307 14:41:08.292568 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4j27g" event={"ID":"65707314-fccf-4ae5-8790-eb24f63ad3ad","Type":"ContainerStarted","Data":"ffb317d00bde88fda0d802ccbdec2e36e5ca6a98a2da441094d0a2ab82edad71"} Mar 07 14:41:08 crc kubenswrapper[4943]: I0307 14:41:08.297346 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" event={"ID":"d340687c-faf0-48dc-a1ac-dec5067ac665","Type":"ContainerStarted","Data":"a64d887ab03c882e1a27a4cf36f359a8d18e3c75b0a6d81f33569e1e3b294280"} Mar 07 14:41:08 crc kubenswrapper[4943]: I0307 14:41:08.297395 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" event={"ID":"d340687c-faf0-48dc-a1ac-dec5067ac665","Type":"ContainerStarted","Data":"6fb3f1769f0d1092192c2f5f152c7d1fc52813d2e722687fb3ea1f6f08fba808"} Mar 07 14:41:08 crc kubenswrapper[4943]: I0307 14:41:08.297410 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" event={"ID":"d340687c-faf0-48dc-a1ac-dec5067ac665","Type":"ContainerStarted","Data":"4dd5c06470991a2bcdddf5a5e429adcf4f4af8d3742de9c0a1e9cdd5705f74c5"} Mar 07 14:41:08 crc kubenswrapper[4943]: I0307 14:41:08.297423 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" event={"ID":"d340687c-faf0-48dc-a1ac-dec5067ac665","Type":"ContainerStarted","Data":"6a089f76e974f8c12f32f42b96b301fda03651276c599841b9c42fa064ffd5fd"} Mar 07 14:41:08 crc kubenswrapper[4943]: I0307 14:41:08.757508 4943 scope.go:117] "RemoveContainer" containerID="ed8d27f5a369a691493bda9dbd6e03c098931b8dab637bbd9bacac6fe7979c80" Mar 07 14:41:08 crc kubenswrapper[4943]: E0307 14:41:08.757903 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 07 14:41:09 crc kubenswrapper[4943]: I0307 14:41:09.304840 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" event={"ID":"d340687c-faf0-48dc-a1ac-dec5067ac665","Type":"ContainerStarted","Data":"50e79fc40569f5b3fd73286933ee0242a1edd807a820edf7ba00a856d81c82c9"} Mar 07 14:41:09 crc kubenswrapper[4943]: I0307 14:41:09.305475 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" event={"ID":"d340687c-faf0-48dc-a1ac-dec5067ac665","Type":"ContainerStarted","Data":"959458a05302e835122b9db5d1eb4c449749a9e7ba2d85459edaeef2eede12de"} Mar 07 14:41:09 crc kubenswrapper[4943]: I0307 14:41:09.307139 4943 generic.go:334] "Generic (PLEG): container finished" podID="65707314-fccf-4ae5-8790-eb24f63ad3ad" containerID="ffb317d00bde88fda0d802ccbdec2e36e5ca6a98a2da441094d0a2ab82edad71" exitCode=0 Mar 07 14:41:09 crc kubenswrapper[4943]: I0307 14:41:09.307208 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4j27g" event={"ID":"65707314-fccf-4ae5-8790-eb24f63ad3ad","Type":"ContainerDied","Data":"ffb317d00bde88fda0d802ccbdec2e36e5ca6a98a2da441094d0a2ab82edad71"} Mar 07 14:41:09 crc kubenswrapper[4943]: I0307 14:41:09.754967 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 07 14:41:09 crc kubenswrapper[4943]: I0307 14:41:09.755064 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 07 14:41:09 crc kubenswrapper[4943]: I0307 14:41:09.755092 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ntc42" Mar 07 14:41:09 crc kubenswrapper[4943]: I0307 14:41:09.755183 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 07 14:41:09 crc kubenswrapper[4943]: E0307 14:41:09.755324 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 07 14:41:09 crc kubenswrapper[4943]: E0307 14:41:09.755469 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 07 14:41:09 crc kubenswrapper[4943]: E0307 14:41:09.755621 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ntc42" podUID="e6f60f83-1b2e-4f75-9065-9971654e878a" Mar 07 14:41:09 crc kubenswrapper[4943]: E0307 14:41:09.755740 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 07 14:41:09 crc kubenswrapper[4943]: I0307 14:41:09.928310 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e6f60f83-1b2e-4f75-9065-9971654e878a-metrics-certs\") pod \"network-metrics-daemon-ntc42\" (UID: \"e6f60f83-1b2e-4f75-9065-9971654e878a\") " pod="openshift-multus/network-metrics-daemon-ntc42" Mar 07 14:41:09 crc kubenswrapper[4943]: E0307 14:41:09.928500 4943 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 07 14:41:09 crc kubenswrapper[4943]: E0307 14:41:09.928561 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e6f60f83-1b2e-4f75-9065-9971654e878a-metrics-certs podName:e6f60f83-1b2e-4f75-9065-9971654e878a nodeName:}" failed. No retries permitted until 2026-03-07 14:41:13.92854627 +0000 UTC m=+115.880682768 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e6f60f83-1b2e-4f75-9065-9971654e878a-metrics-certs") pod "network-metrics-daemon-ntc42" (UID: "e6f60f83-1b2e-4f75-9065-9971654e878a") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 07 14:41:10 crc kubenswrapper[4943]: I0307 14:41:10.313443 4943 generic.go:334] "Generic (PLEG): container finished" podID="65707314-fccf-4ae5-8790-eb24f63ad3ad" containerID="883d02027fb9dd43b1d2536d3180eb52646a210f2f592104313dfde2a35dc8c6" exitCode=0 Mar 07 14:41:10 crc kubenswrapper[4943]: I0307 14:41:10.313513 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4j27g" event={"ID":"65707314-fccf-4ae5-8790-eb24f63ad3ad","Type":"ContainerDied","Data":"883d02027fb9dd43b1d2536d3180eb52646a210f2f592104313dfde2a35dc8c6"} Mar 07 14:41:11 crc kubenswrapper[4943]: I0307 14:41:11.323364 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" event={"ID":"d340687c-faf0-48dc-a1ac-dec5067ac665","Type":"ContainerStarted","Data":"f56c9c195453268c830bf92a70a64f70e75ef30e8ca853bb1f3a054a69c791af"} Mar 07 14:41:11 crc kubenswrapper[4943]: I0307 14:41:11.327453 4943 generic.go:334] "Generic (PLEG): container finished" podID="65707314-fccf-4ae5-8790-eb24f63ad3ad" containerID="3b76f311128c1e27b14a24956c33ea4f977a04b4d2639c94643bb040f94d2b41" exitCode=0 Mar 07 14:41:11 crc kubenswrapper[4943]: I0307 14:41:11.327526 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4j27g" event={"ID":"65707314-fccf-4ae5-8790-eb24f63ad3ad","Type":"ContainerDied","Data":"3b76f311128c1e27b14a24956c33ea4f977a04b4d2639c94643bb040f94d2b41"} Mar 07 14:41:11 crc kubenswrapper[4943]: I0307 14:41:11.754801 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 07 14:41:11 crc kubenswrapper[4943]: I0307 14:41:11.754806 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 07 14:41:11 crc kubenswrapper[4943]: I0307 14:41:11.754817 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 07 14:41:11 crc kubenswrapper[4943]: I0307 14:41:11.754822 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ntc42" Mar 07 14:41:11 crc kubenswrapper[4943]: E0307 14:41:11.755095 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 07 14:41:11 crc kubenswrapper[4943]: E0307 14:41:11.755279 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ntc42" podUID="e6f60f83-1b2e-4f75-9065-9971654e878a" Mar 07 14:41:11 crc kubenswrapper[4943]: E0307 14:41:11.755388 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 07 14:41:11 crc kubenswrapper[4943]: E0307 14:41:11.755492 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 07 14:41:12 crc kubenswrapper[4943]: I0307 14:41:12.336525 4943 generic.go:334] "Generic (PLEG): container finished" podID="65707314-fccf-4ae5-8790-eb24f63ad3ad" containerID="78e670f1ad374acc4d1c938dfaff278c51e9965e0d18efed197db596a351b905" exitCode=0 Mar 07 14:41:12 crc kubenswrapper[4943]: I0307 14:41:12.338377 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4j27g" event={"ID":"65707314-fccf-4ae5-8790-eb24f63ad3ad","Type":"ContainerDied","Data":"78e670f1ad374acc4d1c938dfaff278c51e9965e0d18efed197db596a351b905"} Mar 07 14:41:13 crc kubenswrapper[4943]: I0307 14:41:13.352780 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" event={"ID":"d340687c-faf0-48dc-a1ac-dec5067ac665","Type":"ContainerStarted","Data":"a200f68bc20411f516725f81a930664ba5baaefd21a2792866262ded7e86fae5"} Mar 07 14:41:13 crc kubenswrapper[4943]: I0307 14:41:13.353301 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:13 crc kubenswrapper[4943]: I0307 14:41:13.353508 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:13 crc kubenswrapper[4943]: I0307 14:41:13.353546 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:13 crc kubenswrapper[4943]: I0307 14:41:13.361519 4943 generic.go:334] "Generic (PLEG): container finished" podID="65707314-fccf-4ae5-8790-eb24f63ad3ad" containerID="1a8ff5a7ef6978a796ddb6c7f64ae34e015e482a85020df3c058c25bdd3aae0f" exitCode=0 Mar 07 14:41:13 crc kubenswrapper[4943]: I0307 14:41:13.361570 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4j27g" event={"ID":"65707314-fccf-4ae5-8790-eb24f63ad3ad","Type":"ContainerDied","Data":"1a8ff5a7ef6978a796ddb6c7f64ae34e015e482a85020df3c058c25bdd3aae0f"} Mar 07 14:41:13 crc kubenswrapper[4943]: I0307 14:41:13.441911 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:13 crc kubenswrapper[4943]: I0307 14:41:13.442065 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:13 crc kubenswrapper[4943]: I0307 14:41:13.475348 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" podStartSLOduration=41.475319965 podStartE2EDuration="41.475319965s" podCreationTimestamp="2026-03-07 14:40:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:13.447779741 +0000 UTC m=+115.399916279" watchObservedRunningTime="2026-03-07 14:41:13.475319965 +0000 UTC m=+115.427456503" Mar 07 14:41:13 crc kubenswrapper[4943]: I0307 14:41:13.755022 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 07 14:41:13 crc kubenswrapper[4943]: I0307 14:41:13.755049 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 07 14:41:13 crc kubenswrapper[4943]: I0307 14:41:13.755049 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 07 14:41:13 crc kubenswrapper[4943]: E0307 14:41:13.756394 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 07 14:41:13 crc kubenswrapper[4943]: E0307 14:41:13.756355 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 07 14:41:13 crc kubenswrapper[4943]: I0307 14:41:13.755151 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ntc42" Mar 07 14:41:13 crc kubenswrapper[4943]: E0307 14:41:13.756545 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 07 14:41:13 crc kubenswrapper[4943]: E0307 14:41:13.756644 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ntc42" podUID="e6f60f83-1b2e-4f75-9065-9971654e878a" Mar 07 14:41:13 crc kubenswrapper[4943]: I0307 14:41:13.982051 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e6f60f83-1b2e-4f75-9065-9971654e878a-metrics-certs\") pod \"network-metrics-daemon-ntc42\" (UID: \"e6f60f83-1b2e-4f75-9065-9971654e878a\") " pod="openshift-multus/network-metrics-daemon-ntc42" Mar 07 14:41:13 crc kubenswrapper[4943]: E0307 14:41:13.982309 4943 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 07 14:41:13 crc kubenswrapper[4943]: E0307 14:41:13.982445 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e6f60f83-1b2e-4f75-9065-9971654e878a-metrics-certs podName:e6f60f83-1b2e-4f75-9065-9971654e878a nodeName:}" failed. No retries permitted until 2026-03-07 14:41:21.982411311 +0000 UTC m=+123.934547839 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e6f60f83-1b2e-4f75-9065-9971654e878a-metrics-certs") pod "network-metrics-daemon-ntc42" (UID: "e6f60f83-1b2e-4f75-9065-9971654e878a") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 07 14:41:14 crc kubenswrapper[4943]: I0307 14:41:14.371707 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4j27g" event={"ID":"65707314-fccf-4ae5-8790-eb24f63ad3ad","Type":"ContainerStarted","Data":"ab34d2c2f423be733bd87ab085d286207c65ccb817cf6e6603a279f08679745f"} Mar 07 14:41:14 crc kubenswrapper[4943]: I0307 14:41:14.404554 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-4j27g" podStartSLOduration=42.404492282 podStartE2EDuration="42.404492282s" podCreationTimestamp="2026-03-07 14:40:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:14.402898901 +0000 UTC m=+116.355035439" watchObservedRunningTime="2026-03-07 14:41:14.404492282 +0000 UTC m=+116.356628810" Mar 07 14:41:15 crc kubenswrapper[4943]: I0307 14:41:15.423531 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-ntc42"] Mar 07 14:41:15 crc kubenswrapper[4943]: I0307 14:41:15.423675 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ntc42" Mar 07 14:41:15 crc kubenswrapper[4943]: E0307 14:41:15.423818 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ntc42" podUID="e6f60f83-1b2e-4f75-9065-9971654e878a" Mar 07 14:41:15 crc kubenswrapper[4943]: I0307 14:41:15.599400 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:15 crc kubenswrapper[4943]: E0307 14:41:15.599653 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:47.599612818 +0000 UTC m=+149.551749356 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:15 crc kubenswrapper[4943]: I0307 14:41:15.599755 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 07 14:41:15 crc kubenswrapper[4943]: I0307 14:41:15.599831 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 07 14:41:15 crc kubenswrapper[4943]: E0307 14:41:15.600059 4943 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 07 14:41:15 crc kubenswrapper[4943]: E0307 14:41:15.600118 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-07 14:41:47.60010231 +0000 UTC m=+149.552238838 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 07 14:41:15 crc kubenswrapper[4943]: E0307 14:41:15.600456 4943 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 07 14:41:15 crc kubenswrapper[4943]: E0307 14:41:15.600569 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-07 14:41:47.600540752 +0000 UTC m=+149.552677290 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 07 14:41:15 crc kubenswrapper[4943]: I0307 14:41:15.702515 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 07 14:41:15 crc kubenswrapper[4943]: I0307 14:41:15.702633 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 07 14:41:15 crc kubenswrapper[4943]: E0307 14:41:15.702912 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 07 14:41:15 crc kubenswrapper[4943]: E0307 14:41:15.702986 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 07 14:41:15 crc kubenswrapper[4943]: E0307 14:41:15.703012 4943 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 07 14:41:15 crc kubenswrapper[4943]: E0307 14:41:15.703095 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-07 14:41:47.703066902 +0000 UTC m=+149.655203440 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 07 14:41:15 crc kubenswrapper[4943]: E0307 14:41:15.703731 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 07 14:41:15 crc kubenswrapper[4943]: E0307 14:41:15.703755 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 07 14:41:15 crc kubenswrapper[4943]: E0307 14:41:15.703772 4943 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 07 14:41:15 crc kubenswrapper[4943]: E0307 14:41:15.703820 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-07 14:41:47.703802751 +0000 UTC m=+149.655939289 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 07 14:41:15 crc kubenswrapper[4943]: I0307 14:41:15.754685 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 07 14:41:15 crc kubenswrapper[4943]: I0307 14:41:15.754784 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 07 14:41:15 crc kubenswrapper[4943]: E0307 14:41:15.754871 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 07 14:41:15 crc kubenswrapper[4943]: E0307 14:41:15.754962 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 07 14:41:15 crc kubenswrapper[4943]: I0307 14:41:15.755007 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 07 14:41:15 crc kubenswrapper[4943]: E0307 14:41:15.755264 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 07 14:41:17 crc kubenswrapper[4943]: I0307 14:41:17.754698 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 07 14:41:17 crc kubenswrapper[4943]: I0307 14:41:17.754804 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 07 14:41:17 crc kubenswrapper[4943]: I0307 14:41:17.754821 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ntc42" Mar 07 14:41:17 crc kubenswrapper[4943]: I0307 14:41:17.754870 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 07 14:41:17 crc kubenswrapper[4943]: E0307 14:41:17.755344 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 07 14:41:17 crc kubenswrapper[4943]: E0307 14:41:17.755579 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 07 14:41:17 crc kubenswrapper[4943]: E0307 14:41:17.755719 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 07 14:41:17 crc kubenswrapper[4943]: E0307 14:41:17.755915 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ntc42" podUID="e6f60f83-1b2e-4f75-9065-9971654e878a" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.300123 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.300333 4943 kubelet_node_status.go:538] "Fast updating node status as it just became ready" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.354797 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-bk5xc"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.355534 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-w5mcz"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.356267 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-w5mcz" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.356798 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-bk5xc" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.364901 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.365049 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.365914 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.366233 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.366509 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.368030 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-hr5px"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.375188 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hr5px" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.380401 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.380453 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.382219 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-n2mhr"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.383287 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.383586 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-n2mhr" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.383738 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.384117 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.384762 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.385056 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.385427 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.385493 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.385823 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.386071 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.386154 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.386257 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.391019 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.391099 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.391582 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.391750 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.392062 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.393358 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-6qmrb"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.394431 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6qmrb" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.394486 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.394815 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.423254 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-qml2f"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.423774 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-ltd8c"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.424065 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-cnbpx"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.424207 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qml2f" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.424418 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.424749 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-ltd8c" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.426837 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.427066 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.427194 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.427392 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.427505 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.427612 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.427716 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.427512 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.427611 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.428231 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.428507 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.428788 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.433045 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.433338 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.433564 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.433590 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.433659 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.433742 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.433847 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.433858 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.434171 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.434405 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.438868 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.444090 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dr8bw"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.444676 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dvjg6"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.445021 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-8jjq4"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.445372 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.445418 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dr8bw" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.445456 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.445614 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7k5c\" (UniqueName: \"kubernetes.io/projected/c400fdf2-390c-4c48-8c3e-4854d2e0b58f-kube-api-access-d7k5c\") pod \"machine-api-operator-5694c8668f-bk5xc\" (UID: \"c400fdf2-390c-4c48-8c3e-4854d2e0b58f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bk5xc" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.445652 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/dc37c9ce-2d22-4a80-be9b-0428621f7ca5-etcd-client\") pod \"apiserver-7bbb656c7d-hr5px\" (UID: \"dc37c9ce-2d22-4a80-be9b-0428621f7ca5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hr5px" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.445672 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4d86f7af-32df-419e-a833-d596c2a79ba7-serving-cert\") pod \"apiserver-76f77b778f-w5mcz\" (UID: \"4d86f7af-32df-419e-a833-d596c2a79ba7\") " pod="openshift-apiserver/apiserver-76f77b778f-w5mcz" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.445687 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/4d86f7af-32df-419e-a833-d596c2a79ba7-encryption-config\") pod \"apiserver-76f77b778f-w5mcz\" (UID: \"4d86f7af-32df-419e-a833-d596c2a79ba7\") " pod="openshift-apiserver/apiserver-76f77b778f-w5mcz" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.445707 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c400fdf2-390c-4c48-8c3e-4854d2e0b58f-config\") pod \"machine-api-operator-5694c8668f-bk5xc\" (UID: \"c400fdf2-390c-4c48-8c3e-4854d2e0b58f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bk5xc" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.445729 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/4d86f7af-32df-419e-a833-d596c2a79ba7-image-import-ca\") pod \"apiserver-76f77b778f-w5mcz\" (UID: \"4d86f7af-32df-419e-a833-d596c2a79ba7\") " pod="openshift-apiserver/apiserver-76f77b778f-w5mcz" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.445743 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/dc37c9ce-2d22-4a80-be9b-0428621f7ca5-audit-policies\") pod \"apiserver-7bbb656c7d-hr5px\" (UID: \"dc37c9ce-2d22-4a80-be9b-0428621f7ca5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hr5px" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.445758 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4d86f7af-32df-419e-a833-d596c2a79ba7-audit-dir\") pod \"apiserver-76f77b778f-w5mcz\" (UID: \"4d86f7af-32df-419e-a833-d596c2a79ba7\") " pod="openshift-apiserver/apiserver-76f77b778f-w5mcz" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.445789 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/dc37c9ce-2d22-4a80-be9b-0428621f7ca5-encryption-config\") pod \"apiserver-7bbb656c7d-hr5px\" (UID: \"dc37c9ce-2d22-4a80-be9b-0428621f7ca5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hr5px" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.445382 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-8jjq4" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.445915 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.446105 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dvjg6" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.445951 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.446730 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-zgx7p"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.445958 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.446158 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.445810 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/4d86f7af-32df-419e-a833-d596c2a79ba7-node-pullsecrets\") pod \"apiserver-76f77b778f-w5mcz\" (UID: \"4d86f7af-32df-419e-a833-d596c2a79ba7\") " pod="openshift-apiserver/apiserver-76f77b778f-w5mcz" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.448989 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dc37c9ce-2d22-4a80-be9b-0428621f7ca5-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-hr5px\" (UID: \"dc37c9ce-2d22-4a80-be9b-0428621f7ca5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hr5px" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.449040 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/4d86f7af-32df-419e-a833-d596c2a79ba7-audit\") pod \"apiserver-76f77b778f-w5mcz\" (UID: \"4d86f7af-32df-419e-a833-d596c2a79ba7\") " pod="openshift-apiserver/apiserver-76f77b778f-w5mcz" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.449063 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/c400fdf2-390c-4c48-8c3e-4854d2e0b58f-images\") pod \"machine-api-operator-5694c8668f-bk5xc\" (UID: \"c400fdf2-390c-4c48-8c3e-4854d2e0b58f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bk5xc" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.449088 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/c400fdf2-390c-4c48-8c3e-4854d2e0b58f-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-bk5xc\" (UID: \"c400fdf2-390c-4c48-8c3e-4854d2e0b58f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bk5xc" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.449130 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d86f7af-32df-419e-a833-d596c2a79ba7-config\") pod \"apiserver-76f77b778f-w5mcz\" (UID: \"4d86f7af-32df-419e-a833-d596c2a79ba7\") " pod="openshift-apiserver/apiserver-76f77b778f-w5mcz" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.449172 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lg6bx\" (UniqueName: \"kubernetes.io/projected/dc37c9ce-2d22-4a80-be9b-0428621f7ca5-kube-api-access-lg6bx\") pod \"apiserver-7bbb656c7d-hr5px\" (UID: \"dc37c9ce-2d22-4a80-be9b-0428621f7ca5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hr5px" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.449200 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/4d86f7af-32df-419e-a833-d596c2a79ba7-etcd-serving-ca\") pod \"apiserver-76f77b778f-w5mcz\" (UID: \"4d86f7af-32df-419e-a833-d596c2a79ba7\") " pod="openshift-apiserver/apiserver-76f77b778f-w5mcz" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.449219 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/4d86f7af-32df-419e-a833-d596c2a79ba7-etcd-client\") pod \"apiserver-76f77b778f-w5mcz\" (UID: \"4d86f7af-32df-419e-a833-d596c2a79ba7\") " pod="openshift-apiserver/apiserver-76f77b778f-w5mcz" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.449227 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-zgx7p" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.449241 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nj5lj\" (UniqueName: \"kubernetes.io/projected/4d86f7af-32df-419e-a833-d596c2a79ba7-kube-api-access-nj5lj\") pod \"apiserver-76f77b778f-w5mcz\" (UID: \"4d86f7af-32df-419e-a833-d596c2a79ba7\") " pod="openshift-apiserver/apiserver-76f77b778f-w5mcz" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.449260 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dc37c9ce-2d22-4a80-be9b-0428621f7ca5-serving-cert\") pod \"apiserver-7bbb656c7d-hr5px\" (UID: \"dc37c9ce-2d22-4a80-be9b-0428621f7ca5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hr5px" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.449280 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/dc37c9ce-2d22-4a80-be9b-0428621f7ca5-audit-dir\") pod \"apiserver-7bbb656c7d-hr5px\" (UID: \"dc37c9ce-2d22-4a80-be9b-0428621f7ca5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hr5px" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.449304 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4d86f7af-32df-419e-a833-d596c2a79ba7-trusted-ca-bundle\") pod \"apiserver-76f77b778f-w5mcz\" (UID: \"4d86f7af-32df-419e-a833-d596c2a79ba7\") " pod="openshift-apiserver/apiserver-76f77b778f-w5mcz" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.449325 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/dc37c9ce-2d22-4a80-be9b-0428621f7ca5-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-hr5px\" (UID: \"dc37c9ce-2d22-4a80-be9b-0428621f7ca5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hr5px" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.449170 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-r82fw"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.449783 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-x5qjf"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.450082 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-x5qjf" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.450273 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.450602 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.455362 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.457800 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.458079 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.458839 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.459018 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.459095 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.459270 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.459324 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.459397 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.459434 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.459491 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.459521 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.459534 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.459015 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.459612 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.467858 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.468121 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.469483 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.470185 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.470433 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.470478 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.470701 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.470860 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.471084 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.471283 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.472488 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.474023 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.474157 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.474495 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-8x4b2"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.476064 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-4vkx6"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.476107 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.476962 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.477385 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.477659 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.477842 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.478195 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.478297 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8x4b2" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.478196 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.482262 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-lg6mk"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.485310 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.486215 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.486859 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.489183 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.489765 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.490672 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lk96z"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.490990 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9gzdv"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.491101 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lk96z" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.491128 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-lg6mk" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.491772 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-58n5g"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.492166 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-58n5g" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.492285 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.492336 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9gzdv" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.492714 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4vkx6" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.494263 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-nk422"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.494631 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-nk422" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.494853 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pv7b4"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.495679 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pv7b4" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.496406 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x8mpt"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.496898 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x8mpt" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.497453 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.497647 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.498852 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-j89vd"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.499301 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-7p4vl"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.499939 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sqsdb"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.499967 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-j89vd" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.500372 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-sqsdb" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.500496 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7p4vl" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.506046 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.506280 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.506465 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.506637 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-dqpk4"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.507183 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-bk5xc"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.507202 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d622z"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.507544 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d622z" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.507766 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-dqpk4" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.508504 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-hzqrp"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.509236 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hzqrp" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.510193 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jpq82"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.510555 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jpq82" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.513059 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-ztm29"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.513784 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-ztm29" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.516048 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-w5mcz"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.518398 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-n2mhr"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.519095 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.522577 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-q85kb"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.523502 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-q85kb" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.526061 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-l7g67"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.526846 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.533866 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29548230-zr6fd"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.534211 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-l7g67" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.548447 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-hr5px"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.548497 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-h2f4f"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.549239 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29548230-zr6fd" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.549514 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kmxzj"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.550088 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-h2f4f" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.550441 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-2lnfd"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.550997 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/4d86f7af-32df-419e-a833-d596c2a79ba7-image-import-ca\") pod \"apiserver-76f77b778f-w5mcz\" (UID: \"4d86f7af-32df-419e-a833-d596c2a79ba7\") " pod="openshift-apiserver/apiserver-76f77b778f-w5mcz" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.551038 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/dc37c9ce-2d22-4a80-be9b-0428621f7ca5-audit-policies\") pod \"apiserver-7bbb656c7d-hr5px\" (UID: \"dc37c9ce-2d22-4a80-be9b-0428621f7ca5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hr5px" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.551074 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4d86f7af-32df-419e-a833-d596c2a79ba7-audit-dir\") pod \"apiserver-76f77b778f-w5mcz\" (UID: \"4d86f7af-32df-419e-a833-d596c2a79ba7\") " pod="openshift-apiserver/apiserver-76f77b778f-w5mcz" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.551123 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/dc37c9ce-2d22-4a80-be9b-0428621f7ca5-encryption-config\") pod \"apiserver-7bbb656c7d-hr5px\" (UID: \"dc37c9ce-2d22-4a80-be9b-0428621f7ca5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hr5px" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.551174 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/4d86f7af-32df-419e-a833-d596c2a79ba7-node-pullsecrets\") pod \"apiserver-76f77b778f-w5mcz\" (UID: \"4d86f7af-32df-419e-a833-d596c2a79ba7\") " pod="openshift-apiserver/apiserver-76f77b778f-w5mcz" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.551223 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dc37c9ce-2d22-4a80-be9b-0428621f7ca5-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-hr5px\" (UID: \"dc37c9ce-2d22-4a80-be9b-0428621f7ca5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hr5px" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.551470 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/4d86f7af-32df-419e-a833-d596c2a79ba7-audit\") pod \"apiserver-76f77b778f-w5mcz\" (UID: \"4d86f7af-32df-419e-a833-d596c2a79ba7\") " pod="openshift-apiserver/apiserver-76f77b778f-w5mcz" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.551535 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/c400fdf2-390c-4c48-8c3e-4854d2e0b58f-images\") pod \"machine-api-operator-5694c8668f-bk5xc\" (UID: \"c400fdf2-390c-4c48-8c3e-4854d2e0b58f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bk5xc" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.551566 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/c400fdf2-390c-4c48-8c3e-4854d2e0b58f-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-bk5xc\" (UID: \"c400fdf2-390c-4c48-8c3e-4854d2e0b58f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bk5xc" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.551588 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d86f7af-32df-419e-a833-d596c2a79ba7-config\") pod \"apiserver-76f77b778f-w5mcz\" (UID: \"4d86f7af-32df-419e-a833-d596c2a79ba7\") " pod="openshift-apiserver/apiserver-76f77b778f-w5mcz" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.551658 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kmxzj" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.551888 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lg6bx\" (UniqueName: \"kubernetes.io/projected/dc37c9ce-2d22-4a80-be9b-0428621f7ca5-kube-api-access-lg6bx\") pod \"apiserver-7bbb656c7d-hr5px\" (UID: \"dc37c9ce-2d22-4a80-be9b-0428621f7ca5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hr5px" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.551935 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/4d86f7af-32df-419e-a833-d596c2a79ba7-etcd-serving-ca\") pod \"apiserver-76f77b778f-w5mcz\" (UID: \"4d86f7af-32df-419e-a833-d596c2a79ba7\") " pod="openshift-apiserver/apiserver-76f77b778f-w5mcz" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.552290 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-kdrbq"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.553039 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-2lnfd" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.553344 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/4d86f7af-32df-419e-a833-d596c2a79ba7-image-import-ca\") pod \"apiserver-76f77b778f-w5mcz\" (UID: \"4d86f7af-32df-419e-a833-d596c2a79ba7\") " pod="openshift-apiserver/apiserver-76f77b778f-w5mcz" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.553829 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d86f7af-32df-419e-a833-d596c2a79ba7-config\") pod \"apiserver-76f77b778f-w5mcz\" (UID: \"4d86f7af-32df-419e-a833-d596c2a79ba7\") " pod="openshift-apiserver/apiserver-76f77b778f-w5mcz" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.553840 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/dc37c9ce-2d22-4a80-be9b-0428621f7ca5-audit-policies\") pod \"apiserver-7bbb656c7d-hr5px\" (UID: \"dc37c9ce-2d22-4a80-be9b-0428621f7ca5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hr5px" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.553878 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nj5lj\" (UniqueName: \"kubernetes.io/projected/4d86f7af-32df-419e-a833-d596c2a79ba7-kube-api-access-nj5lj\") pod \"apiserver-76f77b778f-w5mcz\" (UID: \"4d86f7af-32df-419e-a833-d596c2a79ba7\") " pod="openshift-apiserver/apiserver-76f77b778f-w5mcz" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.553907 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/4d86f7af-32df-419e-a833-d596c2a79ba7-etcd-client\") pod \"apiserver-76f77b778f-w5mcz\" (UID: \"4d86f7af-32df-419e-a833-d596c2a79ba7\") " pod="openshift-apiserver/apiserver-76f77b778f-w5mcz" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.555213 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/4d86f7af-32df-419e-a833-d596c2a79ba7-node-pullsecrets\") pod \"apiserver-76f77b778f-w5mcz\" (UID: \"4d86f7af-32df-419e-a833-d596c2a79ba7\") " pod="openshift-apiserver/apiserver-76f77b778f-w5mcz" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.555539 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dc37c9ce-2d22-4a80-be9b-0428621f7ca5-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-hr5px\" (UID: \"dc37c9ce-2d22-4a80-be9b-0428621f7ca5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hr5px" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.552113 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/4d86f7af-32df-419e-a833-d596c2a79ba7-audit\") pod \"apiserver-76f77b778f-w5mcz\" (UID: \"4d86f7af-32df-419e-a833-d596c2a79ba7\") " pod="openshift-apiserver/apiserver-76f77b778f-w5mcz" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.555919 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/4d86f7af-32df-419e-a833-d596c2a79ba7-etcd-serving-ca\") pod \"apiserver-76f77b778f-w5mcz\" (UID: \"4d86f7af-32df-419e-a833-d596c2a79ba7\") " pod="openshift-apiserver/apiserver-76f77b778f-w5mcz" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.556281 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dr8bw"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.556366 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kdrbq" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.556592 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dc37c9ce-2d22-4a80-be9b-0428621f7ca5-serving-cert\") pod \"apiserver-7bbb656c7d-hr5px\" (UID: \"dc37c9ce-2d22-4a80-be9b-0428621f7ca5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hr5px" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.556614 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/dc37c9ce-2d22-4a80-be9b-0428621f7ca5-audit-dir\") pod \"apiserver-7bbb656c7d-hr5px\" (UID: \"dc37c9ce-2d22-4a80-be9b-0428621f7ca5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hr5px" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.556634 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4d86f7af-32df-419e-a833-d596c2a79ba7-trusted-ca-bundle\") pod \"apiserver-76f77b778f-w5mcz\" (UID: \"4d86f7af-32df-419e-a833-d596c2a79ba7\") " pod="openshift-apiserver/apiserver-76f77b778f-w5mcz" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.556657 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/dc37c9ce-2d22-4a80-be9b-0428621f7ca5-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-hr5px\" (UID: \"dc37c9ce-2d22-4a80-be9b-0428621f7ca5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hr5px" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.556678 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/dc37c9ce-2d22-4a80-be9b-0428621f7ca5-etcd-client\") pod \"apiserver-7bbb656c7d-hr5px\" (UID: \"dc37c9ce-2d22-4a80-be9b-0428621f7ca5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hr5px" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.556707 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7k5c\" (UniqueName: \"kubernetes.io/projected/c400fdf2-390c-4c48-8c3e-4854d2e0b58f-kube-api-access-d7k5c\") pod \"machine-api-operator-5694c8668f-bk5xc\" (UID: \"c400fdf2-390c-4c48-8c3e-4854d2e0b58f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bk5xc" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.556728 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4d86f7af-32df-419e-a833-d596c2a79ba7-serving-cert\") pod \"apiserver-76f77b778f-w5mcz\" (UID: \"4d86f7af-32df-419e-a833-d596c2a79ba7\") " pod="openshift-apiserver/apiserver-76f77b778f-w5mcz" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.556752 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/4d86f7af-32df-419e-a833-d596c2a79ba7-encryption-config\") pod \"apiserver-76f77b778f-w5mcz\" (UID: \"4d86f7af-32df-419e-a833-d596c2a79ba7\") " pod="openshift-apiserver/apiserver-76f77b778f-w5mcz" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.556784 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c400fdf2-390c-4c48-8c3e-4854d2e0b58f-config\") pod \"machine-api-operator-5694c8668f-bk5xc\" (UID: \"c400fdf2-390c-4c48-8c3e-4854d2e0b58f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bk5xc" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.564634 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/dc37c9ce-2d22-4a80-be9b-0428621f7ca5-encryption-config\") pod \"apiserver-7bbb656c7d-hr5px\" (UID: \"dc37c9ce-2d22-4a80-be9b-0428621f7ca5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hr5px" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.566178 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/dc37c9ce-2d22-4a80-be9b-0428621f7ca5-audit-dir\") pod \"apiserver-7bbb656c7d-hr5px\" (UID: \"dc37c9ce-2d22-4a80-be9b-0428621f7ca5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hr5px" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.566373 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-8jjq4"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.553881 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4d86f7af-32df-419e-a833-d596c2a79ba7-audit-dir\") pod \"apiserver-76f77b778f-w5mcz\" (UID: \"4d86f7af-32df-419e-a833-d596c2a79ba7\") " pod="openshift-apiserver/apiserver-76f77b778f-w5mcz" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.567729 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4d86f7af-32df-419e-a833-d596c2a79ba7-trusted-ca-bundle\") pod \"apiserver-76f77b778f-w5mcz\" (UID: \"4d86f7af-32df-419e-a833-d596c2a79ba7\") " pod="openshift-apiserver/apiserver-76f77b778f-w5mcz" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.569101 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/dc37c9ce-2d22-4a80-be9b-0428621f7ca5-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-hr5px\" (UID: \"dc37c9ce-2d22-4a80-be9b-0428621f7ca5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hr5px" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.569842 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c400fdf2-390c-4c48-8c3e-4854d2e0b58f-config\") pod \"machine-api-operator-5694c8668f-bk5xc\" (UID: \"c400fdf2-390c-4c48-8c3e-4854d2e0b58f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bk5xc" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.569893 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/c400fdf2-390c-4c48-8c3e-4854d2e0b58f-images\") pod \"machine-api-operator-5694c8668f-bk5xc\" (UID: \"c400fdf2-390c-4c48-8c3e-4854d2e0b58f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bk5xc" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.572396 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4d86f7af-32df-419e-a833-d596c2a79ba7-serving-cert\") pod \"apiserver-76f77b778f-w5mcz\" (UID: \"4d86f7af-32df-419e-a833-d596c2a79ba7\") " pod="openshift-apiserver/apiserver-76f77b778f-w5mcz" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.575071 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/4d86f7af-32df-419e-a833-d596c2a79ba7-encryption-config\") pod \"apiserver-76f77b778f-w5mcz\" (UID: \"4d86f7af-32df-419e-a833-d596c2a79ba7\") " pod="openshift-apiserver/apiserver-76f77b778f-w5mcz" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.576591 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/4d86f7af-32df-419e-a833-d596c2a79ba7-etcd-client\") pod \"apiserver-76f77b778f-w5mcz\" (UID: \"4d86f7af-32df-419e-a833-d596c2a79ba7\") " pod="openshift-apiserver/apiserver-76f77b778f-w5mcz" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.576679 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dvjg6"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.578275 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-x5qjf"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.580352 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dc37c9ce-2d22-4a80-be9b-0428621f7ca5-serving-cert\") pod \"apiserver-7bbb656c7d-hr5px\" (UID: \"dc37c9ce-2d22-4a80-be9b-0428621f7ca5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hr5px" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.580791 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/c400fdf2-390c-4c48-8c3e-4854d2e0b58f-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-bk5xc\" (UID: \"c400fdf2-390c-4c48-8c3e-4854d2e0b58f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bk5xc" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.593786 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/dc37c9ce-2d22-4a80-be9b-0428621f7ca5-etcd-client\") pod \"apiserver-7bbb656c7d-hr5px\" (UID: \"dc37c9ce-2d22-4a80-be9b-0428621f7ca5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hr5px" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.594596 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.594823 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.594984 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.595114 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pv7b4"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.609409 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.614557 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-4vkx6"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.614600 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-zgx7p"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.614610 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-j89vd"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.616731 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-7p7x5"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.618171 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-6qmrb"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.618280 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-7p7x5" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.620072 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.623674 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lk96z"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.625003 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-ltd8c"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.629216 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-cnbpx"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.633474 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-r82fw"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.635158 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-7p4vl"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.636275 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-q85kb"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.637567 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jpq82"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.639023 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-58n5g"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.639489 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x8mpt"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.641483 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-hzqrp"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.643309 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-8x4b2"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.644105 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.644574 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9gzdv"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.647211 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-lg6mk"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.648249 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d622z"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.649207 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-ztm29"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.650097 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-2lnfd"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.651397 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29548230-zr6fd"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.653952 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-dqpk4"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.653985 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-2rd7t"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.654613 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-dmkh8"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.659290 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.662083 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-h2f4f"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.662115 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sqsdb"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.662125 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-kdrbq"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.662137 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-l7g67"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.662209 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-dmkh8" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.662315 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-2rd7t" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.662936 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kmxzj"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.666957 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-2rd7t"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.668556 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-7p7x5"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.672217 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-htv6d"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.673021 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-htv6d" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.673905 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-w286n"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.674359 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-w286n" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.675509 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-w286n"] Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.681120 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.699235 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.718773 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.738707 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.759267 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.779139 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.799808 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.819503 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.839394 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.859393 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.880262 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.899848 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.920054 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.939606 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.960057 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.979818 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Mar 07 14:41:18 crc kubenswrapper[4943]: I0307 14:41:18.999806 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.019954 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.039052 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.059388 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.078957 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.099263 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.120554 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.138995 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.159497 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.179127 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.199651 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.230690 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.239352 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.260297 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.279436 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.299395 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.320114 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.359912 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.365572 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xm5zc\" (UniqueName: \"kubernetes.io/projected/0da2f3db-9ac4-42bf-b4b8-a4d538d9f4e6-kube-api-access-xm5zc\") pod \"openshift-apiserver-operator-796bbdcf4f-dvjg6\" (UID: \"0da2f3db-9ac4-42bf-b4b8-a4d538d9f4e6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dvjg6" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.365775 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f94ec039-60b3-4ecb-b784-04a7ebaf4848-config\") pod \"machine-approver-56656f9798-qml2f\" (UID: \"f94ec039-60b3-4ecb-b784-04a7ebaf4848\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qml2f" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.365987 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0da2f3db-9ac4-42bf-b4b8-a4d538d9f4e6-config\") pod \"openshift-apiserver-operator-796bbdcf4f-dvjg6\" (UID: \"0da2f3db-9ac4-42bf-b4b8-a4d538d9f4e6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dvjg6" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.366222 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/00cc0b59-be8e-402b-b6c9-56e9e20c4b9b-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-x5qjf\" (UID: \"00cc0b59-be8e-402b-b6c9-56e9e20c4b9b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-x5qjf" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.366376 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00cc0b59-be8e-402b-b6c9-56e9e20c4b9b-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-x5qjf\" (UID: \"00cc0b59-be8e-402b-b6c9-56e9e20c4b9b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-x5qjf" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.366533 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-cnbpx\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.366684 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b541cfee-3ec5-4be3-9acd-d58ebb79f0a2-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-ltd8c\" (UID: \"b541cfee-3ec5-4be3-9acd-d58ebb79f0a2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ltd8c" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.366886 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0da2f3db-9ac4-42bf-b4b8-a4d538d9f4e6-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-dvjg6\" (UID: \"0da2f3db-9ac4-42bf-b4b8-a4d538d9f4e6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dvjg6" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.367074 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a04d88f1-a308-42cf-9462-1a06f6338b7b-config\") pod \"controller-manager-879f6c89f-n2mhr\" (UID: \"a04d88f1-a308-42cf-9462-1a06f6338b7b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n2mhr" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.367228 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a04d88f1-a308-42cf-9462-1a06f6338b7b-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-n2mhr\" (UID: \"a04d88f1-a308-42cf-9462-1a06f6338b7b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n2mhr" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.367379 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b541cfee-3ec5-4be3-9acd-d58ebb79f0a2-service-ca-bundle\") pod \"authentication-operator-69f744f599-ltd8c\" (UID: \"b541cfee-3ec5-4be3-9acd-d58ebb79f0a2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ltd8c" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.367554 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pr5v\" (UniqueName: \"kubernetes.io/projected/31ce0220-0f53-4d63-aa7b-99357e9c6b04-kube-api-access-4pr5v\") pod \"oauth-openshift-558db77b4-cnbpx\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.367760 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4tmt\" (UniqueName: \"kubernetes.io/projected/93cd8321-367e-4f02-9f8e-2e044e0f1d89-kube-api-access-w4tmt\") pod \"downloads-7954f5f757-8jjq4\" (UID: \"93cd8321-367e-4f02-9f8e-2e044e0f1d89\") " pod="openshift-console/downloads-7954f5f757-8jjq4" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.368086 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b541cfee-3ec5-4be3-9acd-d58ebb79f0a2-serving-cert\") pod \"authentication-operator-69f744f599-ltd8c\" (UID: \"b541cfee-3ec5-4be3-9acd-d58ebb79f0a2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ltd8c" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.368173 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/80ba5362-78a1-4ed1-a6be-fbda0dbe72df-metrics-tls\") pod \"ingress-operator-5b745b69d9-8x4b2\" (UID: \"80ba5362-78a1-4ed1-a6be-fbda0dbe72df\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8x4b2" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.368214 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/80ba5362-78a1-4ed1-a6be-fbda0dbe72df-trusted-ca\") pod \"ingress-operator-5b745b69d9-8x4b2\" (UID: \"80ba5362-78a1-4ed1-a6be-fbda0dbe72df\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8x4b2" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.368249 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/80ba5362-78a1-4ed1-a6be-fbda0dbe72df-bound-sa-token\") pod \"ingress-operator-5b745b69d9-8x4b2\" (UID: \"80ba5362-78a1-4ed1-a6be-fbda0dbe72df\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8x4b2" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.368318 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a9fcbefd-dfa7-4afd-b198-d872137a9f51-installation-pull-secrets\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.368682 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51fb375a-53bd-420a-930f-6b6cdb69f793-config\") pod \"route-controller-manager-6576b87f9c-6qmrb\" (UID: \"51fb375a-53bd-420a-930f-6b6cdb69f793\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6qmrb" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.368911 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a9fcbefd-dfa7-4afd-b198-d872137a9f51-bound-sa-token\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.368997 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ww544\" (UniqueName: \"kubernetes.io/projected/00cc0b59-be8e-402b-b6c9-56e9e20c4b9b-kube-api-access-ww544\") pod \"openshift-controller-manager-operator-756b6f6bc6-x5qjf\" (UID: \"00cc0b59-be8e-402b-b6c9-56e9e20c4b9b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-x5qjf" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.369044 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/51fb375a-53bd-420a-930f-6b6cdb69f793-serving-cert\") pod \"route-controller-manager-6576b87f9c-6qmrb\" (UID: \"51fb375a-53bd-420a-930f-6b6cdb69f793\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6qmrb" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.369089 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7960fae-df40-482f-a9ab-790fa11aaf8f-config\") pod \"console-operator-58897d9998-lg6mk\" (UID: \"b7960fae-df40-482f-a9ab-790fa11aaf8f\") " pod="openshift-console-operator/console-operator-58897d9998-lg6mk" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.369135 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a9fcbefd-dfa7-4afd-b198-d872137a9f51-registry-certificates\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.369232 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t458n\" (UniqueName: \"kubernetes.io/projected/f94ec039-60b3-4ecb-b784-04a7ebaf4848-kube-api-access-t458n\") pod \"machine-approver-56656f9798-qml2f\" (UID: \"f94ec039-60b3-4ecb-b784-04a7ebaf4848\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qml2f" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.369538 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x46k8\" (UniqueName: \"kubernetes.io/projected/a04d88f1-a308-42cf-9462-1a06f6338b7b-kube-api-access-x46k8\") pod \"controller-manager-879f6c89f-n2mhr\" (UID: \"a04d88f1-a308-42cf-9462-1a06f6338b7b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n2mhr" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.369644 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/698827d2-21c9-4856-a9cd-17bb3df88bce-console-config\") pod \"console-f9d7485db-zgx7p\" (UID: \"698827d2-21c9-4856-a9cd-17bb3df88bce\") " pod="openshift-console/console-f9d7485db-zgx7p" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.369694 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/08f9e92f-8419-4ba3-a532-c13e5bb6cab4-available-featuregates\") pod \"openshift-config-operator-7777fb866f-4vkx6\" (UID: \"08f9e92f-8419-4ba3-a532-c13e5bb6cab4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4vkx6" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.369778 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.369834 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/f94ec039-60b3-4ecb-b784-04a7ebaf4848-machine-approver-tls\") pod \"machine-approver-56656f9798-qml2f\" (UID: \"f94ec039-60b3-4ecb-b784-04a7ebaf4848\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qml2f" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.369884 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-cnbpx\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.369965 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b7960fae-df40-482f-a9ab-790fa11aaf8f-trusted-ca\") pod \"console-operator-58897d9998-lg6mk\" (UID: \"b7960fae-df40-482f-a9ab-790fa11aaf8f\") " pod="openshift-console-operator/console-operator-58897d9998-lg6mk" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.370063 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-cnbpx\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.370158 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qx4fv\" (UniqueName: \"kubernetes.io/projected/51fb375a-53bd-420a-930f-6b6cdb69f793-kube-api-access-qx4fv\") pod \"route-controller-manager-6576b87f9c-6qmrb\" (UID: \"51fb375a-53bd-420a-930f-6b6cdb69f793\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6qmrb" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.370191 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/28995414-2bd3-4b50-90ec-0fbb63a15ef6-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-dr8bw\" (UID: \"28995414-2bd3-4b50-90ec-0fbb63a15ef6\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dr8bw" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.370226 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lmxv\" (UniqueName: \"kubernetes.io/projected/698827d2-21c9-4856-a9cd-17bb3df88bce-kube-api-access-9lmxv\") pod \"console-f9d7485db-zgx7p\" (UID: \"698827d2-21c9-4856-a9cd-17bb3df88bce\") " pod="openshift-console/console-f9d7485db-zgx7p" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.370263 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lh8b9\" (UniqueName: \"kubernetes.io/projected/a9fcbefd-dfa7-4afd-b198-d872137a9f51-kube-api-access-lh8b9\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.370298 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-cnbpx\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.370332 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b541cfee-3ec5-4be3-9acd-d58ebb79f0a2-config\") pod \"authentication-operator-69f744f599-ltd8c\" (UID: \"b541cfee-3ec5-4be3-9acd-d58ebb79f0a2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ltd8c" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.370465 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/698827d2-21c9-4856-a9cd-17bb3df88bce-console-oauth-config\") pod \"console-f9d7485db-zgx7p\" (UID: \"698827d2-21c9-4856-a9cd-17bb3df88bce\") " pod="openshift-console/console-f9d7485db-zgx7p" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.370551 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6pzz\" (UniqueName: \"kubernetes.io/projected/80ba5362-78a1-4ed1-a6be-fbda0dbe72df-kube-api-access-x6pzz\") pod \"ingress-operator-5b745b69d9-8x4b2\" (UID: \"80ba5362-78a1-4ed1-a6be-fbda0dbe72df\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8x4b2" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.370600 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/08f9e92f-8419-4ba3-a532-c13e5bb6cab4-serving-cert\") pod \"openshift-config-operator-7777fb866f-4vkx6\" (UID: \"08f9e92f-8419-4ba3-a532-c13e5bb6cab4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4vkx6" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.370648 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/698827d2-21c9-4856-a9cd-17bb3df88bce-trusted-ca-bundle\") pod \"console-f9d7485db-zgx7p\" (UID: \"698827d2-21c9-4856-a9cd-17bb3df88bce\") " pod="openshift-console/console-f9d7485db-zgx7p" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.370780 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a9fcbefd-dfa7-4afd-b198-d872137a9f51-trusted-ca\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:19 crc kubenswrapper[4943]: E0307 14:41:19.370834 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:19.870802235 +0000 UTC m=+121.822938903 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.370914 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-cnbpx\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.371007 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-cnbpx\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.371037 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/698827d2-21c9-4856-a9cd-17bb3df88bce-service-ca\") pod \"console-f9d7485db-zgx7p\" (UID: \"698827d2-21c9-4856-a9cd-17bb3df88bce\") " pod="openshift-console/console-f9d7485db-zgx7p" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.371077 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f94ec039-60b3-4ecb-b784-04a7ebaf4848-auth-proxy-config\") pod \"machine-approver-56656f9798-qml2f\" (UID: \"f94ec039-60b3-4ecb-b784-04a7ebaf4848\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qml2f" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.371116 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/31ce0220-0f53-4d63-aa7b-99357e9c6b04-audit-dir\") pod \"oauth-openshift-558db77b4-cnbpx\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.371209 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/698827d2-21c9-4856-a9cd-17bb3df88bce-oauth-serving-cert\") pod \"console-f9d7485db-zgx7p\" (UID: \"698827d2-21c9-4856-a9cd-17bb3df88bce\") " pod="openshift-console/console-f9d7485db-zgx7p" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.371265 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/51fb375a-53bd-420a-930f-6b6cdb69f793-client-ca\") pod \"route-controller-manager-6576b87f9c-6qmrb\" (UID: \"51fb375a-53bd-420a-930f-6b6cdb69f793\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6qmrb" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.371307 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-cnbpx\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.371344 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvr2j\" (UniqueName: \"kubernetes.io/projected/08f9e92f-8419-4ba3-a532-c13e5bb6cab4-kube-api-access-fvr2j\") pod \"openshift-config-operator-7777fb866f-4vkx6\" (UID: \"08f9e92f-8419-4ba3-a532-c13e5bb6cab4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4vkx6" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.371378 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b7960fae-df40-482f-a9ab-790fa11aaf8f-serving-cert\") pod \"console-operator-58897d9998-lg6mk\" (UID: \"b7960fae-df40-482f-a9ab-790fa11aaf8f\") " pod="openshift-console-operator/console-operator-58897d9998-lg6mk" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.371432 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-cnbpx\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.371484 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hffv9\" (UniqueName: \"kubernetes.io/projected/b7960fae-df40-482f-a9ab-790fa11aaf8f-kube-api-access-hffv9\") pod \"console-operator-58897d9998-lg6mk\" (UID: \"b7960fae-df40-482f-a9ab-790fa11aaf8f\") " pod="openshift-console-operator/console-operator-58897d9998-lg6mk" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.371689 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bw8vf\" (UniqueName: \"kubernetes.io/projected/28995414-2bd3-4b50-90ec-0fbb63a15ef6-kube-api-access-bw8vf\") pod \"cluster-samples-operator-665b6dd947-dr8bw\" (UID: \"28995414-2bd3-4b50-90ec-0fbb63a15ef6\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dr8bw" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.371818 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-cnbpx\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.371952 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a9fcbefd-dfa7-4afd-b198-d872137a9f51-registry-tls\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.371992 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/31ce0220-0f53-4d63-aa7b-99357e9c6b04-audit-policies\") pod \"oauth-openshift-558db77b4-cnbpx\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.372019 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-cnbpx\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.372048 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a04d88f1-a308-42cf-9462-1a06f6338b7b-serving-cert\") pod \"controller-manager-879f6c89f-n2mhr\" (UID: \"a04d88f1-a308-42cf-9462-1a06f6338b7b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n2mhr" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.372074 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-cnbpx\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.372100 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/698827d2-21c9-4856-a9cd-17bb3df88bce-console-serving-cert\") pod \"console-f9d7485db-zgx7p\" (UID: \"698827d2-21c9-4856-a9cd-17bb3df88bce\") " pod="openshift-console/console-f9d7485db-zgx7p" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.372136 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a9fcbefd-dfa7-4afd-b198-d872137a9f51-ca-trust-extracted\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.372160 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a04d88f1-a308-42cf-9462-1a06f6338b7b-client-ca\") pod \"controller-manager-879f6c89f-n2mhr\" (UID: \"a04d88f1-a308-42cf-9462-1a06f6338b7b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n2mhr" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.372311 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m69bh\" (UniqueName: \"kubernetes.io/projected/b541cfee-3ec5-4be3-9acd-d58ebb79f0a2-kube-api-access-m69bh\") pod \"authentication-operator-69f744f599-ltd8c\" (UID: \"b541cfee-3ec5-4be3-9acd-d58ebb79f0a2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ltd8c" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.379067 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.398860 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.419272 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.439961 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.460108 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.473699 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.473880 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qx4fv\" (UniqueName: \"kubernetes.io/projected/51fb375a-53bd-420a-930f-6b6cdb69f793-kube-api-access-qx4fv\") pod \"route-controller-manager-6576b87f9c-6qmrb\" (UID: \"51fb375a-53bd-420a-930f-6b6cdb69f793\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6qmrb" Mar 07 14:41:19 crc kubenswrapper[4943]: E0307 14:41:19.473956 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:19.973895009 +0000 UTC m=+121.926031547 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.474023 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/28995414-2bd3-4b50-90ec-0fbb63a15ef6-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-dr8bw\" (UID: \"28995414-2bd3-4b50-90ec-0fbb63a15ef6\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dr8bw" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.474076 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lmxv\" (UniqueName: \"kubernetes.io/projected/698827d2-21c9-4856-a9cd-17bb3df88bce-kube-api-access-9lmxv\") pod \"console-f9d7485db-zgx7p\" (UID: \"698827d2-21c9-4856-a9cd-17bb3df88bce\") " pod="openshift-console/console-f9d7485db-zgx7p" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.474120 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7ca6333f-d5ff-4c06-a140-3b9c29710528-metrics-certs\") pod \"router-default-5444994796-nk422\" (UID: \"7ca6333f-d5ff-4c06-a140-3b9c29710528\") " pod="openshift-ingress/router-default-5444994796-nk422" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.474159 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lh8b9\" (UniqueName: \"kubernetes.io/projected/a9fcbefd-dfa7-4afd-b198-d872137a9f51-kube-api-access-lh8b9\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.474195 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/8e19594a-0a69-40b7-8c24-317979edbfbb-srv-cert\") pod \"catalog-operator-68c6474976-jpq82\" (UID: \"8e19594a-0a69-40b7-8c24-317979edbfbb\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jpq82" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.474233 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/203d975c-92a2-424e-a6c4-e4904440dead-tmpfs\") pod \"packageserver-d55dfcdfc-q85kb\" (UID: \"203d975c-92a2-424e-a6c4-e4904440dead\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-q85kb" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.474271 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/08f9e92f-8419-4ba3-a532-c13e5bb6cab4-serving-cert\") pod \"openshift-config-operator-7777fb866f-4vkx6\" (UID: \"08f9e92f-8419-4ba3-a532-c13e5bb6cab4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4vkx6" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.474309 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/698827d2-21c9-4856-a9cd-17bb3df88bce-trusted-ca-bundle\") pod \"console-f9d7485db-zgx7p\" (UID: \"698827d2-21c9-4856-a9cd-17bb3df88bce\") " pod="openshift-console/console-f9d7485db-zgx7p" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.474346 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6wjd\" (UniqueName: \"kubernetes.io/projected/df59699d-5df8-401e-8bb0-0afa130ac06b-kube-api-access-p6wjd\") pod \"cluster-image-registry-operator-dc59b4c8b-lk96z\" (UID: \"df59699d-5df8-401e-8bb0-0afa130ac06b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lk96z" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.474380 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b33c7fd7-2d0f-48b7-9932-e7740577280f-config\") pod \"kube-apiserver-operator-766d6c64bb-9gzdv\" (UID: \"b33c7fd7-2d0f-48b7-9932-e7740577280f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9gzdv" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.474442 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a9fcbefd-dfa7-4afd-b198-d872137a9f51-trusted-ca\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.474478 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-cnbpx\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.474510 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2m79r\" (UniqueName: \"kubernetes.io/projected/415baeb7-0937-47e1-9d7e-8c2078911c3c-kube-api-access-2m79r\") pod \"kube-storage-version-migrator-operator-b67b599dd-d622z\" (UID: \"415baeb7-0937-47e1-9d7e-8c2078911c3c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d622z" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.474541 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/030c2b00-0d20-447c-b7a5-55d86f1674d8-serving-cert\") pod \"etcd-operator-b45778765-dqpk4\" (UID: \"030c2b00-0d20-447c-b7a5-55d86f1674d8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dqpk4" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.474729 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-cnbpx\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.474792 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/618e4221-7bc8-4721-ac10-1a1373d0fade-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-x8mpt\" (UID: \"618e4221-7bc8-4721-ac10-1a1373d0fade\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x8mpt" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.474829 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/ff8cf71a-e914-4244-b1aa-ede95edd30e8-socket-dir\") pod \"csi-hostpathplugin-7p7x5\" (UID: \"ff8cf71a-e914-4244-b1aa-ede95edd30e8\") " pod="hostpath-provisioner/csi-hostpathplugin-7p7x5" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.474887 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b33c7fd7-2d0f-48b7-9932-e7740577280f-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-9gzdv\" (UID: \"b33c7fd7-2d0f-48b7-9932-e7740577280f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9gzdv" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.474922 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/698827d2-21c9-4856-a9cd-17bb3df88bce-oauth-serving-cert\") pod \"console-f9d7485db-zgx7p\" (UID: \"698827d2-21c9-4856-a9cd-17bb3df88bce\") " pod="openshift-console/console-f9d7485db-zgx7p" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.475104 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b7960fae-df40-482f-a9ab-790fa11aaf8f-serving-cert\") pod \"console-operator-58897d9998-lg6mk\" (UID: \"b7960fae-df40-482f-a9ab-790fa11aaf8f\") " pod="openshift-console-operator/console-operator-58897d9998-lg6mk" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.475163 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/df59699d-5df8-401e-8bb0-0afa130ac06b-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-lk96z\" (UID: \"df59699d-5df8-401e-8bb0-0afa130ac06b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lk96z" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.475204 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-cnbpx\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.475240 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hffv9\" (UniqueName: \"kubernetes.io/projected/b7960fae-df40-482f-a9ab-790fa11aaf8f-kube-api-access-hffv9\") pod \"console-operator-58897d9998-lg6mk\" (UID: \"b7960fae-df40-482f-a9ab-790fa11aaf8f\") " pod="openshift-console-operator/console-operator-58897d9998-lg6mk" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.475280 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bw8vf\" (UniqueName: \"kubernetes.io/projected/28995414-2bd3-4b50-90ec-0fbb63a15ef6-kube-api-access-bw8vf\") pod \"cluster-samples-operator-665b6dd947-dr8bw\" (UID: \"28995414-2bd3-4b50-90ec-0fbb63a15ef6\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dr8bw" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.475314 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e03f48ff-595f-4537-824f-650401604f7e-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-h2f4f\" (UID: \"e03f48ff-595f-4537-824f-650401604f7e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-h2f4f" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.475355 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a9fcbefd-dfa7-4afd-b198-d872137a9f51-registry-tls\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.475392 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-cnbpx\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.475424 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a04d88f1-a308-42cf-9462-1a06f6338b7b-serving-cert\") pod \"controller-manager-879f6c89f-n2mhr\" (UID: \"a04d88f1-a308-42cf-9462-1a06f6338b7b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n2mhr" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.475458 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e8c3aeaf-47bf-48ab-b029-2f19e51dae4a-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-hzqrp\" (UID: \"e8c3aeaf-47bf-48ab-b029-2f19e51dae4a\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hzqrp" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.475497 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/698827d2-21c9-4856-a9cd-17bb3df88bce-console-serving-cert\") pod \"console-f9d7485db-zgx7p\" (UID: \"698827d2-21c9-4856-a9cd-17bb3df88bce\") " pod="openshift-console/console-f9d7485db-zgx7p" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.475531 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a9fcbefd-dfa7-4afd-b198-d872137a9f51-ca-trust-extracted\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.475565 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m69bh\" (UniqueName: \"kubernetes.io/projected/b541cfee-3ec5-4be3-9acd-d58ebb79f0a2-kube-api-access-m69bh\") pod \"authentication-operator-69f744f599-ltd8c\" (UID: \"b541cfee-3ec5-4be3-9acd-d58ebb79f0a2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ltd8c" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.475600 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xm5zc\" (UniqueName: \"kubernetes.io/projected/0da2f3db-9ac4-42bf-b4b8-a4d538d9f4e6-kube-api-access-xm5zc\") pod \"openshift-apiserver-operator-796bbdcf4f-dvjg6\" (UID: \"0da2f3db-9ac4-42bf-b4b8-a4d538d9f4e6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dvjg6" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.475632 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f94ec039-60b3-4ecb-b784-04a7ebaf4848-config\") pod \"machine-approver-56656f9798-qml2f\" (UID: \"f94ec039-60b3-4ecb-b784-04a7ebaf4848\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qml2f" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.475663 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqhfd\" (UniqueName: \"kubernetes.io/projected/8e19594a-0a69-40b7-8c24-317979edbfbb-kube-api-access-sqhfd\") pod \"catalog-operator-68c6474976-jpq82\" (UID: \"8e19594a-0a69-40b7-8c24-317979edbfbb\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jpq82" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.475697 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/40cb1361-4bb1-450f-8d78-1a033e611f04-signing-key\") pod \"service-ca-9c57cc56f-2lnfd\" (UID: \"40cb1361-4bb1-450f-8d78-1a033e611f04\") " pod="openshift-service-ca/service-ca-9c57cc56f-2lnfd" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.475729 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgzhw\" (UniqueName: \"kubernetes.io/projected/0de7b0c4-4acf-40de-867d-1fc33069fa51-kube-api-access-dgzhw\") pod \"cni-sysctl-allowlist-ds-htv6d\" (UID: \"0de7b0c4-4acf-40de-867d-1fc33069fa51\") " pod="openshift-multus/cni-sysctl-allowlist-ds-htv6d" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.475763 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjwms\" (UniqueName: \"kubernetes.io/projected/2a45a1ae-558f-4600-b354-dea0169b175f-kube-api-access-gjwms\") pod \"dns-operator-744455d44c-j89vd\" (UID: \"2a45a1ae-558f-4600-b354-dea0169b175f\") " pod="openshift-dns-operator/dns-operator-744455d44c-j89vd" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.475826 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/00cc0b59-be8e-402b-b6c9-56e9e20c4b9b-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-x5qjf\" (UID: \"00cc0b59-be8e-402b-b6c9-56e9e20c4b9b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-x5qjf" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.475870 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00cc0b59-be8e-402b-b6c9-56e9e20c4b9b-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-x5qjf\" (UID: \"00cc0b59-be8e-402b-b6c9-56e9e20c4b9b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-x5qjf" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.475907 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0da2f3db-9ac4-42bf-b4b8-a4d538d9f4e6-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-dvjg6\" (UID: \"0da2f3db-9ac4-42bf-b4b8-a4d538d9f4e6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dvjg6" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.475967 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a04d88f1-a308-42cf-9462-1a06f6338b7b-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-n2mhr\" (UID: \"a04d88f1-a308-42cf-9462-1a06f6338b7b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n2mhr" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.476000 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b541cfee-3ec5-4be3-9acd-d58ebb79f0a2-service-ca-bundle\") pod \"authentication-operator-69f744f599-ltd8c\" (UID: \"b541cfee-3ec5-4be3-9acd-d58ebb79f0a2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ltd8c" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.476037 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/8e19594a-0a69-40b7-8c24-317979edbfbb-profile-collector-cert\") pod \"catalog-operator-68c6474976-jpq82\" (UID: \"8e19594a-0a69-40b7-8c24-317979edbfbb\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jpq82" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.476104 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/415baeb7-0937-47e1-9d7e-8c2078911c3c-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-d622z\" (UID: \"415baeb7-0937-47e1-9d7e-8c2078911c3c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d622z" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.476138 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/0de7b0c4-4acf-40de-867d-1fc33069fa51-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-htv6d\" (UID: \"0de7b0c4-4acf-40de-867d-1fc33069fa51\") " pod="openshift-multus/cni-sysctl-allowlist-ds-htv6d" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.476177 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pr5v\" (UniqueName: \"kubernetes.io/projected/31ce0220-0f53-4d63-aa7b-99357e9c6b04-kube-api-access-4pr5v\") pod \"oauth-openshift-558db77b4-cnbpx\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.476215 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/839b7c38-e05d-4e85-942a-1936e3b2bcf7-certs\") pod \"machine-config-server-dmkh8\" (UID: \"839b7c38-e05d-4e85-942a-1936e3b2bcf7\") " pod="openshift-machine-config-operator/machine-config-server-dmkh8" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.476254 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4tmt\" (UniqueName: \"kubernetes.io/projected/93cd8321-367e-4f02-9f8e-2e044e0f1d89-kube-api-access-w4tmt\") pod \"downloads-7954f5f757-8jjq4\" (UID: \"93cd8321-367e-4f02-9f8e-2e044e0f1d89\") " pod="openshift-console/downloads-7954f5f757-8jjq4" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.476261 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a9fcbefd-dfa7-4afd-b198-d872137a9f51-ca-trust-extracted\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.476397 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b541cfee-3ec5-4be3-9acd-d58ebb79f0a2-serving-cert\") pod \"authentication-operator-69f744f599-ltd8c\" (UID: \"b541cfee-3ec5-4be3-9acd-d58ebb79f0a2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ltd8c" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.476535 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b6ab22d-b401-4727-83f2-94b527c4a366-config\") pod \"service-ca-operator-777779d784-ztm29\" (UID: \"0b6ab22d-b401-4727-83f2-94b527c4a366\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ztm29" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.476615 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x64d9\" (UniqueName: \"kubernetes.io/projected/8d97c7b9-9de3-40a6-ba65-db9c730604c0-kube-api-access-x64d9\") pod \"machine-config-operator-74547568cd-kdrbq\" (UID: \"8d97c7b9-9de3-40a6-ba65-db9c730604c0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kdrbq" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.476669 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/712a2a31-4205-4346-9a32-858a77615eb6-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-sqsdb\" (UID: \"712a2a31-4205-4346-9a32-858a77615eb6\") " pod="openshift-marketplace/marketplace-operator-79b997595-sqsdb" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.476713 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e03f48ff-595f-4537-824f-650401604f7e-config\") pod \"kube-controller-manager-operator-78b949d7b-h2f4f\" (UID: \"e03f48ff-595f-4537-824f-650401604f7e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-h2f4f" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.476828 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/80ba5362-78a1-4ed1-a6be-fbda0dbe72df-metrics-tls\") pod \"ingress-operator-5b745b69d9-8x4b2\" (UID: \"80ba5362-78a1-4ed1-a6be-fbda0dbe72df\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8x4b2" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.476878 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/80ba5362-78a1-4ed1-a6be-fbda0dbe72df-trusted-ca\") pod \"ingress-operator-5b745b69d9-8x4b2\" (UID: \"80ba5362-78a1-4ed1-a6be-fbda0dbe72df\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8x4b2" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.476923 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/80ba5362-78a1-4ed1-a6be-fbda0dbe72df-bound-sa-token\") pod \"ingress-operator-5b745b69d9-8x4b2\" (UID: \"80ba5362-78a1-4ed1-a6be-fbda0dbe72df\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8x4b2" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.477039 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/fdcfa33a-5367-46db-8270-662184116079-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-pv7b4\" (UID: \"fdcfa33a-5367-46db-8270-662184116079\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pv7b4" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.477078 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/0f3dc501-2bc1-4930-a644-f020c92c2827-profile-collector-cert\") pod \"olm-operator-6b444d44fb-kmxzj\" (UID: \"0f3dc501-2bc1-4930-a644-f020c92c2827\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kmxzj" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.477155 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/030c2b00-0d20-447c-b7a5-55d86f1674d8-etcd-client\") pod \"etcd-operator-b45778765-dqpk4\" (UID: \"030c2b00-0d20-447c-b7a5-55d86f1674d8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dqpk4" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.477201 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tktfx\" (UniqueName: \"kubernetes.io/projected/f7170a81-b6f1-4696-9356-737721576cf6-kube-api-access-tktfx\") pod \"migrator-59844c95c7-7p4vl\" (UID: \"f7170a81-b6f1-4696-9356-737721576cf6\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7p4vl" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.477273 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51fb375a-53bd-420a-930f-6b6cdb69f793-config\") pod \"route-controller-manager-6576b87f9c-6qmrb\" (UID: \"51fb375a-53bd-420a-930f-6b6cdb69f793\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6qmrb" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.477378 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-scbtq\" (UniqueName: \"kubernetes.io/projected/fdcfa33a-5367-46db-8270-662184116079-kube-api-access-scbtq\") pod \"control-plane-machine-set-operator-78cbb6b69f-pv7b4\" (UID: \"fdcfa33a-5367-46db-8270-662184116079\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pv7b4" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.477415 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e8c3aeaf-47bf-48ab-b029-2f19e51dae4a-proxy-tls\") pod \"machine-config-controller-84d6567774-hzqrp\" (UID: \"e8c3aeaf-47bf-48ab-b029-2f19e51dae4a\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hzqrp" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.477446 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/23baba8c-4f55-4188-8967-61093c57e913-metrics-tls\") pod \"dns-default-2rd7t\" (UID: \"23baba8c-4f55-4188-8967-61093c57e913\") " pod="openshift-dns/dns-default-2rd7t" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.477494 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ww544\" (UniqueName: \"kubernetes.io/projected/00cc0b59-be8e-402b-b6c9-56e9e20c4b9b-kube-api-access-ww544\") pod \"openshift-controller-manager-operator-756b6f6bc6-x5qjf\" (UID: \"00cc0b59-be8e-402b-b6c9-56e9e20c4b9b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-x5qjf" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.477526 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/51fb375a-53bd-420a-930f-6b6cdb69f793-serving-cert\") pod \"route-controller-manager-6576b87f9c-6qmrb\" (UID: \"51fb375a-53bd-420a-930f-6b6cdb69f793\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6qmrb" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.477560 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7960fae-df40-482f-a9ab-790fa11aaf8f-config\") pod \"console-operator-58897d9998-lg6mk\" (UID: \"b7960fae-df40-482f-a9ab-790fa11aaf8f\") " pod="openshift-console-operator/console-operator-58897d9998-lg6mk" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.477595 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/203d975c-92a2-424e-a6c4-e4904440dead-webhook-cert\") pod \"packageserver-d55dfcdfc-q85kb\" (UID: \"203d975c-92a2-424e-a6c4-e4904440dead\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-q85kb" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.477632 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t458n\" (UniqueName: \"kubernetes.io/projected/f94ec039-60b3-4ecb-b784-04a7ebaf4848-kube-api-access-t458n\") pod \"machine-approver-56656f9798-qml2f\" (UID: \"f94ec039-60b3-4ecb-b784-04a7ebaf4848\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qml2f" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.477665 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22gk4\" (UniqueName: \"kubernetes.io/projected/ff8cf71a-e914-4244-b1aa-ede95edd30e8-kube-api-access-22gk4\") pod \"csi-hostpathplugin-7p7x5\" (UID: \"ff8cf71a-e914-4244-b1aa-ede95edd30e8\") " pod="hostpath-provisioner/csi-hostpathplugin-7p7x5" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.477695 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92fpw\" (UniqueName: \"kubernetes.io/projected/712a2a31-4205-4346-9a32-858a77615eb6-kube-api-access-92fpw\") pod \"marketplace-operator-79b997595-sqsdb\" (UID: \"712a2a31-4205-4346-9a32-858a77615eb6\") " pod="openshift-marketplace/marketplace-operator-79b997595-sqsdb" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.477775 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x46k8\" (UniqueName: \"kubernetes.io/projected/a04d88f1-a308-42cf-9462-1a06f6338b7b-kube-api-access-x46k8\") pod \"controller-manager-879f6c89f-n2mhr\" (UID: \"a04d88f1-a308-42cf-9462-1a06f6338b7b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n2mhr" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.477807 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/0f3dc501-2bc1-4930-a644-f020c92c2827-srv-cert\") pod \"olm-operator-6b444d44fb-kmxzj\" (UID: \"0f3dc501-2bc1-4930-a644-f020c92c2827\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kmxzj" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.477840 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7601d311-698b-4389-a9a2-e4e24fae23ea-config-volume\") pod \"collect-profiles-29548230-zr6fd\" (UID: \"7601d311-698b-4389-a9a2-e4e24fae23ea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29548230-zr6fd" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.477880 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlplm\" (UniqueName: \"kubernetes.io/projected/7601d311-698b-4389-a9a2-e4e24fae23ea-kube-api-access-jlplm\") pod \"collect-profiles-29548230-zr6fd\" (UID: \"7601d311-698b-4389-a9a2-e4e24fae23ea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29548230-zr6fd" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.477914 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/ff8cf71a-e914-4244-b1aa-ede95edd30e8-registration-dir\") pod \"csi-hostpathplugin-7p7x5\" (UID: \"ff8cf71a-e914-4244-b1aa-ede95edd30e8\") " pod="hostpath-provisioner/csi-hostpathplugin-7p7x5" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.477986 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/08f9e92f-8419-4ba3-a532-c13e5bb6cab4-available-featuregates\") pod \"openshift-config-operator-7777fb866f-4vkx6\" (UID: \"08f9e92f-8419-4ba3-a532-c13e5bb6cab4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4vkx6" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.478025 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/f94ec039-60b3-4ecb-b784-04a7ebaf4848-machine-approver-tls\") pod \"machine-approver-56656f9798-qml2f\" (UID: \"f94ec039-60b3-4ecb-b784-04a7ebaf4848\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qml2f" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.478060 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-cnbpx\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.478094 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/ff8cf71a-e914-4244-b1aa-ede95edd30e8-plugins-dir\") pod \"csi-hostpathplugin-7p7x5\" (UID: \"ff8cf71a-e914-4244-b1aa-ede95edd30e8\") " pod="hostpath-provisioner/csi-hostpathplugin-7p7x5" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.478132 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b33c7fd7-2d0f-48b7-9932-e7740577280f-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-9gzdv\" (UID: \"b33c7fd7-2d0f-48b7-9932-e7740577280f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9gzdv" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.478166 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8xj6\" (UniqueName: \"kubernetes.io/projected/203d975c-92a2-424e-a6c4-e4904440dead-kube-api-access-r8xj6\") pod \"packageserver-d55dfcdfc-q85kb\" (UID: \"203d975c-92a2-424e-a6c4-e4904440dead\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-q85kb" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.478199 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8d97c7b9-9de3-40a6-ba65-db9c730604c0-auth-proxy-config\") pod \"machine-config-operator-74547568cd-kdrbq\" (UID: \"8d97c7b9-9de3-40a6-ba65-db9c730604c0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kdrbq" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.478247 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/df59699d-5df8-401e-8bb0-0afa130ac06b-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-lk96z\" (UID: \"df59699d-5df8-401e-8bb0-0afa130ac06b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lk96z" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.478298 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-cnbpx\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.478335 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b541cfee-3ec5-4be3-9acd-d58ebb79f0a2-config\") pod \"authentication-operator-69f744f599-ltd8c\" (UID: \"b541cfee-3ec5-4be3-9acd-d58ebb79f0a2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ltd8c" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.478456 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfkx7\" (UniqueName: \"kubernetes.io/projected/0f3dc501-2bc1-4930-a644-f020c92c2827-kube-api-access-rfkx7\") pod \"olm-operator-6b444d44fb-kmxzj\" (UID: \"0f3dc501-2bc1-4930-a644-f020c92c2827\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kmxzj" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.478530 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/698827d2-21c9-4856-a9cd-17bb3df88bce-console-oauth-config\") pod \"console-f9d7485db-zgx7p\" (UID: \"698827d2-21c9-4856-a9cd-17bb3df88bce\") " pod="openshift-console/console-f9d7485db-zgx7p" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.478567 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/203d975c-92a2-424e-a6c4-e4904440dead-apiservice-cert\") pod \"packageserver-d55dfcdfc-q85kb\" (UID: \"203d975c-92a2-424e-a6c4-e4904440dead\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-q85kb" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.478627 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6pzz\" (UniqueName: \"kubernetes.io/projected/80ba5362-78a1-4ed1-a6be-fbda0dbe72df-kube-api-access-x6pzz\") pod \"ingress-operator-5b745b69d9-8x4b2\" (UID: \"80ba5362-78a1-4ed1-a6be-fbda0dbe72df\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8x4b2" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.478662 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/618e4221-7bc8-4721-ac10-1a1373d0fade-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-x8mpt\" (UID: \"618e4221-7bc8-4721-ac10-1a1373d0fade\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x8mpt" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.478698 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/698827d2-21c9-4856-a9cd-17bb3df88bce-service-ca\") pod \"console-f9d7485db-zgx7p\" (UID: \"698827d2-21c9-4856-a9cd-17bb3df88bce\") " pod="openshift-console/console-f9d7485db-zgx7p" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.478734 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f94ec039-60b3-4ecb-b784-04a7ebaf4848-auth-proxy-config\") pod \"machine-approver-56656f9798-qml2f\" (UID: \"f94ec039-60b3-4ecb-b784-04a7ebaf4848\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qml2f" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.478695 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/08f9e92f-8419-4ba3-a532-c13e5bb6cab4-available-featuregates\") pod \"openshift-config-operator-7777fb866f-4vkx6\" (UID: \"08f9e92f-8419-4ba3-a532-c13e5bb6cab4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4vkx6" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.478913 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/31ce0220-0f53-4d63-aa7b-99357e9c6b04-audit-dir\") pod \"oauth-openshift-558db77b4-cnbpx\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.479034 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/986041cd-e5ea-4269-82b0-5799d343b155-cert\") pod \"ingress-canary-w286n\" (UID: \"986041cd-e5ea-4269-82b0-5799d343b155\") " pod="openshift-ingress-canary/ingress-canary-w286n" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.479081 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/51fb375a-53bd-420a-930f-6b6cdb69f793-client-ca\") pod \"route-controller-manager-6576b87f9c-6qmrb\" (UID: \"51fb375a-53bd-420a-930f-6b6cdb69f793\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6qmrb" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.479118 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-cnbpx\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.479151 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvr2j\" (UniqueName: \"kubernetes.io/projected/08f9e92f-8419-4ba3-a532-c13e5bb6cab4-kube-api-access-fvr2j\") pod \"openshift-config-operator-7777fb866f-4vkx6\" (UID: \"08f9e92f-8419-4ba3-a532-c13e5bb6cab4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4vkx6" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.479182 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/31ce0220-0f53-4d63-aa7b-99357e9c6b04-audit-dir\") pod \"oauth-openshift-558db77b4-cnbpx\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.479205 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.479190 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/030c2b00-0d20-447c-b7a5-55d86f1674d8-etcd-service-ca\") pod \"etcd-operator-b45778765-dqpk4\" (UID: \"030c2b00-0d20-447c-b7a5-55d86f1674d8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dqpk4" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.479364 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/7ca6333f-d5ff-4c06-a140-3b9c29710528-default-certificate\") pod \"router-default-5444994796-nk422\" (UID: \"7ca6333f-d5ff-4c06-a140-3b9c29710528\") " pod="openshift-ingress/router-default-5444994796-nk422" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.479413 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-cnbpx\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.479450 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvzjw\" (UniqueName: \"kubernetes.io/projected/40cb1361-4bb1-450f-8d78-1a033e611f04-kube-api-access-lvzjw\") pod \"service-ca-9c57cc56f-2lnfd\" (UID: \"40cb1361-4bb1-450f-8d78-1a033e611f04\") " pod="openshift-service-ca/service-ca-9c57cc56f-2lnfd" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.479521 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/030c2b00-0d20-447c-b7a5-55d86f1674d8-etcd-ca\") pod \"etcd-operator-b45778765-dqpk4\" (UID: \"030c2b00-0d20-447c-b7a5-55d86f1674d8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dqpk4" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.479605 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/7ca6333f-d5ff-4c06-a140-3b9c29710528-stats-auth\") pod \"router-default-5444994796-nk422\" (UID: \"7ca6333f-d5ff-4c06-a140-3b9c29710528\") " pod="openshift-ingress/router-default-5444994796-nk422" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.479649 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/31ce0220-0f53-4d63-aa7b-99357e9c6b04-audit-policies\") pod \"oauth-openshift-558db77b4-cnbpx\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.479682 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6m8vk\" (UniqueName: \"kubernetes.io/projected/0b6ab22d-b401-4727-83f2-94b527c4a366-kube-api-access-6m8vk\") pod \"service-ca-operator-777779d784-ztm29\" (UID: \"0b6ab22d-b401-4727-83f2-94b527c4a366\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ztm29" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.479714 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/0de7b0c4-4acf-40de-867d-1fc33069fa51-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-htv6d\" (UID: \"0de7b0c4-4acf-40de-867d-1fc33069fa51\") " pod="openshift-multus/cni-sysctl-allowlist-ds-htv6d" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.479754 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/712a2a31-4205-4346-9a32-858a77615eb6-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-sqsdb\" (UID: \"712a2a31-4205-4346-9a32-858a77615eb6\") " pod="openshift-marketplace/marketplace-operator-79b997595-sqsdb" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.479806 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-cnbpx\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.479841 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dlkqz\" (UniqueName: \"kubernetes.io/projected/986041cd-e5ea-4269-82b0-5799d343b155-kube-api-access-dlkqz\") pod \"ingress-canary-w286n\" (UID: \"986041cd-e5ea-4269-82b0-5799d343b155\") " pod="openshift-ingress-canary/ingress-canary-w286n" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.479877 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/214a6dbd-445a-4b41-a3b2-f3efc93f6ed4-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-58n5g\" (UID: \"214a6dbd-445a-4b41-a3b2-f3efc93f6ed4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-58n5g" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.479911 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/839b7c38-e05d-4e85-942a-1936e3b2bcf7-node-bootstrap-token\") pod \"machine-config-server-dmkh8\" (UID: \"839b7c38-e05d-4e85-942a-1936e3b2bcf7\") " pod="openshift-machine-config-operator/machine-config-server-dmkh8" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.479979 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a04d88f1-a308-42cf-9462-1a06f6338b7b-client-ca\") pod \"controller-manager-879f6c89f-n2mhr\" (UID: \"a04d88f1-a308-42cf-9462-1a06f6338b7b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n2mhr" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.480010 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/618e4221-7bc8-4721-ac10-1a1373d0fade-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-x8mpt\" (UID: \"618e4221-7bc8-4721-ac10-1a1373d0fade\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x8mpt" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.480043 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hq95\" (UniqueName: \"kubernetes.io/projected/214a6dbd-445a-4b41-a3b2-f3efc93f6ed4-kube-api-access-5hq95\") pod \"multus-admission-controller-857f4d67dd-58n5g\" (UID: \"214a6dbd-445a-4b41-a3b2-f3efc93f6ed4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-58n5g" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.480079 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/0de7b0c4-4acf-40de-867d-1fc33069fa51-ready\") pod \"cni-sysctl-allowlist-ds-htv6d\" (UID: \"0de7b0c4-4acf-40de-867d-1fc33069fa51\") " pod="openshift-multus/cni-sysctl-allowlist-ds-htv6d" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.480109 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/2886e9c5-32c9-440a-8d58-d8c00d7742de-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-l7g67\" (UID: \"2886e9c5-32c9-440a-8d58-d8c00d7742de\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-l7g67" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.480151 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0da2f3db-9ac4-42bf-b4b8-a4d538d9f4e6-config\") pod \"openshift-apiserver-operator-796bbdcf4f-dvjg6\" (UID: \"0da2f3db-9ac4-42bf-b4b8-a4d538d9f4e6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dvjg6" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.480248 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/ff8cf71a-e914-4244-b1aa-ede95edd30e8-mountpoint-dir\") pod \"csi-hostpathplugin-7p7x5\" (UID: \"ff8cf71a-e914-4244-b1aa-ede95edd30e8\") " pod="hostpath-provisioner/csi-hostpathplugin-7p7x5" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.480296 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7601d311-698b-4389-a9a2-e4e24fae23ea-secret-volume\") pod \"collect-profiles-29548230-zr6fd\" (UID: \"7601d311-698b-4389-a9a2-e4e24fae23ea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29548230-zr6fd" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.480331 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zj4zf\" (UniqueName: \"kubernetes.io/projected/839b7c38-e05d-4e85-942a-1936e3b2bcf7-kube-api-access-zj4zf\") pod \"machine-config-server-dmkh8\" (UID: \"839b7c38-e05d-4e85-942a-1936e3b2bcf7\") " pod="openshift-machine-config-operator/machine-config-server-dmkh8" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.480363 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e03f48ff-595f-4537-824f-650401604f7e-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-h2f4f\" (UID: \"e03f48ff-595f-4537-824f-650401604f7e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-h2f4f" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.480402 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-cnbpx\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.480435 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b541cfee-3ec5-4be3-9acd-d58ebb79f0a2-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-ltd8c\" (UID: \"b541cfee-3ec5-4be3-9acd-d58ebb79f0a2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ltd8c" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.480481 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a04d88f1-a308-42cf-9462-1a06f6338b7b-config\") pod \"controller-manager-879f6c89f-n2mhr\" (UID: \"a04d88f1-a308-42cf-9462-1a06f6338b7b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n2mhr" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.480514 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/030c2b00-0d20-447c-b7a5-55d86f1674d8-config\") pod \"etcd-operator-b45778765-dqpk4\" (UID: \"030c2b00-0d20-447c-b7a5-55d86f1674d8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dqpk4" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.480545 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2a45a1ae-558f-4600-b354-dea0169b175f-metrics-tls\") pod \"dns-operator-744455d44c-j89vd\" (UID: \"2a45a1ae-558f-4600-b354-dea0169b175f\") " pod="openshift-dns-operator/dns-operator-744455d44c-j89vd" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.480584 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8d97c7b9-9de3-40a6-ba65-db9c730604c0-proxy-tls\") pod \"machine-config-operator-74547568cd-kdrbq\" (UID: \"8d97c7b9-9de3-40a6-ba65-db9c730604c0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kdrbq" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.480621 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjpfb\" (UniqueName: \"kubernetes.io/projected/030c2b00-0d20-447c-b7a5-55d86f1674d8-kube-api-access-zjpfb\") pod \"etcd-operator-b45778765-dqpk4\" (UID: \"030c2b00-0d20-447c-b7a5-55d86f1674d8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dqpk4" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.480657 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p96tc\" (UniqueName: \"kubernetes.io/projected/23baba8c-4f55-4188-8967-61093c57e913-kube-api-access-p96tc\") pod \"dns-default-2rd7t\" (UID: \"23baba8c-4f55-4188-8967-61093c57e913\") " pod="openshift-dns/dns-default-2rd7t" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.480717 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a9fcbefd-dfa7-4afd-b198-d872137a9f51-installation-pull-secrets\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.480750 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/df59699d-5df8-401e-8bb0-0afa130ac06b-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-lk96z\" (UID: \"df59699d-5df8-401e-8bb0-0afa130ac06b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lk96z" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.480774 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/08f9e92f-8419-4ba3-a532-c13e5bb6cab4-serving-cert\") pod \"openshift-config-operator-7777fb866f-4vkx6\" (UID: \"08f9e92f-8419-4ba3-a532-c13e5bb6cab4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4vkx6" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.480780 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b6ab22d-b401-4727-83f2-94b527c4a366-serving-cert\") pod \"service-ca-operator-777779d784-ztm29\" (UID: \"0b6ab22d-b401-4727-83f2-94b527c4a366\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ztm29" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.480887 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmbvx\" (UniqueName: \"kubernetes.io/projected/e8c3aeaf-47bf-48ab-b029-2f19e51dae4a-kube-api-access-tmbvx\") pod \"machine-config-controller-84d6567774-hzqrp\" (UID: \"e8c3aeaf-47bf-48ab-b029-2f19e51dae4a\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hzqrp" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.480953 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a9fcbefd-dfa7-4afd-b198-d872137a9f51-registry-certificates\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.480986 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a9fcbefd-dfa7-4afd-b198-d872137a9f51-bound-sa-token\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.481023 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7ca6333f-d5ff-4c06-a140-3b9c29710528-service-ca-bundle\") pod \"router-default-5444994796-nk422\" (UID: \"7ca6333f-d5ff-4c06-a140-3b9c29710528\") " pod="openshift-ingress/router-default-5444994796-nk422" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.481053 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/23baba8c-4f55-4188-8967-61093c57e913-config-volume\") pod \"dns-default-2rd7t\" (UID: \"23baba8c-4f55-4188-8967-61093c57e913\") " pod="openshift-dns/dns-default-2rd7t" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.481086 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/40cb1361-4bb1-450f-8d78-1a033e611f04-signing-cabundle\") pod \"service-ca-9c57cc56f-2lnfd\" (UID: \"40cb1361-4bb1-450f-8d78-1a033e611f04\") " pod="openshift-service-ca/service-ca-9c57cc56f-2lnfd" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.481120 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/ff8cf71a-e914-4244-b1aa-ede95edd30e8-csi-data-dir\") pod \"csi-hostpathplugin-7p7x5\" (UID: \"ff8cf71a-e914-4244-b1aa-ede95edd30e8\") " pod="hostpath-provisioner/csi-hostpathplugin-7p7x5" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.481152 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/698827d2-21c9-4856-a9cd-17bb3df88bce-console-config\") pod \"console-f9d7485db-zgx7p\" (UID: \"698827d2-21c9-4856-a9cd-17bb3df88bce\") " pod="openshift-console/console-f9d7485db-zgx7p" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.481193 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.481229 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b7960fae-df40-482f-a9ab-790fa11aaf8f-trusted-ca\") pod \"console-operator-58897d9998-lg6mk\" (UID: \"b7960fae-df40-482f-a9ab-790fa11aaf8f\") " pod="openshift-console-operator/console-operator-58897d9998-lg6mk" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.481260 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/8d97c7b9-9de3-40a6-ba65-db9c730604c0-images\") pod \"machine-config-operator-74547568cd-kdrbq\" (UID: \"8d97c7b9-9de3-40a6-ba65-db9c730604c0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kdrbq" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.481291 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmgxs\" (UniqueName: \"kubernetes.io/projected/2886e9c5-32c9-440a-8d58-d8c00d7742de-kube-api-access-dmgxs\") pod \"package-server-manager-789f6589d5-l7g67\" (UID: \"2886e9c5-32c9-440a-8d58-d8c00d7742de\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-l7g67" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.481345 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-cnbpx\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.481380 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/415baeb7-0937-47e1-9d7e-8c2078911c3c-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-d622z\" (UID: \"415baeb7-0937-47e1-9d7e-8c2078911c3c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d622z" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.481419 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7r5l\" (UniqueName: \"kubernetes.io/projected/7ca6333f-d5ff-4c06-a140-3b9c29710528-kube-api-access-z7r5l\") pod \"router-default-5444994796-nk422\" (UID: \"7ca6333f-d5ff-4c06-a140-3b9c29710528\") " pod="openshift-ingress/router-default-5444994796-nk422" Mar 07 14:41:19 crc kubenswrapper[4943]: E0307 14:41:19.482080 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:19.982061311 +0000 UTC m=+121.934197839 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.483304 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a9fcbefd-dfa7-4afd-b198-d872137a9f51-registry-certificates\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.499552 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.517615 4943 request.go:700] Waited for 1.009187941s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-etcd-operator/secrets?fieldSelector=metadata.name%3Detcd-operator-dockercfg-r9srn&limit=500&resourceVersion=0 Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.519630 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.539758 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.559658 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.580012 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.583008 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.583157 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b33c7fd7-2d0f-48b7-9932-e7740577280f-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-9gzdv\" (UID: \"b33c7fd7-2d0f-48b7-9932-e7740577280f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9gzdv" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.583243 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/df59699d-5df8-401e-8bb0-0afa130ac06b-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-lk96z\" (UID: \"df59699d-5df8-401e-8bb0-0afa130ac06b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lk96z" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.583314 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e03f48ff-595f-4537-824f-650401604f7e-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-h2f4f\" (UID: \"e03f48ff-595f-4537-824f-650401604f7e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-h2f4f" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.583350 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e8c3aeaf-47bf-48ab-b029-2f19e51dae4a-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-hzqrp\" (UID: \"e8c3aeaf-47bf-48ab-b029-2f19e51dae4a\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hzqrp" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.583438 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgzhw\" (UniqueName: \"kubernetes.io/projected/0de7b0c4-4acf-40de-867d-1fc33069fa51-kube-api-access-dgzhw\") pod \"cni-sysctl-allowlist-ds-htv6d\" (UID: \"0de7b0c4-4acf-40de-867d-1fc33069fa51\") " pod="openshift-multus/cni-sysctl-allowlist-ds-htv6d" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.583470 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjwms\" (UniqueName: \"kubernetes.io/projected/2a45a1ae-558f-4600-b354-dea0169b175f-kube-api-access-gjwms\") pod \"dns-operator-744455d44c-j89vd\" (UID: \"2a45a1ae-558f-4600-b354-dea0169b175f\") " pod="openshift-dns-operator/dns-operator-744455d44c-j89vd" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.583524 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqhfd\" (UniqueName: \"kubernetes.io/projected/8e19594a-0a69-40b7-8c24-317979edbfbb-kube-api-access-sqhfd\") pod \"catalog-operator-68c6474976-jpq82\" (UID: \"8e19594a-0a69-40b7-8c24-317979edbfbb\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jpq82" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.583558 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/40cb1361-4bb1-450f-8d78-1a033e611f04-signing-key\") pod \"service-ca-9c57cc56f-2lnfd\" (UID: \"40cb1361-4bb1-450f-8d78-1a033e611f04\") " pod="openshift-service-ca/service-ca-9c57cc56f-2lnfd" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.583619 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/0de7b0c4-4acf-40de-867d-1fc33069fa51-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-htv6d\" (UID: \"0de7b0c4-4acf-40de-867d-1fc33069fa51\") " pod="openshift-multus/cni-sysctl-allowlist-ds-htv6d" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.583691 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/8e19594a-0a69-40b7-8c24-317979edbfbb-profile-collector-cert\") pod \"catalog-operator-68c6474976-jpq82\" (UID: \"8e19594a-0a69-40b7-8c24-317979edbfbb\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jpq82" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.583726 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/415baeb7-0937-47e1-9d7e-8c2078911c3c-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-d622z\" (UID: \"415baeb7-0937-47e1-9d7e-8c2078911c3c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d622z" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.583791 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/839b7c38-e05d-4e85-942a-1936e3b2bcf7-certs\") pod \"machine-config-server-dmkh8\" (UID: \"839b7c38-e05d-4e85-942a-1936e3b2bcf7\") " pod="openshift-machine-config-operator/machine-config-server-dmkh8" Mar 07 14:41:19 crc kubenswrapper[4943]: E0307 14:41:19.583859 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.083817101 +0000 UTC m=+122.035953639 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.583961 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/712a2a31-4205-4346-9a32-858a77615eb6-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-sqsdb\" (UID: \"712a2a31-4205-4346-9a32-858a77615eb6\") " pod="openshift-marketplace/marketplace-operator-79b997595-sqsdb" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.584053 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e03f48ff-595f-4537-824f-650401604f7e-config\") pod \"kube-controller-manager-operator-78b949d7b-h2f4f\" (UID: \"e03f48ff-595f-4537-824f-650401604f7e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-h2f4f" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.584056 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/0de7b0c4-4acf-40de-867d-1fc33069fa51-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-htv6d\" (UID: \"0de7b0c4-4acf-40de-867d-1fc33069fa51\") " pod="openshift-multus/cni-sysctl-allowlist-ds-htv6d" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.584162 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b6ab22d-b401-4727-83f2-94b527c4a366-config\") pod \"service-ca-operator-777779d784-ztm29\" (UID: \"0b6ab22d-b401-4727-83f2-94b527c4a366\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ztm29" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.584204 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x64d9\" (UniqueName: \"kubernetes.io/projected/8d97c7b9-9de3-40a6-ba65-db9c730604c0-kube-api-access-x64d9\") pod \"machine-config-operator-74547568cd-kdrbq\" (UID: \"8d97c7b9-9de3-40a6-ba65-db9c730604c0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kdrbq" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.584299 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/fdcfa33a-5367-46db-8270-662184116079-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-pv7b4\" (UID: \"fdcfa33a-5367-46db-8270-662184116079\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pv7b4" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.584339 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/0f3dc501-2bc1-4930-a644-f020c92c2827-profile-collector-cert\") pod \"olm-operator-6b444d44fb-kmxzj\" (UID: \"0f3dc501-2bc1-4930-a644-f020c92c2827\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kmxzj" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.584371 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/030c2b00-0d20-447c-b7a5-55d86f1674d8-etcd-client\") pod \"etcd-operator-b45778765-dqpk4\" (UID: \"030c2b00-0d20-447c-b7a5-55d86f1674d8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dqpk4" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.584406 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tktfx\" (UniqueName: \"kubernetes.io/projected/f7170a81-b6f1-4696-9356-737721576cf6-kube-api-access-tktfx\") pod \"migrator-59844c95c7-7p4vl\" (UID: \"f7170a81-b6f1-4696-9356-737721576cf6\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7p4vl" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.584519 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-scbtq\" (UniqueName: \"kubernetes.io/projected/fdcfa33a-5367-46db-8270-662184116079-kube-api-access-scbtq\") pod \"control-plane-machine-set-operator-78cbb6b69f-pv7b4\" (UID: \"fdcfa33a-5367-46db-8270-662184116079\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pv7b4" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.584559 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e8c3aeaf-47bf-48ab-b029-2f19e51dae4a-proxy-tls\") pod \"machine-config-controller-84d6567774-hzqrp\" (UID: \"e8c3aeaf-47bf-48ab-b029-2f19e51dae4a\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hzqrp" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.584591 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/23baba8c-4f55-4188-8967-61093c57e913-metrics-tls\") pod \"dns-default-2rd7t\" (UID: \"23baba8c-4f55-4188-8967-61093c57e913\") " pod="openshift-dns/dns-default-2rd7t" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.584658 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/203d975c-92a2-424e-a6c4-e4904440dead-webhook-cert\") pod \"packageserver-d55dfcdfc-q85kb\" (UID: \"203d975c-92a2-424e-a6c4-e4904440dead\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-q85kb" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.584708 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22gk4\" (UniqueName: \"kubernetes.io/projected/ff8cf71a-e914-4244-b1aa-ede95edd30e8-kube-api-access-22gk4\") pod \"csi-hostpathplugin-7p7x5\" (UID: \"ff8cf71a-e914-4244-b1aa-ede95edd30e8\") " pod="hostpath-provisioner/csi-hostpathplugin-7p7x5" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.584749 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92fpw\" (UniqueName: \"kubernetes.io/projected/712a2a31-4205-4346-9a32-858a77615eb6-kube-api-access-92fpw\") pod \"marketplace-operator-79b997595-sqsdb\" (UID: \"712a2a31-4205-4346-9a32-858a77615eb6\") " pod="openshift-marketplace/marketplace-operator-79b997595-sqsdb" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.584783 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlplm\" (UniqueName: \"kubernetes.io/projected/7601d311-698b-4389-a9a2-e4e24fae23ea-kube-api-access-jlplm\") pod \"collect-profiles-29548230-zr6fd\" (UID: \"7601d311-698b-4389-a9a2-e4e24fae23ea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29548230-zr6fd" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.584894 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/0f3dc501-2bc1-4930-a644-f020c92c2827-srv-cert\") pod \"olm-operator-6b444d44fb-kmxzj\" (UID: \"0f3dc501-2bc1-4930-a644-f020c92c2827\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kmxzj" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.584985 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7601d311-698b-4389-a9a2-e4e24fae23ea-config-volume\") pod \"collect-profiles-29548230-zr6fd\" (UID: \"7601d311-698b-4389-a9a2-e4e24fae23ea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29548230-zr6fd" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.585046 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/ff8cf71a-e914-4244-b1aa-ede95edd30e8-registration-dir\") pod \"csi-hostpathplugin-7p7x5\" (UID: \"ff8cf71a-e914-4244-b1aa-ede95edd30e8\") " pod="hostpath-provisioner/csi-hostpathplugin-7p7x5" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.585181 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/ff8cf71a-e914-4244-b1aa-ede95edd30e8-plugins-dir\") pod \"csi-hostpathplugin-7p7x5\" (UID: \"ff8cf71a-e914-4244-b1aa-ede95edd30e8\") " pod="hostpath-provisioner/csi-hostpathplugin-7p7x5" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.585243 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b33c7fd7-2d0f-48b7-9932-e7740577280f-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-9gzdv\" (UID: \"b33c7fd7-2d0f-48b7-9932-e7740577280f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9gzdv" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.585288 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8xj6\" (UniqueName: \"kubernetes.io/projected/203d975c-92a2-424e-a6c4-e4904440dead-kube-api-access-r8xj6\") pod \"packageserver-d55dfcdfc-q85kb\" (UID: \"203d975c-92a2-424e-a6c4-e4904440dead\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-q85kb" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.585330 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8d97c7b9-9de3-40a6-ba65-db9c730604c0-auth-proxy-config\") pod \"machine-config-operator-74547568cd-kdrbq\" (UID: \"8d97c7b9-9de3-40a6-ba65-db9c730604c0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kdrbq" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.585368 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/df59699d-5df8-401e-8bb0-0afa130ac06b-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-lk96z\" (UID: \"df59699d-5df8-401e-8bb0-0afa130ac06b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lk96z" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.585428 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfkx7\" (UniqueName: \"kubernetes.io/projected/0f3dc501-2bc1-4930-a644-f020c92c2827-kube-api-access-rfkx7\") pod \"olm-operator-6b444d44fb-kmxzj\" (UID: \"0f3dc501-2bc1-4930-a644-f020c92c2827\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kmxzj" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.585464 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/ff8cf71a-e914-4244-b1aa-ede95edd30e8-registration-dir\") pod \"csi-hostpathplugin-7p7x5\" (UID: \"ff8cf71a-e914-4244-b1aa-ede95edd30e8\") " pod="hostpath-provisioner/csi-hostpathplugin-7p7x5" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.585548 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/ff8cf71a-e914-4244-b1aa-ede95edd30e8-plugins-dir\") pod \"csi-hostpathplugin-7p7x5\" (UID: \"ff8cf71a-e914-4244-b1aa-ede95edd30e8\") " pod="hostpath-provisioner/csi-hostpathplugin-7p7x5" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.585478 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/203d975c-92a2-424e-a6c4-e4904440dead-apiservice-cert\") pod \"packageserver-d55dfcdfc-q85kb\" (UID: \"203d975c-92a2-424e-a6c4-e4904440dead\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-q85kb" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.585697 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/618e4221-7bc8-4721-ac10-1a1373d0fade-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-x8mpt\" (UID: \"618e4221-7bc8-4721-ac10-1a1373d0fade\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x8mpt" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.585774 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/986041cd-e5ea-4269-82b0-5799d343b155-cert\") pod \"ingress-canary-w286n\" (UID: \"986041cd-e5ea-4269-82b0-5799d343b155\") " pod="openshift-ingress-canary/ingress-canary-w286n" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.585860 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/030c2b00-0d20-447c-b7a5-55d86f1674d8-etcd-service-ca\") pod \"etcd-operator-b45778765-dqpk4\" (UID: \"030c2b00-0d20-447c-b7a5-55d86f1674d8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dqpk4" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.585895 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/7ca6333f-d5ff-4c06-a140-3b9c29710528-default-certificate\") pod \"router-default-5444994796-nk422\" (UID: \"7ca6333f-d5ff-4c06-a140-3b9c29710528\") " pod="openshift-ingress/router-default-5444994796-nk422" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.586004 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvzjw\" (UniqueName: \"kubernetes.io/projected/40cb1361-4bb1-450f-8d78-1a033e611f04-kube-api-access-lvzjw\") pod \"service-ca-9c57cc56f-2lnfd\" (UID: \"40cb1361-4bb1-450f-8d78-1a033e611f04\") " pod="openshift-service-ca/service-ca-9c57cc56f-2lnfd" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.586044 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/030c2b00-0d20-447c-b7a5-55d86f1674d8-etcd-ca\") pod \"etcd-operator-b45778765-dqpk4\" (UID: \"030c2b00-0d20-447c-b7a5-55d86f1674d8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dqpk4" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.586076 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/7ca6333f-d5ff-4c06-a140-3b9c29710528-stats-auth\") pod \"router-default-5444994796-nk422\" (UID: \"7ca6333f-d5ff-4c06-a140-3b9c29710528\") " pod="openshift-ingress/router-default-5444994796-nk422" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.586123 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6m8vk\" (UniqueName: \"kubernetes.io/projected/0b6ab22d-b401-4727-83f2-94b527c4a366-kube-api-access-6m8vk\") pod \"service-ca-operator-777779d784-ztm29\" (UID: \"0b6ab22d-b401-4727-83f2-94b527c4a366\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ztm29" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.586156 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/0de7b0c4-4acf-40de-867d-1fc33069fa51-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-htv6d\" (UID: \"0de7b0c4-4acf-40de-867d-1fc33069fa51\") " pod="openshift-multus/cni-sysctl-allowlist-ds-htv6d" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.586190 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/712a2a31-4205-4346-9a32-858a77615eb6-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-sqsdb\" (UID: \"712a2a31-4205-4346-9a32-858a77615eb6\") " pod="openshift-marketplace/marketplace-operator-79b997595-sqsdb" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.586238 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dlkqz\" (UniqueName: \"kubernetes.io/projected/986041cd-e5ea-4269-82b0-5799d343b155-kube-api-access-dlkqz\") pod \"ingress-canary-w286n\" (UID: \"986041cd-e5ea-4269-82b0-5799d343b155\") " pod="openshift-ingress-canary/ingress-canary-w286n" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.586270 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/214a6dbd-445a-4b41-a3b2-f3efc93f6ed4-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-58n5g\" (UID: \"214a6dbd-445a-4b41-a3b2-f3efc93f6ed4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-58n5g" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.586309 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/839b7c38-e05d-4e85-942a-1936e3b2bcf7-node-bootstrap-token\") pod \"machine-config-server-dmkh8\" (UID: \"839b7c38-e05d-4e85-942a-1936e3b2bcf7\") " pod="openshift-machine-config-operator/machine-config-server-dmkh8" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.586363 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/618e4221-7bc8-4721-ac10-1a1373d0fade-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-x8mpt\" (UID: \"618e4221-7bc8-4721-ac10-1a1373d0fade\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x8mpt" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.586397 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hq95\" (UniqueName: \"kubernetes.io/projected/214a6dbd-445a-4b41-a3b2-f3efc93f6ed4-kube-api-access-5hq95\") pod \"multus-admission-controller-857f4d67dd-58n5g\" (UID: \"214a6dbd-445a-4b41-a3b2-f3efc93f6ed4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-58n5g" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.586428 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/0de7b0c4-4acf-40de-867d-1fc33069fa51-ready\") pod \"cni-sysctl-allowlist-ds-htv6d\" (UID: \"0de7b0c4-4acf-40de-867d-1fc33069fa51\") " pod="openshift-multus/cni-sysctl-allowlist-ds-htv6d" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.586464 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/2886e9c5-32c9-440a-8d58-d8c00d7742de-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-l7g67\" (UID: \"2886e9c5-32c9-440a-8d58-d8c00d7742de\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-l7g67" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.586500 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e03f48ff-595f-4537-824f-650401604f7e-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-h2f4f\" (UID: \"e03f48ff-595f-4537-824f-650401604f7e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-h2f4f" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.586549 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/ff8cf71a-e914-4244-b1aa-ede95edd30e8-mountpoint-dir\") pod \"csi-hostpathplugin-7p7x5\" (UID: \"ff8cf71a-e914-4244-b1aa-ede95edd30e8\") " pod="hostpath-provisioner/csi-hostpathplugin-7p7x5" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.586584 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7601d311-698b-4389-a9a2-e4e24fae23ea-secret-volume\") pod \"collect-profiles-29548230-zr6fd\" (UID: \"7601d311-698b-4389-a9a2-e4e24fae23ea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29548230-zr6fd" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.586621 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zj4zf\" (UniqueName: \"kubernetes.io/projected/839b7c38-e05d-4e85-942a-1936e3b2bcf7-kube-api-access-zj4zf\") pod \"machine-config-server-dmkh8\" (UID: \"839b7c38-e05d-4e85-942a-1936e3b2bcf7\") " pod="openshift-machine-config-operator/machine-config-server-dmkh8" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.586701 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/030c2b00-0d20-447c-b7a5-55d86f1674d8-config\") pod \"etcd-operator-b45778765-dqpk4\" (UID: \"030c2b00-0d20-447c-b7a5-55d86f1674d8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dqpk4" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.586734 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2a45a1ae-558f-4600-b354-dea0169b175f-metrics-tls\") pod \"dns-operator-744455d44c-j89vd\" (UID: \"2a45a1ae-558f-4600-b354-dea0169b175f\") " pod="openshift-dns-operator/dns-operator-744455d44c-j89vd" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.586765 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8d97c7b9-9de3-40a6-ba65-db9c730604c0-proxy-tls\") pod \"machine-config-operator-74547568cd-kdrbq\" (UID: \"8d97c7b9-9de3-40a6-ba65-db9c730604c0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kdrbq" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.586809 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjpfb\" (UniqueName: \"kubernetes.io/projected/030c2b00-0d20-447c-b7a5-55d86f1674d8-kube-api-access-zjpfb\") pod \"etcd-operator-b45778765-dqpk4\" (UID: \"030c2b00-0d20-447c-b7a5-55d86f1674d8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dqpk4" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.586849 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p96tc\" (UniqueName: \"kubernetes.io/projected/23baba8c-4f55-4188-8967-61093c57e913-kube-api-access-p96tc\") pod \"dns-default-2rd7t\" (UID: \"23baba8c-4f55-4188-8967-61093c57e913\") " pod="openshift-dns/dns-default-2rd7t" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.586862 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/415baeb7-0937-47e1-9d7e-8c2078911c3c-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-d622z\" (UID: \"415baeb7-0937-47e1-9d7e-8c2078911c3c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d622z" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.586906 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/df59699d-5df8-401e-8bb0-0afa130ac06b-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-lk96z\" (UID: \"df59699d-5df8-401e-8bb0-0afa130ac06b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lk96z" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.586916 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/712a2a31-4205-4346-9a32-858a77615eb6-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-sqsdb\" (UID: \"712a2a31-4205-4346-9a32-858a77615eb6\") " pod="openshift-marketplace/marketplace-operator-79b997595-sqsdb" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.587085 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b6ab22d-b401-4727-83f2-94b527c4a366-serving-cert\") pod \"service-ca-operator-777779d784-ztm29\" (UID: \"0b6ab22d-b401-4727-83f2-94b527c4a366\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ztm29" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.587160 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8d97c7b9-9de3-40a6-ba65-db9c730604c0-auth-proxy-config\") pod \"machine-config-operator-74547568cd-kdrbq\" (UID: \"8d97c7b9-9de3-40a6-ba65-db9c730604c0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kdrbq" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.587159 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmbvx\" (UniqueName: \"kubernetes.io/projected/e8c3aeaf-47bf-48ab-b029-2f19e51dae4a-kube-api-access-tmbvx\") pod \"machine-config-controller-84d6567774-hzqrp\" (UID: \"e8c3aeaf-47bf-48ab-b029-2f19e51dae4a\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hzqrp" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.587230 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/030c2b00-0d20-447c-b7a5-55d86f1674d8-etcd-service-ca\") pod \"etcd-operator-b45778765-dqpk4\" (UID: \"030c2b00-0d20-447c-b7a5-55d86f1674d8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dqpk4" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.587289 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7ca6333f-d5ff-4c06-a140-3b9c29710528-service-ca-bundle\") pod \"router-default-5444994796-nk422\" (UID: \"7ca6333f-d5ff-4c06-a140-3b9c29710528\") " pod="openshift-ingress/router-default-5444994796-nk422" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.587334 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/23baba8c-4f55-4188-8967-61093c57e913-config-volume\") pod \"dns-default-2rd7t\" (UID: \"23baba8c-4f55-4188-8967-61093c57e913\") " pod="openshift-dns/dns-default-2rd7t" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.587583 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/ff8cf71a-e914-4244-b1aa-ede95edd30e8-mountpoint-dir\") pod \"csi-hostpathplugin-7p7x5\" (UID: \"ff8cf71a-e914-4244-b1aa-ede95edd30e8\") " pod="hostpath-provisioner/csi-hostpathplugin-7p7x5" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.587764 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/40cb1361-4bb1-450f-8d78-1a033e611f04-signing-cabundle\") pod \"service-ca-9c57cc56f-2lnfd\" (UID: \"40cb1361-4bb1-450f-8d78-1a033e611f04\") " pod="openshift-service-ca/service-ca-9c57cc56f-2lnfd" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.588097 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/0de7b0c4-4acf-40de-867d-1fc33069fa51-ready\") pod \"cni-sysctl-allowlist-ds-htv6d\" (UID: \"0de7b0c4-4acf-40de-867d-1fc33069fa51\") " pod="openshift-multus/cni-sysctl-allowlist-ds-htv6d" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.588432 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/030c2b00-0d20-447c-b7a5-55d86f1674d8-config\") pod \"etcd-operator-b45778765-dqpk4\" (UID: \"030c2b00-0d20-447c-b7a5-55d86f1674d8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dqpk4" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.589237 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e8c3aeaf-47bf-48ab-b029-2f19e51dae4a-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-hzqrp\" (UID: \"e8c3aeaf-47bf-48ab-b029-2f19e51dae4a\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hzqrp" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.589253 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/030c2b00-0d20-447c-b7a5-55d86f1674d8-etcd-ca\") pod \"etcd-operator-b45778765-dqpk4\" (UID: \"030c2b00-0d20-447c-b7a5-55d86f1674d8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dqpk4" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.589458 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/030c2b00-0d20-447c-b7a5-55d86f1674d8-etcd-client\") pod \"etcd-operator-b45778765-dqpk4\" (UID: \"030c2b00-0d20-447c-b7a5-55d86f1674d8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dqpk4" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.589748 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/ff8cf71a-e914-4244-b1aa-ede95edd30e8-csi-data-dir\") pod \"csi-hostpathplugin-7p7x5\" (UID: \"ff8cf71a-e914-4244-b1aa-ede95edd30e8\") " pod="hostpath-provisioner/csi-hostpathplugin-7p7x5" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.589773 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/fdcfa33a-5367-46db-8270-662184116079-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-pv7b4\" (UID: \"fdcfa33a-5367-46db-8270-662184116079\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pv7b4" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.589919 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.589961 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/ff8cf71a-e914-4244-b1aa-ede95edd30e8-csi-data-dir\") pod \"csi-hostpathplugin-7p7x5\" (UID: \"ff8cf71a-e914-4244-b1aa-ede95edd30e8\") " pod="hostpath-provisioner/csi-hostpathplugin-7p7x5" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.590063 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/8d97c7b9-9de3-40a6-ba65-db9c730604c0-images\") pod \"machine-config-operator-74547568cd-kdrbq\" (UID: \"8d97c7b9-9de3-40a6-ba65-db9c730604c0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kdrbq" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.590132 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmgxs\" (UniqueName: \"kubernetes.io/projected/2886e9c5-32c9-440a-8d58-d8c00d7742de-kube-api-access-dmgxs\") pod \"package-server-manager-789f6589d5-l7g67\" (UID: \"2886e9c5-32c9-440a-8d58-d8c00d7742de\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-l7g67" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.590245 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/415baeb7-0937-47e1-9d7e-8c2078911c3c-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-d622z\" (UID: \"415baeb7-0937-47e1-9d7e-8c2078911c3c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d622z" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.590298 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7r5l\" (UniqueName: \"kubernetes.io/projected/7ca6333f-d5ff-4c06-a140-3b9c29710528-kube-api-access-z7r5l\") pod \"router-default-5444994796-nk422\" (UID: \"7ca6333f-d5ff-4c06-a140-3b9c29710528\") " pod="openshift-ingress/router-default-5444994796-nk422" Mar 07 14:41:19 crc kubenswrapper[4943]: E0307 14:41:19.590359 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.09034288 +0000 UTC m=+122.042479408 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.590427 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7ca6333f-d5ff-4c06-a140-3b9c29710528-metrics-certs\") pod \"router-default-5444994796-nk422\" (UID: \"7ca6333f-d5ff-4c06-a140-3b9c29710528\") " pod="openshift-ingress/router-default-5444994796-nk422" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.590478 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/8e19594a-0a69-40b7-8c24-317979edbfbb-srv-cert\") pod \"catalog-operator-68c6474976-jpq82\" (UID: \"8e19594a-0a69-40b7-8c24-317979edbfbb\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jpq82" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.590513 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/203d975c-92a2-424e-a6c4-e4904440dead-tmpfs\") pod \"packageserver-d55dfcdfc-q85kb\" (UID: \"203d975c-92a2-424e-a6c4-e4904440dead\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-q85kb" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.590572 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6wjd\" (UniqueName: \"kubernetes.io/projected/df59699d-5df8-401e-8bb0-0afa130ac06b-kube-api-access-p6wjd\") pod \"cluster-image-registry-operator-dc59b4c8b-lk96z\" (UID: \"df59699d-5df8-401e-8bb0-0afa130ac06b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lk96z" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.590610 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b33c7fd7-2d0f-48b7-9932-e7740577280f-config\") pod \"kube-apiserver-operator-766d6c64bb-9gzdv\" (UID: \"b33c7fd7-2d0f-48b7-9932-e7740577280f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9gzdv" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.590687 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2m79r\" (UniqueName: \"kubernetes.io/projected/415baeb7-0937-47e1-9d7e-8c2078911c3c-kube-api-access-2m79r\") pod \"kube-storage-version-migrator-operator-b67b599dd-d622z\" (UID: \"415baeb7-0937-47e1-9d7e-8c2078911c3c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d622z" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.590722 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/030c2b00-0d20-447c-b7a5-55d86f1674d8-serving-cert\") pod \"etcd-operator-b45778765-dqpk4\" (UID: \"030c2b00-0d20-447c-b7a5-55d86f1674d8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dqpk4" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.590747 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7ca6333f-d5ff-4c06-a140-3b9c29710528-service-ca-bundle\") pod \"router-default-5444994796-nk422\" (UID: \"7ca6333f-d5ff-4c06-a140-3b9c29710528\") " pod="openshift-ingress/router-default-5444994796-nk422" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.590817 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/618e4221-7bc8-4721-ac10-1a1373d0fade-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-x8mpt\" (UID: \"618e4221-7bc8-4721-ac10-1a1373d0fade\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x8mpt" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.590853 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/ff8cf71a-e914-4244-b1aa-ede95edd30e8-socket-dir\") pod \"csi-hostpathplugin-7p7x5\" (UID: \"ff8cf71a-e914-4244-b1aa-ede95edd30e8\") " pod="hostpath-provisioner/csi-hostpathplugin-7p7x5" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.591436 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/203d975c-92a2-424e-a6c4-e4904440dead-tmpfs\") pod \"packageserver-d55dfcdfc-q85kb\" (UID: \"203d975c-92a2-424e-a6c4-e4904440dead\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-q85kb" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.591804 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/ff8cf71a-e914-4244-b1aa-ede95edd30e8-socket-dir\") pod \"csi-hostpathplugin-7p7x5\" (UID: \"ff8cf71a-e914-4244-b1aa-ede95edd30e8\") " pod="hostpath-provisioner/csi-hostpathplugin-7p7x5" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.592409 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/618e4221-7bc8-4721-ac10-1a1373d0fade-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-x8mpt\" (UID: \"618e4221-7bc8-4721-ac10-1a1373d0fade\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x8mpt" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.593094 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b33c7fd7-2d0f-48b7-9932-e7740577280f-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-9gzdv\" (UID: \"b33c7fd7-2d0f-48b7-9932-e7740577280f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9gzdv" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.593428 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/7ca6333f-d5ff-4c06-a140-3b9c29710528-default-certificate\") pod \"router-default-5444994796-nk422\" (UID: \"7ca6333f-d5ff-4c06-a140-3b9c29710528\") " pod="openshift-ingress/router-default-5444994796-nk422" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.594365 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/712a2a31-4205-4346-9a32-858a77615eb6-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-sqsdb\" (UID: \"712a2a31-4205-4346-9a32-858a77615eb6\") " pod="openshift-marketplace/marketplace-operator-79b997595-sqsdb" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.595081 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/7ca6333f-d5ff-4c06-a140-3b9c29710528-stats-auth\") pod \"router-default-5444994796-nk422\" (UID: \"7ca6333f-d5ff-4c06-a140-3b9c29710528\") " pod="openshift-ingress/router-default-5444994796-nk422" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.595630 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7ca6333f-d5ff-4c06-a140-3b9c29710528-metrics-certs\") pod \"router-default-5444994796-nk422\" (UID: \"7ca6333f-d5ff-4c06-a140-3b9c29710528\") " pod="openshift-ingress/router-default-5444994796-nk422" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.596075 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/030c2b00-0d20-447c-b7a5-55d86f1674d8-serving-cert\") pod \"etcd-operator-b45778765-dqpk4\" (UID: \"030c2b00-0d20-447c-b7a5-55d86f1674d8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dqpk4" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.596712 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2a45a1ae-558f-4600-b354-dea0169b175f-metrics-tls\") pod \"dns-operator-744455d44c-j89vd\" (UID: \"2a45a1ae-558f-4600-b354-dea0169b175f\") " pod="openshift-dns-operator/dns-operator-744455d44c-j89vd" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.599188 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/415baeb7-0937-47e1-9d7e-8c2078911c3c-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-d622z\" (UID: \"415baeb7-0937-47e1-9d7e-8c2078911c3c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d622z" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.599596 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/618e4221-7bc8-4721-ac10-1a1373d0fade-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-x8mpt\" (UID: \"618e4221-7bc8-4721-ac10-1a1373d0fade\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x8mpt" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.601595 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.606213 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/214a6dbd-445a-4b41-a3b2-f3efc93f6ed4-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-58n5g\" (UID: \"214a6dbd-445a-4b41-a3b2-f3efc93f6ed4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-58n5g" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.614327 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e8c3aeaf-47bf-48ab-b029-2f19e51dae4a-proxy-tls\") pod \"machine-config-controller-84d6567774-hzqrp\" (UID: \"e8c3aeaf-47bf-48ab-b029-2f19e51dae4a\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hzqrp" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.620043 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.640782 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.661147 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.681053 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.686763 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/8e19594a-0a69-40b7-8c24-317979edbfbb-srv-cert\") pod \"catalog-operator-68c6474976-jpq82\" (UID: \"8e19594a-0a69-40b7-8c24-317979edbfbb\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jpq82" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.691727 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:19 crc kubenswrapper[4943]: E0307 14:41:19.692029 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.191993468 +0000 UTC m=+122.144130006 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.694145 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:19 crc kubenswrapper[4943]: E0307 14:41:19.694735 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.194716688 +0000 UTC m=+122.146853226 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.700356 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.708243 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/8e19594a-0a69-40b7-8c24-317979edbfbb-profile-collector-cert\") pod \"catalog-operator-68c6474976-jpq82\" (UID: \"8e19594a-0a69-40b7-8c24-317979edbfbb\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jpq82" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.710182 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/0f3dc501-2bc1-4930-a644-f020c92c2827-profile-collector-cert\") pod \"olm-operator-6b444d44fb-kmxzj\" (UID: \"0f3dc501-2bc1-4930-a644-f020c92c2827\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kmxzj" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.711102 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7601d311-698b-4389-a9a2-e4e24fae23ea-secret-volume\") pod \"collect-profiles-29548230-zr6fd\" (UID: \"7601d311-698b-4389-a9a2-e4e24fae23ea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29548230-zr6fd" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.718850 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.738918 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.754814 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.754821 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.754918 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ntc42" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.755119 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.759479 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.795075 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:19 crc kubenswrapper[4943]: E0307 14:41:19.795419 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.29538167 +0000 UTC m=+122.247518208 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.795694 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:19 crc kubenswrapper[4943]: E0307 14:41:19.796219 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.296202011 +0000 UTC m=+122.248338549 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.804010 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.805791 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.815399 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b6ab22d-b401-4727-83f2-94b527c4a366-config\") pod \"service-ca-operator-777779d784-ztm29\" (UID: \"0b6ab22d-b401-4727-83f2-94b527c4a366\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ztm29" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.820141 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.833512 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b6ab22d-b401-4727-83f2-94b527c4a366-serving-cert\") pod \"service-ca-operator-777779d784-ztm29\" (UID: \"0b6ab22d-b401-4727-83f2-94b527c4a366\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ztm29" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.839825 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.859292 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.869049 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/203d975c-92a2-424e-a6c4-e4904440dead-webhook-cert\") pod \"packageserver-d55dfcdfc-q85kb\" (UID: \"203d975c-92a2-424e-a6c4-e4904440dead\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-q85kb" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.870993 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/203d975c-92a2-424e-a6c4-e4904440dead-apiservice-cert\") pod \"packageserver-d55dfcdfc-q85kb\" (UID: \"203d975c-92a2-424e-a6c4-e4904440dead\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-q85kb" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.898441 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:19 crc kubenswrapper[4943]: E0307 14:41:19.898608 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.398571177 +0000 UTC m=+122.350707715 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.899642 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.900520 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:19 crc kubenswrapper[4943]: E0307 14:41:19.901111 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.401084882 +0000 UTC m=+122.353221400 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.912371 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/2886e9c5-32c9-440a-8d58-d8c00d7742de-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-l7g67\" (UID: \"2886e9c5-32c9-440a-8d58-d8c00d7742de\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-l7g67" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.919190 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.926396 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7601d311-698b-4389-a9a2-e4e24fae23ea-config-volume\") pod \"collect-profiles-29548230-zr6fd\" (UID: \"7601d311-698b-4389-a9a2-e4e24fae23ea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29548230-zr6fd" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.939611 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.959995 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.980084 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 07 14:41:19 crc kubenswrapper[4943]: I0307 14:41:19.999447 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.001568 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.001786 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.501759304 +0000 UTC m=+122.453895842 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.002798 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.003338 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.503309084 +0000 UTC m=+122.455445612 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.013072 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e03f48ff-595f-4537-824f-650401604f7e-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-h2f4f\" (UID: \"e03f48ff-595f-4537-824f-650401604f7e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-h2f4f" Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.019003 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.025572 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e03f48ff-595f-4537-824f-650401604f7e-config\") pod \"kube-controller-manager-operator-78b949d7b-h2f4f\" (UID: \"e03f48ff-595f-4537-824f-650401604f7e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-h2f4f" Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.079804 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.081463 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/40cb1361-4bb1-450f-8d78-1a033e611f04-signing-cabundle\") pod \"service-ca-9c57cc56f-2lnfd\" (UID: \"40cb1361-4bb1-450f-8d78-1a033e611f04\") " pod="openshift-service-ca/service-ca-9c57cc56f-2lnfd" Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.099434 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.104788 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.105151 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.605117645 +0000 UTC m=+122.557254183 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.105505 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.106036 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.606018079 +0000 UTC m=+122.558154607 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.110070 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/40cb1361-4bb1-450f-8d78-1a033e611f04-signing-key\") pod \"service-ca-9c57cc56f-2lnfd\" (UID: \"40cb1361-4bb1-450f-8d78-1a033e611f04\") " pod="openshift-service-ca/service-ca-9c57cc56f-2lnfd" Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.119138 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.138972 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.160203 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.179326 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.190354 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/0f3dc501-2bc1-4930-a644-f020c92c2827-srv-cert\") pod \"olm-operator-6b444d44fb-kmxzj\" (UID: \"0f3dc501-2bc1-4930-a644-f020c92c2827\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kmxzj" Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.199851 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.202231 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/8d97c7b9-9de3-40a6-ba65-db9c730604c0-images\") pod \"machine-config-operator-74547568cd-kdrbq\" (UID: \"8d97c7b9-9de3-40a6-ba65-db9c730604c0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kdrbq" Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.208213 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.208362 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.708333373 +0000 UTC m=+122.660469911 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.209178 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.209638 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.709620567 +0000 UTC m=+122.661757075 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.219972 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.239365 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.244025 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8d97c7b9-9de3-40a6-ba65-db9c730604c0-proxy-tls\") pod \"machine-config-operator-74547568cd-kdrbq\" (UID: \"8d97c7b9-9de3-40a6-ba65-db9c730604c0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kdrbq" Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.279967 4943 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.299570 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.311430 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.311672 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.811638683 +0000 UTC m=+122.763775221 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.312570 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.313183 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.813155813 +0000 UTC m=+122.765292351 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.319108 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.340240 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.349243 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/839b7c38-e05d-4e85-942a-1936e3b2bcf7-certs\") pod \"machine-config-server-dmkh8\" (UID: \"839b7c38-e05d-4e85-942a-1936e3b2bcf7\") " pod="openshift-machine-config-operator/machine-config-server-dmkh8" Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.359782 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.379487 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.390664 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/23baba8c-4f55-4188-8967-61093c57e913-config-volume\") pod \"dns-default-2rd7t\" (UID: \"23baba8c-4f55-4188-8967-61093c57e913\") " pod="openshift-dns/dns-default-2rd7t" Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.399470 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.414176 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.414322 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.914290067 +0000 UTC m=+122.866426605 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.415510 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.416066 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.916048662 +0000 UTC m=+122.868185210 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.419456 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.433595 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/839b7c38-e05d-4e85-942a-1936e3b2bcf7-node-bootstrap-token\") pod \"machine-config-server-dmkh8\" (UID: \"839b7c38-e05d-4e85-942a-1936e3b2bcf7\") " pod="openshift-machine-config-operator/machine-config-server-dmkh8" Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.440136 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.449906 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/23baba8c-4f55-4188-8967-61093c57e913-metrics-tls\") pod \"dns-default-2rd7t\" (UID: \"23baba8c-4f55-4188-8967-61093c57e913\") " pod="openshift-dns/dns-default-2rd7t" Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.459705 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-sysctl-allowlist" Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.471256 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/0de7b0c4-4acf-40de-867d-1fc33069fa51-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-htv6d\" (UID: \"0de7b0c4-4acf-40de-867d-1fc33069fa51\") " pod="openshift-multus/cni-sysctl-allowlist-ds-htv6d" Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.474958 4943 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-user-idp-0-file-data: failed to sync secret cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.475008 4943 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-system-router-certs: failed to sync secret cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.475073 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-user-idp-0-file-data podName:31ce0220-0f53-4d63-aa7b-99357e9c6b04 nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.975043973 +0000 UTC m=+122.927180511 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-user-idp-0-file-data" (UniqueName: "kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-user-idp-0-file-data") pod "oauth-openshift-558db77b4-cnbpx" (UID: "31ce0220-0f53-4d63-aa7b-99357e9c6b04") : failed to sync secret cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.475093 4943 configmap.go:193] Couldn't get configMap openshift-image-registry/trusted-ca: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.475105 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-router-certs podName:31ce0220-0f53-4d63-aa7b-99357e9c6b04 nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.975091184 +0000 UTC m=+122.927227712 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-system-router-certs" (UniqueName: "kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-router-certs") pod "oauth-openshift-558db77b4-cnbpx" (UID: "31ce0220-0f53-4d63-aa7b-99357e9c6b04") : failed to sync secret cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.475105 4943 secret.go:188] Couldn't get secret openshift-cluster-samples-operator/samples-operator-tls: failed to sync secret cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.475166 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/28995414-2bd3-4b50-90ec-0fbb63a15ef6-samples-operator-tls podName:28995414-2bd3-4b50-90ec-0fbb63a15ef6 nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.975155656 +0000 UTC m=+122.927292194 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "samples-operator-tls" (UniqueName: "kubernetes.io/secret/28995414-2bd3-4b50-90ec-0fbb63a15ef6-samples-operator-tls") pod "cluster-samples-operator-665b6dd947-dr8bw" (UID: "28995414-2bd3-4b50-90ec-0fbb63a15ef6") : failed to sync secret cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.475189 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/a9fcbefd-dfa7-4afd-b198-d872137a9f51-trusted-ca podName:a9fcbefd-dfa7-4afd-b198-d872137a9f51 nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.975178526 +0000 UTC m=+122.927315064 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "trusted-ca" (UniqueName: "kubernetes.io/configmap/a9fcbefd-dfa7-4afd-b198-d872137a9f51-trusted-ca") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.475226 4943 secret.go:188] Couldn't get secret openshift-console-operator/serving-cert: failed to sync secret cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.475223 4943 configmap.go:193] Couldn't get configMap openshift-console/oauth-serving-cert: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.475278 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b7960fae-df40-482f-a9ab-790fa11aaf8f-serving-cert podName:b7960fae-df40-482f-a9ab-790fa11aaf8f nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.975265899 +0000 UTC m=+122.927402427 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/b7960fae-df40-482f-a9ab-790fa11aaf8f-serving-cert") pod "console-operator-58897d9998-lg6mk" (UID: "b7960fae-df40-482f-a9ab-790fa11aaf8f") : failed to sync secret cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.475281 4943 configmap.go:193] Couldn't get configMap openshift-console/trusted-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.475311 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/698827d2-21c9-4856-a9cd-17bb3df88bce-oauth-serving-cert podName:698827d2-21c9-4856-a9cd-17bb3df88bce nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.975288709 +0000 UTC m=+122.927425247 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "oauth-serving-cert" (UniqueName: "kubernetes.io/configmap/698827d2-21c9-4856-a9cd-17bb3df88bce-oauth-serving-cert") pod "console-f9d7485db-zgx7p" (UID: "698827d2-21c9-4856-a9cd-17bb3df88bce") : failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.475356 4943 configmap.go:193] Couldn't get configMap openshift-authentication/v4-0-config-system-trusted-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.475432 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/698827d2-21c9-4856-a9cd-17bb3df88bce-trusted-ca-bundle podName:698827d2-21c9-4856-a9cd-17bb3df88bce nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.975378721 +0000 UTC m=+122.927515249 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "trusted-ca-bundle" (UniqueName: "kubernetes.io/configmap/698827d2-21c9-4856-a9cd-17bb3df88bce-trusted-ca-bundle") pod "console-f9d7485db-zgx7p" (UID: "698827d2-21c9-4856-a9cd-17bb3df88bce") : failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.475503 4943 projected.go:263] Couldn't get secret openshift-image-registry/image-registry-tls: failed to sync secret cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.475524 4943 projected.go:194] Error preparing data for projected volume registry-tls for pod openshift-image-registry/image-registry-697d97f7c8-r82fw: failed to sync secret cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.475527 4943 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-user-template-error: failed to sync secret cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.475572 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a9fcbefd-dfa7-4afd-b198-d872137a9f51-registry-tls podName:a9fcbefd-dfa7-4afd-b198-d872137a9f51 nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.975552166 +0000 UTC m=+122.927688694 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "registry-tls" (UniqueName: "kubernetes.io/projected/a9fcbefd-dfa7-4afd-b198-d872137a9f51-registry-tls") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : failed to sync secret cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.475573 4943 secret.go:188] Couldn't get secret openshift-controller-manager/serving-cert: failed to sync secret cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.475611 4943 secret.go:188] Couldn't get secret openshift-console/console-serving-cert: failed to sync secret cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.475616 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-user-template-error podName:31ce0220-0f53-4d63-aa7b-99357e9c6b04 nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.975590197 +0000 UTC m=+122.927726725 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-user-template-error" (UniqueName: "kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-user-template-error") pod "oauth-openshift-558db77b4-cnbpx" (UID: "31ce0220-0f53-4d63-aa7b-99357e9c6b04") : failed to sync secret cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.475649 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a04d88f1-a308-42cf-9462-1a06f6338b7b-serving-cert podName:a04d88f1-a308-42cf-9462-1a06f6338b7b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.975636448 +0000 UTC m=+122.927772976 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/a04d88f1-a308-42cf-9462-1a06f6338b7b-serving-cert") pod "controller-manager-879f6c89f-n2mhr" (UID: "a04d88f1-a308-42cf-9462-1a06f6338b7b") : failed to sync secret cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.475673 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/698827d2-21c9-4856-a9cd-17bb3df88bce-console-serving-cert podName:698827d2-21c9-4856-a9cd-17bb3df88bce nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.975662549 +0000 UTC m=+122.927799087 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "console-serving-cert" (UniqueName: "kubernetes.io/secret/698827d2-21c9-4856-a9cd-17bb3df88bce-console-serving-cert") pod "console-f9d7485db-zgx7p" (UID: "698827d2-21c9-4856-a9cd-17bb3df88bce") : failed to sync secret cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.476334 4943 configmap.go:193] Couldn't get configMap openshift-cluster-machine-approver/machine-approver-config: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.476357 4943 secret.go:188] Couldn't get secret openshift-controller-manager-operator/openshift-controller-manager-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.476396 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/f94ec039-60b3-4ecb-b784-04a7ebaf4848-config podName:f94ec039-60b3-4ecb-b784-04a7ebaf4848 nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.976372217 +0000 UTC m=+122.928508745 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/f94ec039-60b3-4ecb-b784-04a7ebaf4848-config") pod "machine-approver-56656f9798-qml2f" (UID: "f94ec039-60b3-4ecb-b784-04a7ebaf4848") : failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.476431 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/00cc0b59-be8e-402b-b6c9-56e9e20c4b9b-serving-cert podName:00cc0b59-be8e-402b-b6c9-56e9e20c4b9b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.976410508 +0000 UTC m=+122.928547036 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/00cc0b59-be8e-402b-b6c9-56e9e20c4b9b-serving-cert") pod "openshift-controller-manager-operator-756b6f6bc6-x5qjf" (UID: "00cc0b59-be8e-402b-b6c9-56e9e20c4b9b") : failed to sync secret cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.476451 4943 configmap.go:193] Couldn't get configMap openshift-authentication-operator/service-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.476451 4943 configmap.go:193] Couldn't get configMap openshift-controller-manager-operator/openshift-controller-manager-operator-config: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.476474 4943 secret.go:188] Couldn't get secret openshift-apiserver-operator/openshift-apiserver-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.476499 4943 configmap.go:193] Couldn't get configMap openshift-controller-manager/openshift-global-ca: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.476528 4943 secret.go:188] Couldn't get secret openshift-authentication-operator/serving-cert: failed to sync secret cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.476504 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b541cfee-3ec5-4be3-9acd-d58ebb79f0a2-service-ca-bundle podName:b541cfee-3ec5-4be3-9acd-d58ebb79f0a2 nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.97648945 +0000 UTC m=+122.928625978 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "service-ca-bundle" (UniqueName: "kubernetes.io/configmap/b541cfee-3ec5-4be3-9acd-d58ebb79f0a2-service-ca-bundle") pod "authentication-operator-69f744f599-ltd8c" (UID: "b541cfee-3ec5-4be3-9acd-d58ebb79f0a2") : failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.476594 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/00cc0b59-be8e-402b-b6c9-56e9e20c4b9b-config podName:00cc0b59-be8e-402b-b6c9-56e9e20c4b9b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.976575212 +0000 UTC m=+122.928711750 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/00cc0b59-be8e-402b-b6c9-56e9e20c4b9b-config") pod "openshift-controller-manager-operator-756b6f6bc6-x5qjf" (UID: "00cc0b59-be8e-402b-b6c9-56e9e20c4b9b") : failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.476618 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0da2f3db-9ac4-42bf-b4b8-a4d538d9f4e6-serving-cert podName:0da2f3db-9ac4-42bf-b4b8-a4d538d9f4e6 nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.976606173 +0000 UTC m=+122.928742701 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/0da2f3db-9ac4-42bf-b4b8-a4d538d9f4e6-serving-cert") pod "openshift-apiserver-operator-796bbdcf4f-dvjg6" (UID: "0da2f3db-9ac4-42bf-b4b8-a4d538d9f4e6") : failed to sync secret cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.476651 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/a04d88f1-a308-42cf-9462-1a06f6338b7b-proxy-ca-bundles podName:a04d88f1-a308-42cf-9462-1a06f6338b7b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.976640594 +0000 UTC m=+122.928777132 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "proxy-ca-bundles" (UniqueName: "kubernetes.io/configmap/a04d88f1-a308-42cf-9462-1a06f6338b7b-proxy-ca-bundles") pod "controller-manager-879f6c89f-n2mhr" (UID: "a04d88f1-a308-42cf-9462-1a06f6338b7b") : failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.476677 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b541cfee-3ec5-4be3-9acd-d58ebb79f0a2-serving-cert podName:b541cfee-3ec5-4be3-9acd-d58ebb79f0a2 nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.976666965 +0000 UTC m=+122.928803503 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/b541cfee-3ec5-4be3-9acd-d58ebb79f0a2-serving-cert") pod "authentication-operator-69f744f599-ltd8c" (UID: "b541cfee-3ec5-4be3-9acd-d58ebb79f0a2") : failed to sync secret cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.477669 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-trusted-ca-bundle podName:31ce0220-0f53-4d63-aa7b-99357e9c6b04 nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.97764124 +0000 UTC m=+122.929777808 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-system-trusted-ca-bundle" (UniqueName: "kubernetes.io/configmap/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-trusted-ca-bundle") pod "oauth-openshift-558db77b4-cnbpx" (UID: "31ce0220-0f53-4d63-aa7b-99357e9c6b04") : failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.477732 4943 configmap.go:193] Couldn't get configMap openshift-ingress-operator/trusted-ca: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.477792 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/80ba5362-78a1-4ed1-a6be-fbda0dbe72df-trusted-ca podName:80ba5362-78a1-4ed1-a6be-fbda0dbe72df nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.977775484 +0000 UTC m=+122.929912012 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "trusted-ca" (UniqueName: "kubernetes.io/configmap/80ba5362-78a1-4ed1-a6be-fbda0dbe72df-trusted-ca") pod "ingress-operator-5b745b69d9-8x4b2" (UID: "80ba5362-78a1-4ed1-a6be-fbda0dbe72df") : failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.477795 4943 configmap.go:193] Couldn't get configMap openshift-route-controller-manager/config: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.477832 4943 configmap.go:193] Couldn't get configMap openshift-console-operator/console-operator-config: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.477872 4943 secret.go:188] Couldn't get secret openshift-ingress-operator/metrics-tls: failed to sync secret cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.477898 4943 secret.go:188] Couldn't get secret openshift-route-controller-manager/serving-cert: failed to sync secret cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.477842 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/51fb375a-53bd-420a-930f-6b6cdb69f793-config podName:51fb375a-53bd-420a-930f-6b6cdb69f793 nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.977830955 +0000 UTC m=+122.929967483 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/51fb375a-53bd-420a-930f-6b6cdb69f793-config") pod "route-controller-manager-6576b87f9c-6qmrb" (UID: "51fb375a-53bd-420a-930f-6b6cdb69f793") : failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.478035 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b7960fae-df40-482f-a9ab-790fa11aaf8f-config podName:b7960fae-df40-482f-a9ab-790fa11aaf8f nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.977961778 +0000 UTC m=+122.930098336 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/b7960fae-df40-482f-a9ab-790fa11aaf8f-config") pod "console-operator-58897d9998-lg6mk" (UID: "b7960fae-df40-482f-a9ab-790fa11aaf8f") : failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.478068 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/80ba5362-78a1-4ed1-a6be-fbda0dbe72df-metrics-tls podName:80ba5362-78a1-4ed1-a6be-fbda0dbe72df nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.978055011 +0000 UTC m=+122.930191549 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/80ba5362-78a1-4ed1-a6be-fbda0dbe72df-metrics-tls") pod "ingress-operator-5b745b69d9-8x4b2" (UID: "80ba5362-78a1-4ed1-a6be-fbda0dbe72df") : failed to sync secret cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.478097 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/51fb375a-53bd-420a-930f-6b6cdb69f793-serving-cert podName:51fb375a-53bd-420a-930f-6b6cdb69f793 nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.978081912 +0000 UTC m=+122.930218540 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/51fb375a-53bd-420a-930f-6b6cdb69f793-serving-cert") pod "route-controller-manager-6576b87f9c-6qmrb" (UID: "51fb375a-53bd-420a-930f-6b6cdb69f793") : failed to sync secret cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.478333 4943 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-system-session: failed to sync secret cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.478405 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-session podName:31ce0220-0f53-4d63-aa7b-99357e9c6b04 nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.97838869 +0000 UTC m=+122.930525218 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-system-session" (UniqueName: "kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-session") pod "oauth-openshift-558db77b4-cnbpx" (UID: "31ce0220-0f53-4d63-aa7b-99357e9c6b04") : failed to sync secret cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.478498 4943 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-system-ocp-branding-template: failed to sync secret cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.478545 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-ocp-branding-template podName:31ce0220-0f53-4d63-aa7b-99357e9c6b04 nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.978532643 +0000 UTC m=+122.930669181 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-system-ocp-branding-template" (UniqueName: "kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-ocp-branding-template") pod "oauth-openshift-558db77b4-cnbpx" (UID: "31ce0220-0f53-4d63-aa7b-99357e9c6b04") : failed to sync secret cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.478565 4943 configmap.go:193] Couldn't get configMap openshift-authentication-operator/authentication-operator-config: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.478656 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b541cfee-3ec5-4be3-9acd-d58ebb79f0a2-config podName:b541cfee-3ec5-4be3-9acd-d58ebb79f0a2 nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.978634126 +0000 UTC m=+122.930770714 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/b541cfee-3ec5-4be3-9acd-d58ebb79f0a2-config") pod "authentication-operator-69f744f599-ltd8c" (UID: "b541cfee-3ec5-4be3-9acd-d58ebb79f0a2") : failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.478572 4943 secret.go:188] Couldn't get secret openshift-cluster-machine-approver/machine-approver-tls: failed to sync secret cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.478733 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f94ec039-60b3-4ecb-b784-04a7ebaf4848-machine-approver-tls podName:f94ec039-60b3-4ecb-b784-04a7ebaf4848 nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.978715768 +0000 UTC m=+122.930852506 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "machine-approver-tls" (UniqueName: "kubernetes.io/secret/f94ec039-60b3-4ecb-b784-04a7ebaf4848-machine-approver-tls") pod "machine-approver-56656f9798-qml2f" (UID: "f94ec039-60b3-4ecb-b784-04a7ebaf4848") : failed to sync secret cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.478758 4943 secret.go:188] Couldn't get secret openshift-console/console-oauth-config: failed to sync secret cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.478831 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/698827d2-21c9-4856-a9cd-17bb3df88bce-console-oauth-config podName:698827d2-21c9-4856-a9cd-17bb3df88bce nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.97880852 +0000 UTC m=+122.930945048 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "console-oauth-config" (UniqueName: "kubernetes.io/secret/698827d2-21c9-4856-a9cd-17bb3df88bce-console-oauth-config") pod "console-f9d7485db-zgx7p" (UID: "698827d2-21c9-4856-a9cd-17bb3df88bce") : failed to sync secret cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.478905 4943 configmap.go:193] Couldn't get configMap openshift-console/service-ca: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.479016 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/698827d2-21c9-4856-a9cd-17bb3df88bce-service-ca podName:698827d2-21c9-4856-a9cd-17bb3df88bce nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.978996465 +0000 UTC m=+122.931132993 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "service-ca" (UniqueName: "kubernetes.io/configmap/698827d2-21c9-4856-a9cd-17bb3df88bce-service-ca") pod "console-f9d7485db-zgx7p" (UID: "698827d2-21c9-4856-a9cd-17bb3df88bce") : failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.479065 4943 configmap.go:193] Couldn't get configMap openshift-cluster-machine-approver/kube-rbac-proxy: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.479114 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/f94ec039-60b3-4ecb-b784-04a7ebaf4848-auth-proxy-config podName:f94ec039-60b3-4ecb-b784-04a7ebaf4848 nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.979100758 +0000 UTC m=+122.931237296 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "auth-proxy-config" (UniqueName: "kubernetes.io/configmap/f94ec039-60b3-4ecb-b784-04a7ebaf4848-auth-proxy-config") pod "machine-approver-56656f9798-qml2f" (UID: "f94ec039-60b3-4ecb-b784-04a7ebaf4848") : failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.479340 4943 configmap.go:193] Couldn't get configMap openshift-route-controller-manager/client-ca: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.479401 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/51fb375a-53bd-420a-930f-6b6cdb69f793-client-ca podName:51fb375a-53bd-420a-930f-6b6cdb69f793 nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.979386155 +0000 UTC m=+122.931522693 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "client-ca" (UniqueName: "kubernetes.io/configmap/51fb375a-53bd-420a-930f-6b6cdb69f793-client-ca") pod "route-controller-manager-6576b87f9c-6qmrb" (UID: "51fb375a-53bd-420a-930f-6b6cdb69f793") : failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.479460 4943 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-user-template-login: failed to sync secret cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.479504 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-user-template-login podName:31ce0220-0f53-4d63-aa7b-99357e9c6b04 nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.979492108 +0000 UTC m=+122.931628636 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-user-template-login" (UniqueName: "kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-user-template-login") pod "oauth-openshift-558db77b4-cnbpx" (UID: "31ce0220-0f53-4d63-aa7b-99357e9c6b04") : failed to sync secret cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.479758 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.479812 4943 configmap.go:193] Couldn't get configMap openshift-authentication/v4-0-config-system-cliconfig: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.479881 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-cliconfig podName:31ce0220-0f53-4d63-aa7b-99357e9c6b04 nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.979861998 +0000 UTC m=+122.931998526 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-system-cliconfig" (UniqueName: "kubernetes.io/configmap/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-cliconfig") pod "oauth-openshift-558db77b4-cnbpx" (UID: "31ce0220-0f53-4d63-aa7b-99357e9c6b04") : failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.479951 4943 configmap.go:193] Couldn't get configMap openshift-authentication/audit: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.479999 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/31ce0220-0f53-4d63-aa7b-99357e9c6b04-audit-policies podName:31ce0220-0f53-4d63-aa7b-99357e9c6b04 nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.979986181 +0000 UTC m=+122.932122709 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "audit-policies" (UniqueName: "kubernetes.io/configmap/31ce0220-0f53-4d63-aa7b-99357e9c6b04-audit-policies") pod "oauth-openshift-558db77b4-cnbpx" (UID: "31ce0220-0f53-4d63-aa7b-99357e9c6b04") : failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.480021 4943 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-system-serving-cert: failed to sync secret cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.480097 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-serving-cert podName:31ce0220-0f53-4d63-aa7b-99357e9c6b04 nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.980079933 +0000 UTC m=+122.932216471 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-system-serving-cert" (UniqueName: "kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-serving-cert") pod "oauth-openshift-558db77b4-cnbpx" (UID: "31ce0220-0f53-4d63-aa7b-99357e9c6b04") : failed to sync secret cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.480158 4943 configmap.go:193] Couldn't get configMap openshift-controller-manager/client-ca: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.480218 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/a04d88f1-a308-42cf-9462-1a06f6338b7b-client-ca podName:a04d88f1-a308-42cf-9462-1a06f6338b7b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.980203757 +0000 UTC m=+122.932340285 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "client-ca" (UniqueName: "kubernetes.io/configmap/a04d88f1-a308-42cf-9462-1a06f6338b7b-client-ca") pod "controller-manager-879f6c89f-n2mhr" (UID: "a04d88f1-a308-42cf-9462-1a06f6338b7b") : failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.480268 4943 configmap.go:193] Couldn't get configMap openshift-apiserver-operator/openshift-apiserver-operator-config: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.480314 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0da2f3db-9ac4-42bf-b4b8-a4d538d9f4e6-config podName:0da2f3db-9ac4-42bf-b4b8-a4d538d9f4e6 nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.980303649 +0000 UTC m=+122.932440177 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/0da2f3db-9ac4-42bf-b4b8-a4d538d9f4e6-config") pod "openshift-apiserver-operator-796bbdcf4f-dvjg6" (UID: "0da2f3db-9ac4-42bf-b4b8-a4d538d9f4e6") : failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.480570 4943 configmap.go:193] Couldn't get configMap openshift-authentication/v4-0-config-system-service-ca: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.480629 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-service-ca podName:31ce0220-0f53-4d63-aa7b-99357e9c6b04 nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.980615407 +0000 UTC m=+122.932751945 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-system-service-ca" (UniqueName: "kubernetes.io/configmap/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-service-ca") pod "oauth-openshift-558db77b4-cnbpx" (UID: "31ce0220-0f53-4d63-aa7b-99357e9c6b04") : failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.480671 4943 configmap.go:193] Couldn't get configMap openshift-controller-manager/config: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.480674 4943 configmap.go:193] Couldn't get configMap openshift-authentication-operator/trusted-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.480727 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b541cfee-3ec5-4be3-9acd-d58ebb79f0a2-trusted-ca-bundle podName:b541cfee-3ec5-4be3-9acd-d58ebb79f0a2 nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.9807138 +0000 UTC m=+122.932850328 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "trusted-ca-bundle" (UniqueName: "kubernetes.io/configmap/b541cfee-3ec5-4be3-9acd-d58ebb79f0a2-trusted-ca-bundle") pod "authentication-operator-69f744f599-ltd8c" (UID: "b541cfee-3ec5-4be3-9acd-d58ebb79f0a2") : failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.480753 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/a04d88f1-a308-42cf-9462-1a06f6338b7b-config podName:a04d88f1-a308-42cf-9462-1a06f6338b7b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.980739451 +0000 UTC m=+122.932875989 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/a04d88f1-a308-42cf-9462-1a06f6338b7b-config") pod "controller-manager-879f6c89f-n2mhr" (UID: "a04d88f1-a308-42cf-9462-1a06f6338b7b") : failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.481709 4943 secret.go:188] Couldn't get secret openshift-image-registry/installation-pull-secrets: failed to sync secret cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.481776 4943 configmap.go:193] Couldn't get configMap openshift-console/console-config: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.481842 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/698827d2-21c9-4856-a9cd-17bb3df88bce-console-config podName:698827d2-21c9-4856-a9cd-17bb3df88bce nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.981822499 +0000 UTC m=+122.933959037 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "console-config" (UniqueName: "kubernetes.io/configmap/698827d2-21c9-4856-a9cd-17bb3df88bce-console-config") pod "console-f9d7485db-zgx7p" (UID: "698827d2-21c9-4856-a9cd-17bb3df88bce") : failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.481869 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a9fcbefd-dfa7-4afd-b198-d872137a9f51-installation-pull-secrets podName:a9fcbefd-dfa7-4afd-b198-d872137a9f51 nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.98185695 +0000 UTC m=+122.933993488 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "installation-pull-secrets" (UniqueName: "kubernetes.io/secret/a9fcbefd-dfa7-4afd-b198-d872137a9f51-installation-pull-secrets") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : failed to sync secret cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.482804 4943 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-user-template-provider-selection: failed to sync secret cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.482880 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-user-template-provider-selection podName:31ce0220-0f53-4d63-aa7b-99357e9c6b04 nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.982861156 +0000 UTC m=+122.934997684 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-user-template-provider-selection" (UniqueName: "kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-user-template-provider-selection") pod "oauth-openshift-558db77b4-cnbpx" (UID: "31ce0220-0f53-4d63-aa7b-99357e9c6b04") : failed to sync secret cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.482973 4943 configmap.go:193] Couldn't get configMap openshift-console-operator/trusted-ca: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.483061 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b7960fae-df40-482f-a9ab-790fa11aaf8f-trusted-ca podName:b7960fae-df40-482f-a9ab-790fa11aaf8f nodeName:}" failed. No retries permitted until 2026-03-07 14:41:20.98303722 +0000 UTC m=+122.935173768 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "trusted-ca" (UniqueName: "kubernetes.io/configmap/b7960fae-df40-482f-a9ab-790fa11aaf8f-trusted-ca") pod "console-operator-58897d9998-lg6mk" (UID: "b7960fae-df40-482f-a9ab-790fa11aaf8f") : failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.499190 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.511876 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/986041cd-e5ea-4269-82b0-5799d343b155-cert\") pod \"ingress-canary-w286n\" (UID: \"986041cd-e5ea-4269-82b0-5799d343b155\") " pod="openshift-ingress-canary/ingress-canary-w286n" Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.516422 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.516601 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:21.01656791 +0000 UTC m=+122.968704438 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.517907 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.519476 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:21.019426274 +0000 UTC m=+122.971562812 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.522002 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.537805 4943 request.go:700] Waited for 1.863169787s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-ingress-canary/configmaps?fieldSelector=metadata.name%3Dopenshift-service-ca.crt&limit=500&resourceVersion=0 Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.540072 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.584671 4943 configmap.go:193] Couldn't get configMap openshift-image-registry/trusted-ca: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.585056 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/df59699d-5df8-401e-8bb0-0afa130ac06b-trusted-ca podName:df59699d-5df8-401e-8bb0-0afa130ac06b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:21.085015616 +0000 UTC m=+123.037152154 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "trusted-ca" (UniqueName: "kubernetes.io/configmap/df59699d-5df8-401e-8bb0-0afa130ac06b-trusted-ca") pod "cluster-image-registry-operator-dc59b4c8b-lk96z" (UID: "df59699d-5df8-401e-8bb0-0afa130ac06b") : failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.588015 4943 secret.go:188] Couldn't get secret openshift-image-registry/image-registry-operator-tls: failed to sync secret cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.588136 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/df59699d-5df8-401e-8bb0-0afa130ac06b-image-registry-operator-tls podName:df59699d-5df8-401e-8bb0-0afa130ac06b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:21.088106406 +0000 UTC m=+123.040242934 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "image-registry-operator-tls" (UniqueName: "kubernetes.io/secret/df59699d-5df8-401e-8bb0-0afa130ac06b-image-registry-operator-tls") pod "cluster-image-registry-operator-dc59b4c8b-lk96z" (UID: "df59699d-5df8-401e-8bb0-0afa130ac06b") : failed to sync secret cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.592241 4943 configmap.go:193] Couldn't get configMap openshift-kube-apiserver-operator/kube-apiserver-operator-config: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.592336 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b33c7fd7-2d0f-48b7-9932-e7740577280f-config podName:b33c7fd7-2d0f-48b7-9932-e7740577280f nodeName:}" failed. No retries permitted until 2026-03-07 14:41:21.092312845 +0000 UTC m=+123.044449373 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/b33c7fd7-2d0f-48b7-9932-e7740577280f-config") pod "kube-apiserver-operator-766d6c64bb-9gzdv" (UID: "b33c7fd7-2d0f-48b7-9932-e7740577280f") : failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.600700 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.622206 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.623148 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:21.123108114 +0000 UTC m=+123.075244652 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.649383 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.649596 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lh8b9\" (UniqueName: \"kubernetes.io/projected/a9fcbefd-dfa7-4afd-b198-d872137a9f51-kube-api-access-lh8b9\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.669295 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.679518 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.698857 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.720083 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.726093 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.726922 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:21.226893627 +0000 UTC m=+123.179030165 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.740689 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.769534 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.819775 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.829695 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.830008 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:21.329973691 +0000 UTC m=+123.282110229 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.830395 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.831273 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:21.331228924 +0000 UTC m=+123.283365452 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.839684 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.858974 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.880007 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.932479 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.932704 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:21.432673595 +0000 UTC m=+123.384810133 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.933213 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:20 crc kubenswrapper[4943]: E0307 14:41:20.933738 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:21.433712512 +0000 UTC m=+123.385849130 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.939226 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.959571 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Mar 07 14:41:20 crc kubenswrapper[4943]: I0307 14:41:20.978823 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.010730 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.019704 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.058681 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:21 crc kubenswrapper[4943]: E0307 14:41:21.059332 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:21.55929431 +0000 UTC m=+123.511430848 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.059451 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f94ec039-60b3-4ecb-b784-04a7ebaf4848-config\") pod \"machine-approver-56656f9798-qml2f\" (UID: \"f94ec039-60b3-4ecb-b784-04a7ebaf4848\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qml2f" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.059524 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/00cc0b59-be8e-402b-b6c9-56e9e20c4b9b-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-x5qjf\" (UID: \"00cc0b59-be8e-402b-b6c9-56e9e20c4b9b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-x5qjf" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.059570 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00cc0b59-be8e-402b-b6c9-56e9e20c4b9b-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-x5qjf\" (UID: \"00cc0b59-be8e-402b-b6c9-56e9e20c4b9b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-x5qjf" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.059610 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0da2f3db-9ac4-42bf-b4b8-a4d538d9f4e6-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-dvjg6\" (UID: \"0da2f3db-9ac4-42bf-b4b8-a4d538d9f4e6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dvjg6" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.059645 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a04d88f1-a308-42cf-9462-1a06f6338b7b-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-n2mhr\" (UID: \"a04d88f1-a308-42cf-9462-1a06f6338b7b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n2mhr" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.059682 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b541cfee-3ec5-4be3-9acd-d58ebb79f0a2-service-ca-bundle\") pod \"authentication-operator-69f744f599-ltd8c\" (UID: \"b541cfee-3ec5-4be3-9acd-d58ebb79f0a2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ltd8c" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.059759 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b541cfee-3ec5-4be3-9acd-d58ebb79f0a2-serving-cert\") pod \"authentication-operator-69f744f599-ltd8c\" (UID: \"b541cfee-3ec5-4be3-9acd-d58ebb79f0a2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ltd8c" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.059795 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/80ba5362-78a1-4ed1-a6be-fbda0dbe72df-metrics-tls\") pod \"ingress-operator-5b745b69d9-8x4b2\" (UID: \"80ba5362-78a1-4ed1-a6be-fbda0dbe72df\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8x4b2" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.059829 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/80ba5362-78a1-4ed1-a6be-fbda0dbe72df-trusted-ca\") pod \"ingress-operator-5b745b69d9-8x4b2\" (UID: \"80ba5362-78a1-4ed1-a6be-fbda0dbe72df\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8x4b2" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.059905 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51fb375a-53bd-420a-930f-6b6cdb69f793-config\") pod \"route-controller-manager-6576b87f9c-6qmrb\" (UID: \"51fb375a-53bd-420a-930f-6b6cdb69f793\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6qmrb" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.060001 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/51fb375a-53bd-420a-930f-6b6cdb69f793-serving-cert\") pod \"route-controller-manager-6576b87f9c-6qmrb\" (UID: \"51fb375a-53bd-420a-930f-6b6cdb69f793\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6qmrb" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.060035 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7960fae-df40-482f-a9ab-790fa11aaf8f-config\") pod \"console-operator-58897d9998-lg6mk\" (UID: \"b7960fae-df40-482f-a9ab-790fa11aaf8f\") " pod="openshift-console-operator/console-operator-58897d9998-lg6mk" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.060149 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/f94ec039-60b3-4ecb-b784-04a7ebaf4848-machine-approver-tls\") pod \"machine-approver-56656f9798-qml2f\" (UID: \"f94ec039-60b3-4ecb-b784-04a7ebaf4848\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qml2f" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.060187 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-cnbpx\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.060254 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-cnbpx\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.060288 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b541cfee-3ec5-4be3-9acd-d58ebb79f0a2-config\") pod \"authentication-operator-69f744f599-ltd8c\" (UID: \"b541cfee-3ec5-4be3-9acd-d58ebb79f0a2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ltd8c" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.060332 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/698827d2-21c9-4856-a9cd-17bb3df88bce-console-oauth-config\") pod \"console-f9d7485db-zgx7p\" (UID: \"698827d2-21c9-4856-a9cd-17bb3df88bce\") " pod="openshift-console/console-f9d7485db-zgx7p" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.060406 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/698827d2-21c9-4856-a9cd-17bb3df88bce-service-ca\") pod \"console-f9d7485db-zgx7p\" (UID: \"698827d2-21c9-4856-a9cd-17bb3df88bce\") " pod="openshift-console/console-f9d7485db-zgx7p" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.060440 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f94ec039-60b3-4ecb-b784-04a7ebaf4848-auth-proxy-config\") pod \"machine-approver-56656f9798-qml2f\" (UID: \"f94ec039-60b3-4ecb-b784-04a7ebaf4848\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qml2f" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.060477 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/51fb375a-53bd-420a-930f-6b6cdb69f793-client-ca\") pod \"route-controller-manager-6576b87f9c-6qmrb\" (UID: \"51fb375a-53bd-420a-930f-6b6cdb69f793\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6qmrb" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.060513 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-cnbpx\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.060569 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-cnbpx\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.060692 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f94ec039-60b3-4ecb-b784-04a7ebaf4848-config\") pod \"machine-approver-56656f9798-qml2f\" (UID: \"f94ec039-60b3-4ecb-b784-04a7ebaf4848\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qml2f" Mar 07 14:41:21 crc kubenswrapper[4943]: E0307 14:41:21.060831 4943 projected.go:288] Couldn't get configMap openshift-apiserver/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.060628 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/31ce0220-0f53-4d63-aa7b-99357e9c6b04-audit-policies\") pod \"oauth-openshift-558db77b4-cnbpx\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.061197 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-cnbpx\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.061317 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a04d88f1-a308-42cf-9462-1a06f6338b7b-client-ca\") pod \"controller-manager-879f6c89f-n2mhr\" (UID: \"a04d88f1-a308-42cf-9462-1a06f6338b7b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n2mhr" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.061393 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0da2f3db-9ac4-42bf-b4b8-a4d538d9f4e6-config\") pod \"openshift-apiserver-operator-796bbdcf4f-dvjg6\" (UID: \"0da2f3db-9ac4-42bf-b4b8-a4d538d9f4e6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dvjg6" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.061434 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-cnbpx\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.061509 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b541cfee-3ec5-4be3-9acd-d58ebb79f0a2-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-ltd8c\" (UID: \"b541cfee-3ec5-4be3-9acd-d58ebb79f0a2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ltd8c" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.061550 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00cc0b59-be8e-402b-b6c9-56e9e20c4b9b-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-x5qjf\" (UID: \"00cc0b59-be8e-402b-b6c9-56e9e20c4b9b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-x5qjf" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.061582 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a04d88f1-a308-42cf-9462-1a06f6338b7b-config\") pod \"controller-manager-879f6c89f-n2mhr\" (UID: \"a04d88f1-a308-42cf-9462-1a06f6338b7b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n2mhr" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.061638 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.061675 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a9fcbefd-dfa7-4afd-b198-d872137a9f51-installation-pull-secrets\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.061735 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b541cfee-3ec5-4be3-9acd-d58ebb79f0a2-service-ca-bundle\") pod \"authentication-operator-69f744f599-ltd8c\" (UID: \"b541cfee-3ec5-4be3-9acd-d58ebb79f0a2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ltd8c" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.061804 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/698827d2-21c9-4856-a9cd-17bb3df88bce-console-config\") pod \"console-f9d7485db-zgx7p\" (UID: \"698827d2-21c9-4856-a9cd-17bb3df88bce\") " pod="openshift-console/console-f9d7485db-zgx7p" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.061867 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.061909 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b7960fae-df40-482f-a9ab-790fa11aaf8f-trusted-ca\") pod \"console-operator-58897d9998-lg6mk\" (UID: \"b7960fae-df40-482f-a9ab-790fa11aaf8f\") " pod="openshift-console-operator/console-operator-58897d9998-lg6mk" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.062012 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-cnbpx\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.062119 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/28995414-2bd3-4b50-90ec-0fbb63a15ef6-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-dr8bw\" (UID: \"28995414-2bd3-4b50-90ec-0fbb63a15ef6\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dr8bw" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.062188 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/698827d2-21c9-4856-a9cd-17bb3df88bce-trusted-ca-bundle\") pod \"console-f9d7485db-zgx7p\" (UID: \"698827d2-21c9-4856-a9cd-17bb3df88bce\") " pod="openshift-console/console-f9d7485db-zgx7p" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.062269 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a9fcbefd-dfa7-4afd-b198-d872137a9f51-trusted-ca\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.062306 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-cnbpx\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:21 crc kubenswrapper[4943]: E0307 14:41:21.062345 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:21.562321709 +0000 UTC m=+123.514458237 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.065217 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a04d88f1-a308-42cf-9462-1a06f6338b7b-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-n2mhr\" (UID: \"a04d88f1-a308-42cf-9462-1a06f6338b7b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n2mhr" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.065651 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/698827d2-21c9-4856-a9cd-17bb3df88bce-trusted-ca-bundle\") pod \"console-f9d7485db-zgx7p\" (UID: \"698827d2-21c9-4856-a9cd-17bb3df88bce\") " pod="openshift-console/console-f9d7485db-zgx7p" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.062419 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-cnbpx\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.065917 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/698827d2-21c9-4856-a9cd-17bb3df88bce-oauth-serving-cert\") pod \"console-f9d7485db-zgx7p\" (UID: \"698827d2-21c9-4856-a9cd-17bb3df88bce\") " pod="openshift-console/console-f9d7485db-zgx7p" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.066091 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b7960fae-df40-482f-a9ab-790fa11aaf8f-serving-cert\") pod \"console-operator-58897d9998-lg6mk\" (UID: \"b7960fae-df40-482f-a9ab-790fa11aaf8f\") " pod="openshift-console-operator/console-operator-58897d9998-lg6mk" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.066369 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-cnbpx\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.066568 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a04d88f1-a308-42cf-9462-1a06f6338b7b-serving-cert\") pod \"controller-manager-879f6c89f-n2mhr\" (UID: \"a04d88f1-a308-42cf-9462-1a06f6338b7b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n2mhr" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.066668 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a9fcbefd-dfa7-4afd-b198-d872137a9f51-registry-tls\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.067072 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-cnbpx\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.067232 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/698827d2-21c9-4856-a9cd-17bb3df88bce-console-serving-cert\") pod \"console-f9d7485db-zgx7p\" (UID: \"698827d2-21c9-4856-a9cd-17bb3df88bce\") " pod="openshift-console/console-f9d7485db-zgx7p" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.068290 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a9fcbefd-dfa7-4afd-b198-d872137a9f51-trusted-ca\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.071497 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-cnbpx\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.073858 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/698827d2-21c9-4856-a9cd-17bb3df88bce-oauth-serving-cert\") pod \"console-f9d7485db-zgx7p\" (UID: \"698827d2-21c9-4856-a9cd-17bb3df88bce\") " pod="openshift-console/console-f9d7485db-zgx7p" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.074820 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a04d88f1-a308-42cf-9462-1a06f6338b7b-serving-cert\") pod \"controller-manager-879f6c89f-n2mhr\" (UID: \"a04d88f1-a308-42cf-9462-1a06f6338b7b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n2mhr" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.074888 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/698827d2-21c9-4856-a9cd-17bb3df88bce-console-serving-cert\") pod \"console-f9d7485db-zgx7p\" (UID: \"698827d2-21c9-4856-a9cd-17bb3df88bce\") " pod="openshift-console/console-f9d7485db-zgx7p" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.076876 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b7960fae-df40-482f-a9ab-790fa11aaf8f-serving-cert\") pod \"console-operator-58897d9998-lg6mk\" (UID: \"b7960fae-df40-482f-a9ab-790fa11aaf8f\") " pod="openshift-console-operator/console-operator-58897d9998-lg6mk" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.078031 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a9fcbefd-dfa7-4afd-b198-d872137a9f51-registry-tls\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:21 crc kubenswrapper[4943]: E0307 14:41:21.078620 4943 projected.go:288] Couldn't get configMap openshift-oauth-apiserver/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.079697 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-cnbpx\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.082264 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-cnbpx\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.085391 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-cnbpx\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.085514 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/00cc0b59-be8e-402b-b6c9-56e9e20c4b9b-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-x5qjf\" (UID: \"00cc0b59-be8e-402b-b6c9-56e9e20c4b9b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-x5qjf" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.085626 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/28995414-2bd3-4b50-90ec-0fbb63a15ef6-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-dr8bw\" (UID: \"28995414-2bd3-4b50-90ec-0fbb63a15ef6\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dr8bw" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.088085 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0da2f3db-9ac4-42bf-b4b8-a4d538d9f4e6-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-dvjg6\" (UID: \"0da2f3db-9ac4-42bf-b4b8-a4d538d9f4e6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dvjg6" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.100197 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.105576 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b541cfee-3ec5-4be3-9acd-d58ebb79f0a2-serving-cert\") pod \"authentication-operator-69f744f599-ltd8c\" (UID: \"b541cfee-3ec5-4be3-9acd-d58ebb79f0a2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ltd8c" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.129704 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.132746 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/80ba5362-78a1-4ed1-a6be-fbda0dbe72df-trusted-ca\") pod \"ingress-operator-5b745b69d9-8x4b2\" (UID: \"80ba5362-78a1-4ed1-a6be-fbda0dbe72df\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8x4b2" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.139553 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.147100 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/80ba5362-78a1-4ed1-a6be-fbda0dbe72df-metrics-tls\") pod \"ingress-operator-5b745b69d9-8x4b2\" (UID: \"80ba5362-78a1-4ed1-a6be-fbda0dbe72df\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8x4b2" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.168544 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:21 crc kubenswrapper[4943]: E0307 14:41:21.168738 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:21.668708409 +0000 UTC m=+123.620844937 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.169188 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/df59699d-5df8-401e-8bb0-0afa130ac06b-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-lk96z\" (UID: \"df59699d-5df8-401e-8bb0-0afa130ac06b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lk96z" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.169298 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.169455 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b33c7fd7-2d0f-48b7-9932-e7740577280f-config\") pod \"kube-apiserver-operator-766d6c64bb-9gzdv\" (UID: \"b33c7fd7-2d0f-48b7-9932-e7740577280f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9gzdv" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.169538 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/df59699d-5df8-401e-8bb0-0afa130ac06b-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-lk96z\" (UID: \"df59699d-5df8-401e-8bb0-0afa130ac06b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lk96z" Mar 07 14:41:21 crc kubenswrapper[4943]: E0307 14:41:21.169690 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:21.669673754 +0000 UTC m=+123.621810292 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.171570 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/df59699d-5df8-401e-8bb0-0afa130ac06b-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-lk96z\" (UID: \"df59699d-5df8-401e-8bb0-0afa130ac06b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lk96z" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.178826 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.183535 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51fb375a-53bd-420a-930f-6b6cdb69f793-config\") pod \"route-controller-manager-6576b87f9c-6qmrb\" (UID: \"51fb375a-53bd-420a-930f-6b6cdb69f793\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6qmrb" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.186815 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/80ba5362-78a1-4ed1-a6be-fbda0dbe72df-bound-sa-token\") pod \"ingress-operator-5b745b69d9-8x4b2\" (UID: \"80ba5362-78a1-4ed1-a6be-fbda0dbe72df\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8x4b2" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.219815 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.226759 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/51fb375a-53bd-420a-930f-6b6cdb69f793-serving-cert\") pod \"route-controller-manager-6576b87f9c-6qmrb\" (UID: \"51fb375a-53bd-420a-930f-6b6cdb69f793\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6qmrb" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.239974 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.242609 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7960fae-df40-482f-a9ab-790fa11aaf8f-config\") pod \"console-operator-58897d9998-lg6mk\" (UID: \"b7960fae-df40-482f-a9ab-790fa11aaf8f\") " pod="openshift-console-operator/console-operator-58897d9998-lg6mk" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.271275 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:21 crc kubenswrapper[4943]: E0307 14:41:21.271575 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:21.771542837 +0000 UTC m=+123.723679365 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.272011 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:21 crc kubenswrapper[4943]: E0307 14:41:21.272554 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:21.772530542 +0000 UTC m=+123.724667070 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:21 crc kubenswrapper[4943]: E0307 14:41:21.281590 4943 projected.go:288] Couldn't get configMap openshift-machine-api/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.300045 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.306218 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/f94ec039-60b3-4ecb-b784-04a7ebaf4848-machine-approver-tls\") pod \"machine-approver-56656f9798-qml2f\" (UID: \"f94ec039-60b3-4ecb-b784-04a7ebaf4848\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qml2f" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.319445 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.325703 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-cnbpx\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.339702 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.342588 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b541cfee-3ec5-4be3-9acd-d58ebb79f0a2-config\") pod \"authentication-operator-69f744f599-ltd8c\" (UID: \"b541cfee-3ec5-4be3-9acd-d58ebb79f0a2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ltd8c" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.373722 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:21 crc kubenswrapper[4943]: E0307 14:41:21.374216 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:21.874171579 +0000 UTC m=+123.826308117 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.375406 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:21 crc kubenswrapper[4943]: E0307 14:41:21.376058 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:21.876028408 +0000 UTC m=+123.828164946 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.380306 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.380405 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.386860 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/698827d2-21c9-4856-a9cd-17bb3df88bce-console-oauth-config\") pod \"console-f9d7485db-zgx7p\" (UID: \"698827d2-21c9-4856-a9cd-17bb3df88bce\") " pod="openshift-console/console-f9d7485db-zgx7p" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.388350 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-cnbpx\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.420618 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.422031 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/698827d2-21c9-4856-a9cd-17bb3df88bce-service-ca\") pod \"console-f9d7485db-zgx7p\" (UID: \"698827d2-21c9-4856-a9cd-17bb3df88bce\") " pod="openshift-console/console-f9d7485db-zgx7p" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.440446 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.442591 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f94ec039-60b3-4ecb-b784-04a7ebaf4848-auth-proxy-config\") pod \"machine-approver-56656f9798-qml2f\" (UID: \"f94ec039-60b3-4ecb-b784-04a7ebaf4848\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qml2f" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.458911 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.463338 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/51fb375a-53bd-420a-930f-6b6cdb69f793-client-ca\") pod \"route-controller-manager-6576b87f9c-6qmrb\" (UID: \"51fb375a-53bd-420a-930f-6b6cdb69f793\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6qmrb" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.476329 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:21 crc kubenswrapper[4943]: E0307 14:41:21.476508 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:21.976474024 +0000 UTC m=+123.928610562 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.477534 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:21 crc kubenswrapper[4943]: E0307 14:41:21.478435 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:21.978403694 +0000 UTC m=+123.930540202 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.486521 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.496824 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-cnbpx\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.519394 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.519592 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvr2j\" (UniqueName: \"kubernetes.io/projected/08f9e92f-8419-4ba3-a532-c13e5bb6cab4-kube-api-access-fvr2j\") pod \"openshift-config-operator-7777fb866f-4vkx6\" (UID: \"08f9e92f-8419-4ba3-a532-c13e5bb6cab4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4vkx6" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.523130 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-cnbpx\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.539543 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.542295 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/31ce0220-0f53-4d63-aa7b-99357e9c6b04-audit-policies\") pod \"oauth-openshift-558db77b4-cnbpx\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.557637 4943 request.go:700] Waited for 2.077576281s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/secrets?fieldSelector=metadata.name%3Dv4-0-config-system-serving-cert&resourceVersion=27093 Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.560006 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.567586 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-cnbpx\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.578242 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:21 crc kubenswrapper[4943]: E0307 14:41:21.578547 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:22.078504441 +0000 UTC m=+124.030640969 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.578974 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:21 crc kubenswrapper[4943]: E0307 14:41:21.580113 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:22.080081242 +0000 UTC m=+124.032217820 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.580694 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 07 14:41:21 crc kubenswrapper[4943]: E0307 14:41:21.583151 4943 projected.go:288] Couldn't get configMap openshift-route-controller-manager/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.588651 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a04d88f1-a308-42cf-9462-1a06f6338b7b-client-ca\") pod \"controller-manager-879f6c89f-n2mhr\" (UID: \"a04d88f1-a308-42cf-9462-1a06f6338b7b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n2mhr" Mar 07 14:41:21 crc kubenswrapper[4943]: E0307 14:41:21.596875 4943 projected.go:288] Couldn't get configMap openshift-console/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.603204 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.612972 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0da2f3db-9ac4-42bf-b4b8-a4d538d9f4e6-config\") pod \"openshift-apiserver-operator-796bbdcf4f-dvjg6\" (UID: \"0da2f3db-9ac4-42bf-b4b8-a4d538d9f4e6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dvjg6" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.619199 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.623403 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-cnbpx\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.651382 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.653723 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b541cfee-3ec5-4be3-9acd-d58ebb79f0a2-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-ltd8c\" (UID: \"b541cfee-3ec5-4be3-9acd-d58ebb79f0a2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ltd8c" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.660193 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.663899 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a04d88f1-a308-42cf-9462-1a06f6338b7b-config\") pod \"controller-manager-879f6c89f-n2mhr\" (UID: \"a04d88f1-a308-42cf-9462-1a06f6338b7b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n2mhr" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.680055 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:21 crc kubenswrapper[4943]: E0307 14:41:21.680215 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:22.180178428 +0000 UTC m=+124.132314976 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.680225 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.680366 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:21 crc kubenswrapper[4943]: E0307 14:41:21.680918 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:22.180880697 +0000 UTC m=+124.133017235 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.687977 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a9fcbefd-dfa7-4afd-b198-d872137a9f51-installation-pull-secrets\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.720902 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.723984 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/698827d2-21c9-4856-a9cd-17bb3df88bce-console-config\") pod \"console-f9d7485db-zgx7p\" (UID: \"698827d2-21c9-4856-a9cd-17bb3df88bce\") " pod="openshift-console/console-f9d7485db-zgx7p" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.724078 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4vkx6" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.728456 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a9fcbefd-dfa7-4afd-b198-d872137a9f51-bound-sa-token\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.750821 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.754068 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b7960fae-df40-482f-a9ab-790fa11aaf8f-trusted-ca\") pod \"console-operator-58897d9998-lg6mk\" (UID: \"b7960fae-df40-482f-a9ab-790fa11aaf8f\") " pod="openshift-console-operator/console-operator-58897d9998-lg6mk" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.755020 4943 scope.go:117] "RemoveContainer" containerID="ed8d27f5a369a691493bda9dbd6e03c098931b8dab637bbd9bacac6fe7979c80" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.758858 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.768156 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-cnbpx\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.785201 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:21 crc kubenswrapper[4943]: E0307 14:41:21.786616 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:22.286569669 +0000 UTC m=+124.238706177 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.787151 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:21 crc kubenswrapper[4943]: E0307 14:41:21.787899 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:22.287880053 +0000 UTC m=+124.240016561 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:21 crc kubenswrapper[4943]: E0307 14:41:21.800760 4943 projected.go:288] Couldn't get configMap openshift-console-operator/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.810263 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e03f48ff-595f-4537-824f-650401604f7e-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-h2f4f\" (UID: \"e03f48ff-595f-4537-824f-650401604f7e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-h2f4f" Mar 07 14:41:21 crc kubenswrapper[4943]: E0307 14:41:21.820477 4943 projected.go:288] Couldn't get configMap openshift-cluster-samples-operator/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.832911 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqhfd\" (UniqueName: \"kubernetes.io/projected/8e19594a-0a69-40b7-8c24-317979edbfbb-kube-api-access-sqhfd\") pod \"catalog-operator-68c6474976-jpq82\" (UID: \"8e19594a-0a69-40b7-8c24-317979edbfbb\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jpq82" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.833365 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-h2f4f" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.840291 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgzhw\" (UniqueName: \"kubernetes.io/projected/0de7b0c4-4acf-40de-867d-1fc33069fa51-kube-api-access-dgzhw\") pod \"cni-sysctl-allowlist-ds-htv6d\" (UID: \"0de7b0c4-4acf-40de-867d-1fc33069fa51\") " pod="openshift-multus/cni-sysctl-allowlist-ds-htv6d" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.868602 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjwms\" (UniqueName: \"kubernetes.io/projected/2a45a1ae-558f-4600-b354-dea0169b175f-kube-api-access-gjwms\") pod \"dns-operator-744455d44c-j89vd\" (UID: \"2a45a1ae-558f-4600-b354-dea0169b175f\") " pod="openshift-dns-operator/dns-operator-744455d44c-j89vd" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.888167 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:21 crc kubenswrapper[4943]: E0307 14:41:21.888836 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:22.388817771 +0000 UTC m=+124.340954279 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.889159 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b33c7fd7-2d0f-48b7-9932-e7740577280f-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-9gzdv\" (UID: \"b33c7fd7-2d0f-48b7-9932-e7740577280f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9gzdv" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.894312 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x64d9\" (UniqueName: \"kubernetes.io/projected/8d97c7b9-9de3-40a6-ba65-db9c730604c0-kube-api-access-x64d9\") pod \"machine-config-operator-74547568cd-kdrbq\" (UID: \"8d97c7b9-9de3-40a6-ba65-db9c730604c0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kdrbq" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.899819 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-htv6d" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.919808 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tktfx\" (UniqueName: \"kubernetes.io/projected/f7170a81-b6f1-4696-9356-737721576cf6-kube-api-access-tktfx\") pod \"migrator-59844c95c7-7p4vl\" (UID: \"f7170a81-b6f1-4696-9356-737721576cf6\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7p4vl" Mar 07 14:41:21 crc kubenswrapper[4943]: E0307 14:41:21.921087 4943 projected.go:288] Couldn't get configMap openshift-authentication-operator/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.940036 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22gk4\" (UniqueName: \"kubernetes.io/projected/ff8cf71a-e914-4244-b1aa-ede95edd30e8-kube-api-access-22gk4\") pod \"csi-hostpathplugin-7p7x5\" (UID: \"ff8cf71a-e914-4244-b1aa-ede95edd30e8\") " pod="hostpath-provisioner/csi-hostpathplugin-7p7x5" Mar 07 14:41:21 crc kubenswrapper[4943]: E0307 14:41:21.942821 4943 projected.go:288] Couldn't get configMap openshift-apiserver-operator/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.979478 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92fpw\" (UniqueName: \"kubernetes.io/projected/712a2a31-4205-4346-9a32-858a77615eb6-kube-api-access-92fpw\") pod \"marketplace-operator-79b997595-sqsdb\" (UID: \"712a2a31-4205-4346-9a32-858a77615eb6\") " pod="openshift-marketplace/marketplace-operator-79b997595-sqsdb" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.989486 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e6f60f83-1b2e-4f75-9065-9971654e878a-metrics-certs\") pod \"network-metrics-daemon-ntc42\" (UID: \"e6f60f83-1b2e-4f75-9065-9971654e878a\") " pod="openshift-multus/network-metrics-daemon-ntc42" Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.989729 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:21 crc kubenswrapper[4943]: E0307 14:41:21.990226 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:22.490202842 +0000 UTC m=+124.442339360 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:21 crc kubenswrapper[4943]: I0307 14:41:21.996831 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlplm\" (UniqueName: \"kubernetes.io/projected/7601d311-698b-4389-a9a2-e4e24fae23ea-kube-api-access-jlplm\") pod \"collect-profiles-29548230-zr6fd\" (UID: \"7601d311-698b-4389-a9a2-e4e24fae23ea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29548230-zr6fd" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.004535 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-4vkx6"] Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.024158 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8xj6\" (UniqueName: \"kubernetes.io/projected/203d975c-92a2-424e-a6c4-e4904440dead-kube-api-access-r8xj6\") pod \"packageserver-d55dfcdfc-q85kb\" (UID: \"203d975c-92a2-424e-a6c4-e4904440dead\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-q85kb" Mar 07 14:41:22 crc kubenswrapper[4943]: W0307 14:41:22.026142 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod08f9e92f_8419_4ba3_a532_c13e5bb6cab4.slice/crio-81f613fef78c2801c453866d2530e8f2ab7919600f562a28341f728ec2c1902c WatchSource:0}: Error finding container 81f613fef78c2801c453866d2530e8f2ab7919600f562a28341f728ec2c1902c: Status 404 returned error can't find the container with id 81f613fef78c2801c453866d2530e8f2ab7919600f562a28341f728ec2c1902c Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.040586 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfkx7\" (UniqueName: \"kubernetes.io/projected/0f3dc501-2bc1-4930-a644-f020c92c2827-kube-api-access-rfkx7\") pod \"olm-operator-6b444d44fb-kmxzj\" (UID: \"0f3dc501-2bc1-4930-a644-f020c92c2827\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kmxzj" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.049265 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-h2f4f"] Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.060238 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-j89vd" Mar 07 14:41:22 crc kubenswrapper[4943]: E0307 14:41:22.061662 4943 projected.go:288] Couldn't get configMap openshift-apiserver/openshift-service-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:22 crc kubenswrapper[4943]: E0307 14:41:22.061691 4943 projected.go:194] Error preparing data for projected volume kube-api-access-nj5lj for pod openshift-apiserver/apiserver-76f77b778f-w5mcz: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:22 crc kubenswrapper[4943]: E0307 14:41:22.061758 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4d86f7af-32df-419e-a833-d596c2a79ba7-kube-api-access-nj5lj podName:4d86f7af-32df-419e-a833-d596c2a79ba7 nodeName:}" failed. No retries permitted until 2026-03-07 14:41:22.561737007 +0000 UTC m=+124.513873515 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-nj5lj" (UniqueName: "kubernetes.io/projected/4d86f7af-32df-419e-a833-d596c2a79ba7-kube-api-access-nj5lj") pod "apiserver-76f77b778f-w5mcz" (UID: "4d86f7af-32df-419e-a833-d596c2a79ba7") : failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.063207 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/df59699d-5df8-401e-8bb0-0afa130ac06b-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-lk96z\" (UID: \"df59699d-5df8-401e-8bb0-0afa130ac06b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lk96z" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.063353 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-sqsdb" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.072046 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7p4vl" Mar 07 14:41:22 crc kubenswrapper[4943]: E0307 14:41:22.078725 4943 projected.go:288] Couldn't get configMap openshift-oauth-apiserver/openshift-service-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:22 crc kubenswrapper[4943]: E0307 14:41:22.078762 4943 projected.go:194] Error preparing data for projected volume kube-api-access-lg6bx for pod openshift-oauth-apiserver/apiserver-7bbb656c7d-hr5px: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:22 crc kubenswrapper[4943]: E0307 14:41:22.078841 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/dc37c9ce-2d22-4a80-be9b-0428621f7ca5-kube-api-access-lg6bx podName:dc37c9ce-2d22-4a80-be9b-0428621f7ca5 nodeName:}" failed. No retries permitted until 2026-03-07 14:41:22.578816251 +0000 UTC m=+124.530952839 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-lg6bx" (UniqueName: "kubernetes.io/projected/dc37c9ce-2d22-4a80-be9b-0428621f7ca5-kube-api-access-lg6bx") pod "apiserver-7bbb656c7d-hr5px" (UID: "dc37c9ce-2d22-4a80-be9b-0428621f7ca5") : failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.083590 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/618e4221-7bc8-4721-ac10-1a1373d0fade-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-x8mpt\" (UID: \"618e4221-7bc8-4721-ac10-1a1373d0fade\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x8mpt" Mar 07 14:41:22 crc kubenswrapper[4943]: E0307 14:41:22.087030 4943 projected.go:288] Couldn't get configMap openshift-authentication/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.090311 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:22 crc kubenswrapper[4943]: E0307 14:41:22.090576 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:22.590541135 +0000 UTC m=+124.542677633 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.090832 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:22 crc kubenswrapper[4943]: E0307 14:41:22.091207 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:22.591187202 +0000 UTC m=+124.543323890 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.098722 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jpq82" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.102831 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvzjw\" (UniqueName: \"kubernetes.io/projected/40cb1361-4bb1-450f-8d78-1a033e611f04-kube-api-access-lvzjw\") pod \"service-ca-9c57cc56f-2lnfd\" (UID: \"40cb1361-4bb1-450f-8d78-1a033e611f04\") " pod="openshift-service-ca/service-ca-9c57cc56f-2lnfd" Mar 07 14:41:22 crc kubenswrapper[4943]: E0307 14:41:22.103337 4943 projected.go:288] Couldn't get configMap openshift-console/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.115633 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmbvx\" (UniqueName: \"kubernetes.io/projected/e8c3aeaf-47bf-48ab-b029-2f19e51dae4a-kube-api-access-tmbvx\") pod \"machine-config-controller-84d6567774-hzqrp\" (UID: \"e8c3aeaf-47bf-48ab-b029-2f19e51dae4a\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hzqrp" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.119765 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-q85kb" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.132376 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29548230-zr6fd" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.135411 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zj4zf\" (UniqueName: \"kubernetes.io/projected/839b7c38-e05d-4e85-942a-1936e3b2bcf7-kube-api-access-zj4zf\") pod \"machine-config-server-dmkh8\" (UID: \"839b7c38-e05d-4e85-942a-1936e3b2bcf7\") " pod="openshift-machine-config-operator/machine-config-server-dmkh8" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.140470 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kmxzj" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.153063 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-2lnfd" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.155970 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kdrbq" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.158380 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dlkqz\" (UniqueName: \"kubernetes.io/projected/986041cd-e5ea-4269-82b0-5799d343b155-kube-api-access-dlkqz\") pod \"ingress-canary-w286n\" (UID: \"986041cd-e5ea-4269-82b0-5799d343b155\") " pod="openshift-ingress-canary/ingress-canary-w286n" Mar 07 14:41:22 crc kubenswrapper[4943]: E0307 14:41:22.175394 4943 configmap.go:193] Couldn't get configMap openshift-kube-apiserver-operator/kube-apiserver-operator-config: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:22 crc kubenswrapper[4943]: E0307 14:41:22.175511 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b33c7fd7-2d0f-48b7-9932-e7740577280f-config podName:b33c7fd7-2d0f-48b7-9932-e7740577280f nodeName:}" failed. No retries permitted until 2026-03-07 14:41:23.175472618 +0000 UTC m=+125.127609106 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/b33c7fd7-2d0f-48b7-9932-e7740577280f-config") pod "kube-apiserver-operator-766d6c64bb-9gzdv" (UID: "b33c7fd7-2d0f-48b7-9932-e7740577280f") : failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.176094 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-7p7x5" Mar 07 14:41:22 crc kubenswrapper[4943]: E0307 14:41:22.177911 4943 secret.go:188] Couldn't get secret openshift-image-registry/image-registry-operator-tls: failed to sync secret cache: timed out waiting for the condition Mar 07 14:41:22 crc kubenswrapper[4943]: E0307 14:41:22.178027 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/df59699d-5df8-401e-8bb0-0afa130ac06b-image-registry-operator-tls podName:df59699d-5df8-401e-8bb0-0afa130ac06b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:23.178002444 +0000 UTC m=+125.130138942 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "image-registry-operator-tls" (UniqueName: "kubernetes.io/secret/df59699d-5df8-401e-8bb0-0afa130ac06b-image-registry-operator-tls") pod "cluster-image-registry-operator-dc59b4c8b-lk96z" (UID: "df59699d-5df8-401e-8bb0-0afa130ac06b") : failed to sync secret cache: timed out waiting for the condition Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.187288 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-dmkh8" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.191469 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.191494 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6m8vk\" (UniqueName: \"kubernetes.io/projected/0b6ab22d-b401-4727-83f2-94b527c4a366-kube-api-access-6m8vk\") pod \"service-ca-operator-777779d784-ztm29\" (UID: \"0b6ab22d-b401-4727-83f2-94b527c4a366\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ztm29" Mar 07 14:41:22 crc kubenswrapper[4943]: E0307 14:41:22.191850 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:22.691834673 +0000 UTC m=+124.643971171 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.198980 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p96tc\" (UniqueName: \"kubernetes.io/projected/23baba8c-4f55-4188-8967-61093c57e913-kube-api-access-p96tc\") pod \"dns-default-2rd7t\" (UID: \"23baba8c-4f55-4188-8967-61093c57e913\") " pod="openshift-dns/dns-default-2rd7t" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.203385 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.212004 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-w286n" Mar 07 14:41:22 crc kubenswrapper[4943]: E0307 14:41:22.222116 4943 projected.go:288] Couldn't get configMap openshift-controller-manager-operator/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:22 crc kubenswrapper[4943]: W0307 14:41:22.224154 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod839b7c38_e05d_4e85_942a_1936e3b2bcf7.slice/crio-8f0e4d884bc48d348035eb3cc85baef34cd32b6071e362bec39addf13751d61b WatchSource:0}: Error finding container 8f0e4d884bc48d348035eb3cc85baef34cd32b6071e362bec39addf13751d61b: Status 404 returned error can't find the container with id 8f0e4d884bc48d348035eb3cc85baef34cd32b6071e362bec39addf13751d61b Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.267778 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjpfb\" (UniqueName: \"kubernetes.io/projected/030c2b00-0d20-447c-b7a5-55d86f1674d8-kube-api-access-zjpfb\") pod \"etcd-operator-b45778765-dqpk4\" (UID: \"030c2b00-0d20-447c-b7a5-55d86f1674d8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dqpk4" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.267955 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hq95\" (UniqueName: \"kubernetes.io/projected/214a6dbd-445a-4b41-a3b2-f3efc93f6ed4-kube-api-access-5hq95\") pod \"multus-admission-controller-857f4d67dd-58n5g\" (UID: \"214a6dbd-445a-4b41-a3b2-f3efc93f6ed4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-58n5g" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.280775 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmgxs\" (UniqueName: \"kubernetes.io/projected/2886e9c5-32c9-440a-8d58-d8c00d7742de-kube-api-access-dmgxs\") pod \"package-server-manager-789f6589d5-l7g67\" (UID: \"2886e9c5-32c9-440a-8d58-d8c00d7742de\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-l7g67" Mar 07 14:41:22 crc kubenswrapper[4943]: E0307 14:41:22.282751 4943 projected.go:288] Couldn't get configMap openshift-machine-api/openshift-service-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:22 crc kubenswrapper[4943]: E0307 14:41:22.282773 4943 projected.go:194] Error preparing data for projected volume kube-api-access-d7k5c for pod openshift-machine-api/machine-api-operator-5694c8668f-bk5xc: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:22 crc kubenswrapper[4943]: E0307 14:41:22.282820 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c400fdf2-390c-4c48-8c3e-4854d2e0b58f-kube-api-access-d7k5c podName:c400fdf2-390c-4c48-8c3e-4854d2e0b58f nodeName:}" failed. No retries permitted until 2026-03-07 14:41:22.782802463 +0000 UTC m=+124.734938961 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-d7k5c" (UniqueName: "kubernetes.io/projected/c400fdf2-390c-4c48-8c3e-4854d2e0b58f-kube-api-access-d7k5c") pod "machine-api-operator-5694c8668f-bk5xc" (UID: "c400fdf2-390c-4c48-8c3e-4854d2e0b58f") : failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:22 crc kubenswrapper[4943]: E0307 14:41:22.284164 4943 projected.go:288] Couldn't get configMap openshift-controller-manager/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.292977 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:22 crc kubenswrapper[4943]: E0307 14:41:22.293333 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:22.793324406 +0000 UTC m=+124.745460904 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.296823 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7r5l\" (UniqueName: \"kubernetes.io/projected/7ca6333f-d5ff-4c06-a140-3b9c29710528-kube-api-access-z7r5l\") pod \"router-default-5444994796-nk422\" (UID: \"7ca6333f-d5ff-4c06-a140-3b9c29710528\") " pod="openshift-ingress/router-default-5444994796-nk422" Mar 07 14:41:22 crc kubenswrapper[4943]: E0307 14:41:22.303722 4943 projected.go:288] Couldn't get configMap openshift-cluster-machine-approver/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.319787 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.322050 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6wjd\" (UniqueName: \"kubernetes.io/projected/df59699d-5df8-401e-8bb0-0afa130ac06b-kube-api-access-p6wjd\") pod \"cluster-image-registry-operator-dc59b4c8b-lk96z\" (UID: \"df59699d-5df8-401e-8bb0-0afa130ac06b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lk96z" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.333370 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-nk422" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.346449 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x8mpt" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.348496 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-j89vd"] Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.356872 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2m79r\" (UniqueName: \"kubernetes.io/projected/415baeb7-0937-47e1-9d7e-8c2078911c3c-kube-api-access-2m79r\") pod \"kube-storage-version-migrator-operator-b67b599dd-d622z\" (UID: \"415baeb7-0937-47e1-9d7e-8c2078911c3c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d622z" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.367675 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.380275 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d622z" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.382534 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.387144 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-dqpk4" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.392237 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hzqrp" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.394564 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:22 crc kubenswrapper[4943]: E0307 14:41:22.395330 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:22.895313352 +0000 UTC m=+124.847449850 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.405714 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.407045 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-ztm29" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.412742 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-l7g67" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.421496 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 07 14:41:22 crc kubenswrapper[4943]: E0307 14:41:22.423081 4943 projected.go:288] Couldn't get configMap openshift-ingress-operator/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.441103 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.447488 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sqsdb"] Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.458832 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-nk422" event={"ID":"7ca6333f-d5ff-4c06-a140-3b9c29710528","Type":"ContainerStarted","Data":"0665870a6e5207e57884abd95d02f01b1e7eb5160a3c06332ceaddfafc26a4ef"} Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.459034 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.460391 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-dmkh8" event={"ID":"839b7c38-e05d-4e85-942a-1936e3b2bcf7","Type":"ContainerStarted","Data":"8f0e4d884bc48d348035eb3cc85baef34cd32b6071e362bec39addf13751d61b"} Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.461672 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-j89vd" event={"ID":"2a45a1ae-558f-4600-b354-dea0169b175f","Type":"ContainerStarted","Data":"ba4a7e705b39de440389663968409de2f036cdc51302835a708fd8034f3775b1"} Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.464741 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-htv6d" event={"ID":"0de7b0c4-4acf-40de-867d-1fc33069fa51","Type":"ContainerStarted","Data":"5ad27bce7efee616ce74cde00b2b89d44c20dc06fb06394eddde07f604433c89"} Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.464789 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-htv6d" event={"ID":"0de7b0c4-4acf-40de-867d-1fc33069fa51","Type":"ContainerStarted","Data":"a503af39597f4a531b75289a655cb79ea7f4c380985b8d0744618afd83475c86"} Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.465314 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-multus/cni-sysctl-allowlist-ds-htv6d" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.466147 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e6f60f83-1b2e-4f75-9065-9971654e878a-metrics-certs\") pod \"network-metrics-daemon-ntc42\" (UID: \"e6f60f83-1b2e-4f75-9065-9971654e878a\") " pod="openshift-multus/network-metrics-daemon-ntc42" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.468704 4943 generic.go:334] "Generic (PLEG): container finished" podID="08f9e92f-8419-4ba3-a532-c13e5bb6cab4" containerID="b21ec4da88d5ad120e58096e646a2c849087b3b261ddfbddf499f2fb243576a8" exitCode=0 Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.469173 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4vkx6" event={"ID":"08f9e92f-8419-4ba3-a532-c13e5bb6cab4","Type":"ContainerDied","Data":"b21ec4da88d5ad120e58096e646a2c849087b3b261ddfbddf499f2fb243576a8"} Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.469196 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4vkx6" event={"ID":"08f9e92f-8419-4ba3-a532-c13e5bb6cab4","Type":"ContainerStarted","Data":"81f613fef78c2801c453866d2530e8f2ab7919600f562a28341f728ec2c1902c"} Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.474567 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.493128 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-2rd7t" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.498367 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jpq82"] Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.503208 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.504104 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:22 crc kubenswrapper[4943]: E0307 14:41:22.506208 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:23.006196348 +0000 UTC m=+124.958332846 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.518962 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"36053e9d136906da3b2eb3d258bf5b73bb956ea6141497e74f1f18c8f9a937a9"} Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.519726 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.522208 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-h2f4f" event={"ID":"e03f48ff-595f-4537-824f-650401604f7e","Type":"ContainerStarted","Data":"68db1ce39e81ce76acc732b7b2fecbab4c68969fabf1fcc3fc4347bee6afd0dd"} Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.529577 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ntc42" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.531065 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.544061 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.545224 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29548230-zr6fd"] Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.546289 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-w286n"] Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.559783 4943 request.go:700] Waited for 1.976861757s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/configmaps?fieldSelector=metadata.name%3Dkube-root-ca.crt&resourceVersion=27130 Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.562208 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.580985 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 07 14:41:22 crc kubenswrapper[4943]: E0307 14:41:22.583696 4943 projected.go:288] Couldn't get configMap openshift-route-controller-manager/openshift-service-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:22 crc kubenswrapper[4943]: E0307 14:41:22.583725 4943 projected.go:194] Error preparing data for projected volume kube-api-access-qx4fv for pod openshift-route-controller-manager/route-controller-manager-6576b87f9c-6qmrb: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:22 crc kubenswrapper[4943]: E0307 14:41:22.583791 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/51fb375a-53bd-420a-930f-6b6cdb69f793-kube-api-access-qx4fv podName:51fb375a-53bd-420a-930f-6b6cdb69f793 nodeName:}" failed. No retries permitted until 2026-03-07 14:41:23.083770401 +0000 UTC m=+125.035906899 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-qx4fv" (UniqueName: "kubernetes.io/projected/51fb375a-53bd-420a-930f-6b6cdb69f793-kube-api-access-qx4fv") pod "route-controller-manager-6576b87f9c-6qmrb" (UID: "51fb375a-53bd-420a-930f-6b6cdb69f793") : failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:22 crc kubenswrapper[4943]: E0307 14:41:22.597575 4943 projected.go:288] Couldn't get configMap openshift-console/openshift-service-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:22 crc kubenswrapper[4943]: E0307 14:41:22.597609 4943 projected.go:194] Error preparing data for projected volume kube-api-access-9lmxv for pod openshift-console/console-f9d7485db-zgx7p: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:22 crc kubenswrapper[4943]: E0307 14:41:22.597656 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/698827d2-21c9-4856-a9cd-17bb3df88bce-kube-api-access-9lmxv podName:698827d2-21c9-4856-a9cd-17bb3df88bce nodeName:}" failed. No retries permitted until 2026-03-07 14:41:23.097641431 +0000 UTC m=+125.049777929 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-9lmxv" (UniqueName: "kubernetes.io/projected/698827d2-21c9-4856-a9cd-17bb3df88bce-kube-api-access-9lmxv") pod "console-f9d7485db-zgx7p" (UID: "698827d2-21c9-4856-a9cd-17bb3df88bce") : failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.600346 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.604769 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.604865 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lg6bx\" (UniqueName: \"kubernetes.io/projected/dc37c9ce-2d22-4a80-be9b-0428621f7ca5-kube-api-access-lg6bx\") pod \"apiserver-7bbb656c7d-hr5px\" (UID: \"dc37c9ce-2d22-4a80-be9b-0428621f7ca5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hr5px" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.604897 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nj5lj\" (UniqueName: \"kubernetes.io/projected/4d86f7af-32df-419e-a833-d596c2a79ba7-kube-api-access-nj5lj\") pod \"apiserver-76f77b778f-w5mcz\" (UID: \"4d86f7af-32df-419e-a833-d596c2a79ba7\") " pod="openshift-apiserver/apiserver-76f77b778f-w5mcz" Mar 07 14:41:22 crc kubenswrapper[4943]: E0307 14:41:22.605814 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:23.105804223 +0000 UTC m=+125.057940721 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.618857 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.639333 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.659507 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Mar 07 14:41:22 crc kubenswrapper[4943]: W0307 14:41:22.661719 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod986041cd_e5ea_4269_82b0_5799d343b155.slice/crio-3862d42f7649cd1559295d019c60ed4caed2632dc42a12bd2b770ca8167c59ab WatchSource:0}: Error finding container 3862d42f7649cd1559295d019c60ed4caed2632dc42a12bd2b770ca8167c59ab: Status 404 returned error can't find the container with id 3862d42f7649cd1559295d019c60ed4caed2632dc42a12bd2b770ca8167c59ab Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.679757 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-7p4vl"] Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.679550 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.693017 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nj5lj\" (UniqueName: \"kubernetes.io/projected/4d86f7af-32df-419e-a833-d596c2a79ba7-kube-api-access-nj5lj\") pod \"apiserver-76f77b778f-w5mcz\" (UID: \"4d86f7af-32df-419e-a833-d596c2a79ba7\") " pod="openshift-apiserver/apiserver-76f77b778f-w5mcz" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.703245 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.705869 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:22 crc kubenswrapper[4943]: E0307 14:41:22.706782 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:23.206759162 +0000 UTC m=+125.158895660 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.712937 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lg6bx\" (UniqueName: \"kubernetes.io/projected/dc37c9ce-2d22-4a80-be9b-0428621f7ca5-kube-api-access-lg6bx\") pod \"apiserver-7bbb656c7d-hr5px\" (UID: \"dc37c9ce-2d22-4a80-be9b-0428621f7ca5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hr5px" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.724336 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 07 14:41:22 crc kubenswrapper[4943]: W0307 14:41:22.735778 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf7170a81_b6f1_4696_9356_737721576cf6.slice/crio-04f433185b319a856a6e86d26f8a9a6f2355b54dbe5fcf90d0d61e8771e5328c WatchSource:0}: Error finding container 04f433185b319a856a6e86d26f8a9a6f2355b54dbe5fcf90d0d61e8771e5328c: Status 404 returned error can't find the container with id 04f433185b319a856a6e86d26f8a9a6f2355b54dbe5fcf90d0d61e8771e5328c Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.739473 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.767193 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.787110 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.795617 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-scbtq\" (UniqueName: \"kubernetes.io/projected/fdcfa33a-5367-46db-8270-662184116079-kube-api-access-scbtq\") pod \"control-plane-machine-set-operator-78cbb6b69f-pv7b4\" (UID: \"fdcfa33a-5367-46db-8270-662184116079\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pv7b4" Mar 07 14:41:22 crc kubenswrapper[4943]: E0307 14:41:22.801039 4943 projected.go:288] Couldn't get configMap openshift-console-operator/openshift-service-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:22 crc kubenswrapper[4943]: E0307 14:41:22.801453 4943 projected.go:194] Error preparing data for projected volume kube-api-access-hffv9 for pod openshift-console-operator/console-operator-58897d9998-lg6mk: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:22 crc kubenswrapper[4943]: E0307 14:41:22.801536 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/b7960fae-df40-482f-a9ab-790fa11aaf8f-kube-api-access-hffv9 podName:b7960fae-df40-482f-a9ab-790fa11aaf8f nodeName:}" failed. No retries permitted until 2026-03-07 14:41:23.30149107 +0000 UTC m=+125.253627568 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-hffv9" (UniqueName: "kubernetes.io/projected/b7960fae-df40-482f-a9ab-790fa11aaf8f-kube-api-access-hffv9") pod "console-operator-58897d9998-lg6mk" (UID: "b7960fae-df40-482f-a9ab-790fa11aaf8f") : failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.806534 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.808145 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.808562 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7k5c\" (UniqueName: \"kubernetes.io/projected/c400fdf2-390c-4c48-8c3e-4854d2e0b58f-kube-api-access-d7k5c\") pod \"machine-api-operator-5694c8668f-bk5xc\" (UID: \"c400fdf2-390c-4c48-8c3e-4854d2e0b58f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bk5xc" Mar 07 14:41:22 crc kubenswrapper[4943]: E0307 14:41:22.810001 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:23.309976 +0000 UTC m=+125.262112498 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.810810 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-kdrbq"] Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.813444 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7k5c\" (UniqueName: \"kubernetes.io/projected/c400fdf2-390c-4c48-8c3e-4854d2e0b58f-kube-api-access-d7k5c\") pod \"machine-api-operator-5694c8668f-bk5xc\" (UID: \"c400fdf2-390c-4c48-8c3e-4854d2e0b58f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bk5xc" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.816530 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kmxzj"] Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.816580 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-q85kb"] Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.819538 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 07 14:41:22 crc kubenswrapper[4943]: E0307 14:41:22.821950 4943 projected.go:288] Couldn't get configMap openshift-cluster-samples-operator/openshift-service-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:22 crc kubenswrapper[4943]: E0307 14:41:22.821999 4943 projected.go:194] Error preparing data for projected volume kube-api-access-bw8vf for pod openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dr8bw: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:22 crc kubenswrapper[4943]: E0307 14:41:22.822119 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/28995414-2bd3-4b50-90ec-0fbb63a15ef6-kube-api-access-bw8vf podName:28995414-2bd3-4b50-90ec-0fbb63a15ef6 nodeName:}" failed. No retries permitted until 2026-03-07 14:41:23.322070353 +0000 UTC m=+125.274206851 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-bw8vf" (UniqueName: "kubernetes.io/projected/28995414-2bd3-4b50-90ec-0fbb63a15ef6-kube-api-access-bw8vf") pod "cluster-samples-operator-665b6dd947-dr8bw" (UID: "28995414-2bd3-4b50-90ec-0fbb63a15ef6") : failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.825679 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-7p7x5"] Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.838877 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.864260 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 07 14:41:22 crc kubenswrapper[4943]: E0307 14:41:22.876444 4943 projected.go:194] Error preparing data for projected volume kube-api-access-w4tmt for pod openshift-console/downloads-7954f5f757-8jjq4: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:22 crc kubenswrapper[4943]: E0307 14:41:22.876546 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/93cd8321-367e-4f02-9f8e-2e044e0f1d89-kube-api-access-w4tmt podName:93cd8321-367e-4f02-9f8e-2e044e0f1d89 nodeName:}" failed. No retries permitted until 2026-03-07 14:41:23.376523296 +0000 UTC m=+125.328659794 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-w4tmt" (UniqueName: "kubernetes.io/projected/93cd8321-367e-4f02-9f8e-2e044e0f1d89-kube-api-access-w4tmt") pod "downloads-7954f5f757-8jjq4" (UID: "93cd8321-367e-4f02-9f8e-2e044e0f1d89") : failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.881843 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.890428 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-2lnfd"] Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.899821 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.910012 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:22 crc kubenswrapper[4943]: E0307 14:41:22.910384 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:23.410360914 +0000 UTC m=+125.362497412 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.918854 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Mar 07 14:41:22 crc kubenswrapper[4943]: E0307 14:41:22.923268 4943 projected.go:194] Error preparing data for projected volume kube-api-access-m69bh for pod openshift-authentication-operator/authentication-operator-69f744f599-ltd8c: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:22 crc kubenswrapper[4943]: E0307 14:41:22.923333 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/b541cfee-3ec5-4be3-9acd-d58ebb79f0a2-kube-api-access-m69bh podName:b541cfee-3ec5-4be3-9acd-d58ebb79f0a2 nodeName:}" failed. No retries permitted until 2026-03-07 14:41:23.42331617 +0000 UTC m=+125.375452668 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-m69bh" (UniqueName: "kubernetes.io/projected/b541cfee-3ec5-4be3-9acd-d58ebb79f0a2-kube-api-access-m69bh") pod "authentication-operator-69f744f599-ltd8c" (UID: "b541cfee-3ec5-4be3-9acd-d58ebb79f0a2") : failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.938414 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pv7b4" Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.940460 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Mar 07 14:41:22 crc kubenswrapper[4943]: W0307 14:41:22.942701 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod40cb1361_4bb1_450f_8d78_1a033e611f04.slice/crio-0911cf9bfce0ed9949142f1c192b79c4ae56c0579aca6b24b879ea1ca6fb1587 WatchSource:0}: Error finding container 0911cf9bfce0ed9949142f1c192b79c4ae56c0579aca6b24b879ea1ca6fb1587: Status 404 returned error can't find the container with id 0911cf9bfce0ed9949142f1c192b79c4ae56c0579aca6b24b879ea1ca6fb1587 Mar 07 14:41:22 crc kubenswrapper[4943]: E0307 14:41:22.943174 4943 projected.go:194] Error preparing data for projected volume kube-api-access-xm5zc for pod openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dvjg6: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:22 crc kubenswrapper[4943]: E0307 14:41:22.943316 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0da2f3db-9ac4-42bf-b4b8-a4d538d9f4e6-kube-api-access-xm5zc podName:0da2f3db-9ac4-42bf-b4b8-a4d538d9f4e6 nodeName:}" failed. No retries permitted until 2026-03-07 14:41:23.443285197 +0000 UTC m=+125.395421695 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-xm5zc" (UniqueName: "kubernetes.io/projected/0da2f3db-9ac4-42bf-b4b8-a4d538d9f4e6-kube-api-access-xm5zc") pod "openshift-apiserver-operator-796bbdcf4f-dvjg6" (UID: "0da2f3db-9ac4-42bf-b4b8-a4d538d9f4e6") : failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.959660 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 07 14:41:22 crc kubenswrapper[4943]: E0307 14:41:22.967310 4943 projected.go:194] Error preparing data for projected volume kube-api-access-4pr5v for pod openshift-authentication/oauth-openshift-558db77b4-cnbpx: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:22 crc kubenswrapper[4943]: E0307 14:41:22.967372 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/31ce0220-0f53-4d63-aa7b-99357e9c6b04-kube-api-access-4pr5v podName:31ce0220-0f53-4d63-aa7b-99357e9c6b04 nodeName:}" failed. No retries permitted until 2026-03-07 14:41:23.467354452 +0000 UTC m=+125.419490950 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-4pr5v" (UniqueName: "kubernetes.io/projected/31ce0220-0f53-4d63-aa7b-99357e9c6b04-kube-api-access-4pr5v") pod "oauth-openshift-558db77b4-cnbpx" (UID: "31ce0220-0f53-4d63-aa7b-99357e9c6b04") : failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:22 crc kubenswrapper[4943]: I0307 14:41:22.984378 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 07 14:41:22 crc kubenswrapper[4943]: E0307 14:41:22.984397 4943 projected.go:194] Error preparing data for projected volume kube-api-access-ww544 for pod openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-x5qjf: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:22 crc kubenswrapper[4943]: E0307 14:41:22.984490 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/00cc0b59-be8e-402b-b6c9-56e9e20c4b9b-kube-api-access-ww544 podName:00cc0b59-be8e-402b-b6c9-56e9e20c4b9b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:23.484464236 +0000 UTC m=+125.436600774 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-ww544" (UniqueName: "kubernetes.io/projected/00cc0b59-be8e-402b-b6c9-56e9e20c4b9b-kube-api-access-ww544") pod "openshift-controller-manager-operator-756b6f6bc6-x5qjf" (UID: "00cc0b59-be8e-402b-b6c9-56e9e20c4b9b") : failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:22.999220 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 07 14:41:23 crc kubenswrapper[4943]: E0307 14:41:23.007083 4943 projected.go:194] Error preparing data for projected volume kube-api-access-x46k8 for pod openshift-controller-manager/controller-manager-879f6c89f-n2mhr: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:23 crc kubenswrapper[4943]: E0307 14:41:23.007167 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a04d88f1-a308-42cf-9462-1a06f6338b7b-kube-api-access-x46k8 podName:a04d88f1-a308-42cf-9462-1a06f6338b7b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:23.507148184 +0000 UTC m=+125.459284682 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-x46k8" (UniqueName: "kubernetes.io/projected/a04d88f1-a308-42cf-9462-1a06f6338b7b-kube-api-access-x46k8") pod "controller-manager-879f6c89f-n2mhr" (UID: "a04d88f1-a308-42cf-9462-1a06f6338b7b") : failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.011114 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:23 crc kubenswrapper[4943]: E0307 14:41:23.011487 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:23.511477066 +0000 UTC m=+125.463613564 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.018891 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 07 14:41:23 crc kubenswrapper[4943]: E0307 14:41:23.023853 4943 projected.go:194] Error preparing data for projected volume kube-api-access-t458n for pod openshift-cluster-machine-approver/machine-approver-56656f9798-qml2f: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:23 crc kubenswrapper[4943]: E0307 14:41:23.023953 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f94ec039-60b3-4ecb-b784-04a7ebaf4848-kube-api-access-t458n podName:f94ec039-60b3-4ecb-b784-04a7ebaf4848 nodeName:}" failed. No retries permitted until 2026-03-07 14:41:23.523918019 +0000 UTC m=+125.476054517 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-t458n" (UniqueName: "kubernetes.io/projected/f94ec039-60b3-4ecb-b784-04a7ebaf4848-kube-api-access-t458n") pod "machine-approver-56656f9798-qml2f" (UID: "f94ec039-60b3-4ecb-b784-04a7ebaf4848") : failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.025695 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-dqpk4"] Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.028539 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x8mpt"] Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.051480 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d622z"] Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.053283 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.060920 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.061122 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-58n5g" Mar 07 14:41:23 crc kubenswrapper[4943]: E0307 14:41:23.063730 4943 projected.go:194] Error preparing data for projected volume kube-api-access-x6pzz for pod openshift-ingress-operator/ingress-operator-5b745b69d9-8x4b2: failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:23 crc kubenswrapper[4943]: E0307 14:41:23.063792 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/80ba5362-78a1-4ed1-a6be-fbda0dbe72df-kube-api-access-x6pzz podName:80ba5362-78a1-4ed1-a6be-fbda0dbe72df nodeName:}" failed. No retries permitted until 2026-03-07 14:41:23.563769603 +0000 UTC m=+125.515906101 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-x6pzz" (UniqueName: "kubernetes.io/projected/80ba5362-78a1-4ed1-a6be-fbda0dbe72df-kube-api-access-x6pzz") pod "ingress-operator-5b745b69d9-8x4b2" (UID: "80ba5362-78a1-4ed1-a6be-fbda0dbe72df") : failed to sync configmap cache: timed out waiting for the condition Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.079657 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-l7g67"] Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.107229 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.108508 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-w5mcz" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.115818 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.115869 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qx4fv\" (UniqueName: \"kubernetes.io/projected/51fb375a-53bd-420a-930f-6b6cdb69f793-kube-api-access-qx4fv\") pod \"route-controller-manager-6576b87f9c-6qmrb\" (UID: \"51fb375a-53bd-420a-930f-6b6cdb69f793\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6qmrb" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.115891 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lmxv\" (UniqueName: \"kubernetes.io/projected/698827d2-21c9-4856-a9cd-17bb3df88bce-kube-api-access-9lmxv\") pod \"console-f9d7485db-zgx7p\" (UID: \"698827d2-21c9-4856-a9cd-17bb3df88bce\") " pod="openshift-console/console-f9d7485db-zgx7p" Mar 07 14:41:23 crc kubenswrapper[4943]: E0307 14:41:23.116355 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:23.616333107 +0000 UTC m=+125.568469605 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.127792 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.129786 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qx4fv\" (UniqueName: \"kubernetes.io/projected/51fb375a-53bd-420a-930f-6b6cdb69f793-kube-api-access-qx4fv\") pod \"route-controller-manager-6576b87f9c-6qmrb\" (UID: \"51fb375a-53bd-420a-930f-6b6cdb69f793\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6qmrb" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.129879 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-ntc42"] Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.132028 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-bk5xc" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.135011 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9lmxv\" (UniqueName: \"kubernetes.io/projected/698827d2-21c9-4856-a9cd-17bb3df88bce-kube-api-access-9lmxv\") pod \"console-f9d7485db-zgx7p\" (UID: \"698827d2-21c9-4856-a9cd-17bb3df88bce\") " pod="openshift-console/console-f9d7485db-zgx7p" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.148676 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.149962 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hr5px" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.179144 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.182057 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-zgx7p" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.203157 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-hzqrp"] Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.204775 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-2rd7t"] Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.216908 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:23 crc kubenswrapper[4943]: E0307 14:41:23.217095 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:23.71706838 +0000 UTC m=+125.669204878 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.217143 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/df59699d-5df8-401e-8bb0-0afa130ac06b-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-lk96z\" (UID: \"df59699d-5df8-401e-8bb0-0afa130ac06b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lk96z" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.217174 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.217208 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b33c7fd7-2d0f-48b7-9932-e7740577280f-config\") pod \"kube-apiserver-operator-766d6c64bb-9gzdv\" (UID: \"b33c7fd7-2d0f-48b7-9932-e7740577280f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9gzdv" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.217910 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b33c7fd7-2d0f-48b7-9932-e7740577280f-config\") pod \"kube-apiserver-operator-766d6c64bb-9gzdv\" (UID: \"b33c7fd7-2d0f-48b7-9932-e7740577280f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9gzdv" Mar 07 14:41:23 crc kubenswrapper[4943]: E0307 14:41:23.218741 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:23.718731153 +0000 UTC m=+125.670867701 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.229617 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pv7b4"] Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.230356 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/df59699d-5df8-401e-8bb0-0afa130ac06b-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-lk96z\" (UID: \"df59699d-5df8-401e-8bb0-0afa130ac06b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lk96z" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.247106 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-ztm29"] Mar 07 14:41:23 crc kubenswrapper[4943]: W0307 14:41:23.283732 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0b6ab22d_b401_4727_83f2_94b527c4a366.slice/crio-6786d0e3905f702a9966d73bd48e30b1bb3b397731febe6ab615a6099cd8f89e WatchSource:0}: Error finding container 6786d0e3905f702a9966d73bd48e30b1bb3b397731febe6ab615a6099cd8f89e: Status 404 returned error can't find the container with id 6786d0e3905f702a9966d73bd48e30b1bb3b397731febe6ab615a6099cd8f89e Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.322998 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.323026 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 07 14:41:23 crc kubenswrapper[4943]: E0307 14:41:23.323170 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:23.823151742 +0000 UTC m=+125.775288240 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.324320 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hffv9\" (UniqueName: \"kubernetes.io/projected/b7960fae-df40-482f-a9ab-790fa11aaf8f-kube-api-access-hffv9\") pod \"console-operator-58897d9998-lg6mk\" (UID: \"b7960fae-df40-482f-a9ab-790fa11aaf8f\") " pod="openshift-console-operator/console-operator-58897d9998-lg6mk" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.324356 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bw8vf\" (UniqueName: \"kubernetes.io/projected/28995414-2bd3-4b50-90ec-0fbb63a15ef6-kube-api-access-bw8vf\") pod \"cluster-samples-operator-665b6dd947-dr8bw\" (UID: \"28995414-2bd3-4b50-90ec-0fbb63a15ef6\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dr8bw" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.324627 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:23 crc kubenswrapper[4943]: E0307 14:41:23.324920 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:23.824913068 +0000 UTC m=+125.777049566 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.328342 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6qmrb" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.329325 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bw8vf\" (UniqueName: \"kubernetes.io/projected/28995414-2bd3-4b50-90ec-0fbb63a15ef6-kube-api-access-bw8vf\") pod \"cluster-samples-operator-665b6dd947-dr8bw\" (UID: \"28995414-2bd3-4b50-90ec-0fbb63a15ef6\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dr8bw" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.338780 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hffv9\" (UniqueName: \"kubernetes.io/projected/b7960fae-df40-482f-a9ab-790fa11aaf8f-kube-api-access-hffv9\") pod \"console-operator-58897d9998-lg6mk\" (UID: \"b7960fae-df40-482f-a9ab-790fa11aaf8f\") " pod="openshift-console-operator/console-operator-58897d9998-lg6mk" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.425653 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.425984 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m69bh\" (UniqueName: \"kubernetes.io/projected/b541cfee-3ec5-4be3-9acd-d58ebb79f0a2-kube-api-access-m69bh\") pod \"authentication-operator-69f744f599-ltd8c\" (UID: \"b541cfee-3ec5-4be3-9acd-d58ebb79f0a2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ltd8c" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.426029 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4tmt\" (UniqueName: \"kubernetes.io/projected/93cd8321-367e-4f02-9f8e-2e044e0f1d89-kube-api-access-w4tmt\") pod \"downloads-7954f5f757-8jjq4\" (UID: \"93cd8321-367e-4f02-9f8e-2e044e0f1d89\") " pod="openshift-console/downloads-7954f5f757-8jjq4" Mar 07 14:41:23 crc kubenswrapper[4943]: E0307 14:41:23.428716 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:23.92868655 +0000 UTC m=+125.880823048 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.432405 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4tmt\" (UniqueName: \"kubernetes.io/projected/93cd8321-367e-4f02-9f8e-2e044e0f1d89-kube-api-access-w4tmt\") pod \"downloads-7954f5f757-8jjq4\" (UID: \"93cd8321-367e-4f02-9f8e-2e044e0f1d89\") " pod="openshift-console/downloads-7954f5f757-8jjq4" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.438466 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-58n5g"] Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.439475 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m69bh\" (UniqueName: \"kubernetes.io/projected/b541cfee-3ec5-4be3-9acd-d58ebb79f0a2-kube-api-access-m69bh\") pod \"authentication-operator-69f744f599-ltd8c\" (UID: \"b541cfee-3ec5-4be3-9acd-d58ebb79f0a2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ltd8c" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.442265 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.448494 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dr8bw" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.481273 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.490166 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-8jjq4" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.516369 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9gzdv" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.520253 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.524075 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lk96z" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.527425 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.527491 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xm5zc\" (UniqueName: \"kubernetes.io/projected/0da2f3db-9ac4-42bf-b4b8-a4d538d9f4e6-kube-api-access-xm5zc\") pod \"openshift-apiserver-operator-796bbdcf4f-dvjg6\" (UID: \"0da2f3db-9ac4-42bf-b4b8-a4d538d9f4e6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dvjg6" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.527519 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pr5v\" (UniqueName: \"kubernetes.io/projected/31ce0220-0f53-4d63-aa7b-99357e9c6b04-kube-api-access-4pr5v\") pod \"oauth-openshift-558db77b4-cnbpx\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.527553 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ww544\" (UniqueName: \"kubernetes.io/projected/00cc0b59-be8e-402b-b6c9-56e9e20c4b9b-kube-api-access-ww544\") pod \"openshift-controller-manager-operator-756b6f6bc6-x5qjf\" (UID: \"00cc0b59-be8e-402b-b6c9-56e9e20c4b9b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-x5qjf" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.527576 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t458n\" (UniqueName: \"kubernetes.io/projected/f94ec039-60b3-4ecb-b784-04a7ebaf4848-kube-api-access-t458n\") pod \"machine-approver-56656f9798-qml2f\" (UID: \"f94ec039-60b3-4ecb-b784-04a7ebaf4848\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qml2f" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.527594 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x46k8\" (UniqueName: \"kubernetes.io/projected/a04d88f1-a308-42cf-9462-1a06f6338b7b-kube-api-access-x46k8\") pod \"controller-manager-879f6c89f-n2mhr\" (UID: \"a04d88f1-a308-42cf-9462-1a06f6338b7b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n2mhr" Mar 07 14:41:23 crc kubenswrapper[4943]: E0307 14:41:23.527782 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:24.027765861 +0000 UTC m=+125.979902359 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.539271 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-j89vd" event={"ID":"2a45a1ae-558f-4600-b354-dea0169b175f","Type":"ContainerStarted","Data":"15d56328496eddaa649cf3e2ad98d6adeb3d185b34c6a430aceb29b9e2ea6f95"} Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.539946 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x46k8\" (UniqueName: \"kubernetes.io/projected/a04d88f1-a308-42cf-9462-1a06f6338b7b-kube-api-access-x46k8\") pod \"controller-manager-879f6c89f-n2mhr\" (UID: \"a04d88f1-a308-42cf-9462-1a06f6338b7b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n2mhr" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.540433 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ww544\" (UniqueName: \"kubernetes.io/projected/00cc0b59-be8e-402b-b6c9-56e9e20c4b9b-kube-api-access-ww544\") pod \"openshift-controller-manager-operator-756b6f6bc6-x5qjf\" (UID: \"00cc0b59-be8e-402b-b6c9-56e9e20c4b9b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-x5qjf" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.543154 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pr5v\" (UniqueName: \"kubernetes.io/projected/31ce0220-0f53-4d63-aa7b-99357e9c6b04-kube-api-access-4pr5v\") pod \"oauth-openshift-558db77b4-cnbpx\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.543595 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t458n\" (UniqueName: \"kubernetes.io/projected/f94ec039-60b3-4ecb-b784-04a7ebaf4848-kube-api-access-t458n\") pod \"machine-approver-56656f9798-qml2f\" (UID: \"f94ec039-60b3-4ecb-b784-04a7ebaf4848\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qml2f" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.547756 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-dmkh8" event={"ID":"839b7c38-e05d-4e85-942a-1936e3b2bcf7","Type":"ContainerStarted","Data":"c95fe69c039ee569e882774035bca6d1c174736aefb1b56e552576af9c374e1d"} Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.557432 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xm5zc\" (UniqueName: \"kubernetes.io/projected/0da2f3db-9ac4-42bf-b4b8-a4d538d9f4e6-kube-api-access-xm5zc\") pod \"openshift-apiserver-operator-796bbdcf4f-dvjg6\" (UID: \"0da2f3db-9ac4-42bf-b4b8-a4d538d9f4e6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dvjg6" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.561817 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.563697 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-lg6mk" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.564835 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-h2f4f" event={"ID":"e03f48ff-595f-4537-824f-650401604f7e","Type":"ContainerStarted","Data":"23f26ed4015d58dcc0e784a3da373c142842412f30173bdc7680f3676ce4a759"} Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.568031 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-ztm29" event={"ID":"0b6ab22d-b401-4727-83f2-94b527c4a366","Type":"ContainerStarted","Data":"6786d0e3905f702a9966d73bd48e30b1bb3b397731febe6ab615a6099cd8f89e"} Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.569424 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-dqpk4" event={"ID":"030c2b00-0d20-447c-b7a5-55d86f1674d8","Type":"ContainerStarted","Data":"eb82e2b668e8a069d1d19553f1e3ef8c5ff320d80c2c85e4284fc86756691a58"} Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.576742 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4vkx6" event={"ID":"08f9e92f-8419-4ba3-a532-c13e5bb6cab4","Type":"ContainerStarted","Data":"09723f9347b6204bade403a607b004157e915f2b98de21d15900dd2c1e1f1910"} Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.577756 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4vkx6" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.579250 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-ntc42" event={"ID":"e6f60f83-1b2e-4f75-9065-9971654e878a","Type":"ContainerStarted","Data":"f5315cf24b7b14253102506089e18d48dca4cd8385c2f69a8a865530219d39c8"} Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.583896 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pv7b4" event={"ID":"fdcfa33a-5367-46db-8270-662184116079","Type":"ContainerStarted","Data":"bb582bea2478f9ecbb0403b7f51978f1ea07a486e7678a7e4f2370906e28f214"} Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.589477 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-w286n" event={"ID":"986041cd-e5ea-4269-82b0-5799d343b155","Type":"ContainerStarted","Data":"5e14d69413bc550ab7fd2a362fa257f99a07e4a881771961f89c565b453cdd99"} Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.589528 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-w286n" event={"ID":"986041cd-e5ea-4269-82b0-5799d343b155","Type":"ContainerStarted","Data":"3862d42f7649cd1559295d019c60ed4caed2632dc42a12bd2b770ca8167c59ab"} Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.593286 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kdrbq" event={"ID":"8d97c7b9-9de3-40a6-ba65-db9c730604c0","Type":"ContainerStarted","Data":"269640dfdad86ea6159936327c6b7c0903d50b49bb85d0f5d8698925f0d8b177"} Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.593312 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kdrbq" event={"ID":"8d97c7b9-9de3-40a6-ba65-db9c730604c0","Type":"ContainerStarted","Data":"cf083838791392b9c092377769e487bb813338730cf19565b4bdcf28eac0c4b1"} Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.597786 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-q85kb" event={"ID":"203d975c-92a2-424e-a6c4-e4904440dead","Type":"ContainerStarted","Data":"88831bce8fca3e7fc306f110cf28d928309c3d0847163e9e037bba69b5221f58"} Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.597808 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-q85kb" event={"ID":"203d975c-92a2-424e-a6c4-e4904440dead","Type":"ContainerStarted","Data":"5e78f96dc5ebeccca3d9e3e0b511367bb7cb0a9848e1c62a5d0bbe9b2768b2d1"} Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.598831 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-q85kb" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.615407 4943 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-q85kb container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.25:5443/healthz\": dial tcp 10.217.0.25:5443: connect: connection refused" start-of-body= Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.615441 4943 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-q85kb" podUID="203d975c-92a2-424e-a6c4-e4904440dead" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.25:5443/healthz\": dial tcp 10.217.0.25:5443: connect: connection refused" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.615684 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-sqsdb" event={"ID":"712a2a31-4205-4346-9a32-858a77615eb6","Type":"ContainerStarted","Data":"51fa91e109cdf7e64fa128d5ebe2fa2b77ae12ee40c73925027012b869163b0c"} Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.615705 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-sqsdb" event={"ID":"712a2a31-4205-4346-9a32-858a77615eb6","Type":"ContainerStarted","Data":"07df969ccc164094099b25423ccbaaff05f842977aea3d1c5560e14d8169fd99"} Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.616468 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-sqsdb" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.627061 4943 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-sqsdb container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" start-of-body= Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.627338 4943 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-sqsdb" podUID="712a2a31-4205-4346-9a32-858a77615eb6" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.627618 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.629018 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.629227 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6pzz\" (UniqueName: \"kubernetes.io/projected/80ba5362-78a1-4ed1-a6be-fbda0dbe72df-kube-api-access-x6pzz\") pod \"ingress-operator-5b745b69d9-8x4b2\" (UID: \"80ba5362-78a1-4ed1-a6be-fbda0dbe72df\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8x4b2" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.630789 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-n2mhr" Mar 07 14:41:23 crc kubenswrapper[4943]: E0307 14:41:23.631162 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:24.131147933 +0000 UTC m=+126.083284431 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.633578 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6pzz\" (UniqueName: \"kubernetes.io/projected/80ba5362-78a1-4ed1-a6be-fbda0dbe72df-kube-api-access-x6pzz\") pod \"ingress-operator-5b745b69d9-8x4b2\" (UID: \"80ba5362-78a1-4ed1-a6be-fbda0dbe72df\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8x4b2" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.633686 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-2rd7t" event={"ID":"23baba8c-4f55-4188-8967-61093c57e913","Type":"ContainerStarted","Data":"4484671e94661607b37690ee3ae732d8818c4d0c8a7f8091bb2f6b0d4145e026"} Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.634973 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-7p7x5" event={"ID":"ff8cf71a-e914-4244-b1aa-ede95edd30e8","Type":"ContainerStarted","Data":"7d19529214ee87d7d7729d189a4308375bf10029577f48855b802ffdbf575828"} Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.636241 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-l7g67" event={"ID":"2886e9c5-32c9-440a-8d58-d8c00d7742de","Type":"ContainerStarted","Data":"2727a2d2fb9a968f565e9f5776f8f8c9bc61c3518bc81f675b101c6c05a92997"} Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.639914 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x8mpt" event={"ID":"618e4221-7bc8-4721-ac10-1a1373d0fade","Type":"ContainerStarted","Data":"2f5777e49a18efd8aaff6aedbd8fa7c50fc8f3471bc731e8ba267c9afc7102b9"} Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.642533 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d622z" event={"ID":"415baeb7-0937-47e1-9d7e-8c2078911c3c","Type":"ContainerStarted","Data":"b094812154c3d97f5cf6450fcba1a3ab5a8012881cc63703e65835b0ee86acc4"} Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.658431 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-2lnfd" event={"ID":"40cb1361-4bb1-450f-8d78-1a033e611f04","Type":"ContainerStarted","Data":"46181e8200fa6c71968c8f61f32c203bc12ec227a318c3cd57371cf71bd08c7c"} Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.658486 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-2lnfd" event={"ID":"40cb1361-4bb1-450f-8d78-1a033e611f04","Type":"ContainerStarted","Data":"0911cf9bfce0ed9949142f1c192b79c4ae56c0579aca6b24b879ea1ca6fb1587"} Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.662642 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.666460 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kmxzj" event={"ID":"0f3dc501-2bc1-4930-a644-f020c92c2827","Type":"ContainerStarted","Data":"4889e4a94bb0d76c919421994a6fdef3b122d477af2457f3bda14bca1cb6bc65"} Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.666502 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kmxzj" event={"ID":"0f3dc501-2bc1-4930-a644-f020c92c2827","Type":"ContainerStarted","Data":"ae1d61e41901b1689abe5705f1123fb77f6f8665affd2ce0334531d3e53dd034"} Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.667028 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kmxzj" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.669003 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qml2f" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.676561 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jpq82" event={"ID":"8e19594a-0a69-40b7-8c24-317979edbfbb","Type":"ContainerStarted","Data":"5480e3700dc30c8242b2b0c8f87dde41dc7404916ebf4e61cdd26b540bcb4923"} Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.676604 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jpq82" event={"ID":"8e19594a-0a69-40b7-8c24-317979edbfbb","Type":"ContainerStarted","Data":"46825d1409a877bde3a1f1e4bdbad22d318688ea808ace6b6a7485fed08b2d20"} Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.677248 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jpq82" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.678218 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hzqrp" event={"ID":"e8c3aeaf-47bf-48ab-b029-2f19e51dae4a","Type":"ContainerStarted","Data":"a108ea54a4395c692e093620a1c157b6e9e0090d6a78e542f6d2f90f8bca09b5"} Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.680277 4943 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-kmxzj container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.42:8443/healthz\": dial tcp 10.217.0.42:8443: connect: connection refused" start-of-body= Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.680303 4943 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kmxzj" podUID="0f3dc501-2bc1-4930-a644-f020c92c2827" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.42:8443/healthz\": dial tcp 10.217.0.42:8443: connect: connection refused" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.687234 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.690403 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-nk422" event={"ID":"7ca6333f-d5ff-4c06-a140-3b9c29710528","Type":"ContainerStarted","Data":"dc05890ad8cf5948b5153b5a081bdf1376c0088fd6001871a2bc68efe46bf01e"} Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.693314 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-58n5g" event={"ID":"214a6dbd-445a-4b41-a3b2-f3efc93f6ed4","Type":"ContainerStarted","Data":"2ec205647fcfcf729b7b65ff7575cceb96774ee024231c71016b8846c7262d44"} Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.694753 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.697245 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7p4vl" event={"ID":"f7170a81-b6f1-4696-9356-737721576cf6","Type":"ContainerStarted","Data":"779ec1c5424452a0d88795cc50429350d33eb53e50ead9d1ff430a80334c139e"} Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.697272 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7p4vl" event={"ID":"f7170a81-b6f1-4696-9356-737721576cf6","Type":"ContainerStarted","Data":"491a02326b5342d19ab542bcc5b45440117127434d227834bb7d67a1dbec7728"} Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.697282 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7p4vl" event={"ID":"f7170a81-b6f1-4696-9356-737721576cf6","Type":"ContainerStarted","Data":"04f433185b319a856a6e86d26f8a9a6f2355b54dbe5fcf90d0d61e8771e5328c"} Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.703452 4943 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-jpq82 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.703507 4943 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jpq82" podUID="8e19594a-0a69-40b7-8c24-317979edbfbb" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.722059 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.725611 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-ltd8c" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.732171 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.744801 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29548230-zr6fd" event={"ID":"7601d311-698b-4389-a9a2-e4e24fae23ea","Type":"ContainerStarted","Data":"77e2176d82c678c6306ab7d0f7ca6234ef2c1b3bce73ba85d718298ef1acdc57"} Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.745144 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29548230-zr6fd" event={"ID":"7601d311-698b-4389-a9a2-e4e24fae23ea","Type":"ContainerStarted","Data":"d9cdc1fb988957b10fce7af01e8029c6f8ede0b5a7a6a7856492bfff0b6c326f"} Mar 07 14:41:23 crc kubenswrapper[4943]: E0307 14:41:23.746092 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:24.246079075 +0000 UTC m=+126.198215563 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.759904 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.774176 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dvjg6" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.803152 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.816989 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-x5qjf" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.820293 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.829624 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8x4b2" Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.835126 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:23 crc kubenswrapper[4943]: E0307 14:41:23.838412 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:24.33839339 +0000 UTC m=+126.290529888 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.838734 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:23 crc kubenswrapper[4943]: E0307 14:41:23.841262 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:24.341239093 +0000 UTC m=+126.293375591 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.898604 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-bk5xc"] Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.916120 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-w5mcz"] Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.946252 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:23 crc kubenswrapper[4943]: E0307 14:41:23.946508 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:24.446484234 +0000 UTC m=+126.398620732 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.946608 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:23 crc kubenswrapper[4943]: E0307 14:41:23.947075 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:24.447055279 +0000 UTC m=+126.399191777 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:23 crc kubenswrapper[4943]: I0307 14:41:23.951398 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-multus/cni-sysctl-allowlist-ds-htv6d" Mar 07 14:41:24 crc kubenswrapper[4943]: I0307 14:41:24.048526 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:24 crc kubenswrapper[4943]: E0307 14:41:24.048894 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:24.54886598 +0000 UTC m=+126.501002478 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:24 crc kubenswrapper[4943]: I0307 14:41:24.133318 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-zgx7p"] Mar 07 14:41:24 crc kubenswrapper[4943]: I0307 14:41:24.160173 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:24 crc kubenswrapper[4943]: E0307 14:41:24.160662 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:24.66065075 +0000 UTC m=+126.612787248 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:24 crc kubenswrapper[4943]: I0307 14:41:24.261974 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:24 crc kubenswrapper[4943]: E0307 14:41:24.262358 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:24.762325258 +0000 UTC m=+126.714461756 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:24 crc kubenswrapper[4943]: I0307 14:41:24.268282 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:24 crc kubenswrapper[4943]: E0307 14:41:24.268741 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:24.768721344 +0000 UTC m=+126.720857842 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:24 crc kubenswrapper[4943]: I0307 14:41:24.264015 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dr8bw"] Mar 07 14:41:24 crc kubenswrapper[4943]: I0307 14:41:24.268897 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-6qmrb"] Mar 07 14:41:24 crc kubenswrapper[4943]: I0307 14:41:24.289813 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-8jjq4"] Mar 07 14:41:24 crc kubenswrapper[4943]: I0307 14:41:24.299011 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-hr5px"] Mar 07 14:41:24 crc kubenswrapper[4943]: I0307 14:41:24.342014 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-nk422" Mar 07 14:41:24 crc kubenswrapper[4943]: I0307 14:41:24.369887 4943 patch_prober.go:28] interesting pod/router-default-5444994796-nk422 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 07 14:41:24 crc kubenswrapper[4943]: [-]has-synced failed: reason withheld Mar 07 14:41:24 crc kubenswrapper[4943]: [+]process-running ok Mar 07 14:41:24 crc kubenswrapper[4943]: healthz check failed Mar 07 14:41:24 crc kubenswrapper[4943]: I0307 14:41:24.370018 4943 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-nk422" podUID="7ca6333f-d5ff-4c06-a140-3b9c29710528" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 07 14:41:24 crc kubenswrapper[4943]: I0307 14:41:24.371551 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:24 crc kubenswrapper[4943]: E0307 14:41:24.372200 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:24.872180028 +0000 UTC m=+126.824316536 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:24 crc kubenswrapper[4943]: I0307 14:41:24.472871 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:24 crc kubenswrapper[4943]: E0307 14:41:24.473860 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:24.973845346 +0000 UTC m=+126.925981844 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:24 crc kubenswrapper[4943]: I0307 14:41:24.516674 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-lg6mk"] Mar 07 14:41:24 crc kubenswrapper[4943]: I0307 14:41:24.577189 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:24 crc kubenswrapper[4943]: E0307 14:41:24.577508 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:25.077451534 +0000 UTC m=+127.029588032 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:24 crc kubenswrapper[4943]: I0307 14:41:24.678494 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:24 crc kubenswrapper[4943]: E0307 14:41:24.678959 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:25.178941267 +0000 UTC m=+127.131077765 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:24 crc kubenswrapper[4943]: I0307 14:41:24.688355 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4vkx6" podStartSLOduration=52.68833688 podStartE2EDuration="52.68833688s" podCreationTimestamp="2026-03-07 14:40:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:24.63783437 +0000 UTC m=+126.589970868" watchObservedRunningTime="2026-03-07 14:41:24.68833688 +0000 UTC m=+126.640473378" Mar 07 14:41:24 crc kubenswrapper[4943]: I0307 14:41:24.787335 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:24 crc kubenswrapper[4943]: E0307 14:41:24.789707 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:25.2896885 +0000 UTC m=+127.241824998 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:24 crc kubenswrapper[4943]: I0307 14:41:24.787389 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29548230-zr6fd" podStartSLOduration=52.78736624 podStartE2EDuration="52.78736624s" podCreationTimestamp="2026-03-07 14:40:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:24.730554146 +0000 UTC m=+126.682690644" watchObservedRunningTime="2026-03-07 14:41:24.78736624 +0000 UTC m=+126.739502728" Mar 07 14:41:24 crc kubenswrapper[4943]: I0307 14:41:24.795547 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-dmkh8" podStartSLOduration=6.795108021 podStartE2EDuration="6.795108021s" podCreationTimestamp="2026-03-07 14:41:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:24.785916942 +0000 UTC m=+126.738053440" watchObservedRunningTime="2026-03-07 14:41:24.795108021 +0000 UTC m=+126.747244519" Mar 07 14:41:24 crc kubenswrapper[4943]: I0307 14:41:24.805950 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:24 crc kubenswrapper[4943]: E0307 14:41:24.806546 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:25.306532377 +0000 UTC m=+127.258668875 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:24 crc kubenswrapper[4943]: I0307 14:41:24.895890 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Mar 07 14:41:24 crc kubenswrapper[4943]: I0307 14:41:24.908347 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-ntc42" event={"ID":"e6f60f83-1b2e-4f75-9065-9971654e878a","Type":"ContainerStarted","Data":"ebaddf4c1888f1aee26045bcfb841cde1586deba01dafb79c5715080ae08b372"} Mar 07 14:41:24 crc kubenswrapper[4943]: I0307 14:41:24.909130 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:24 crc kubenswrapper[4943]: E0307 14:41:24.909368 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:25.409338954 +0000 UTC m=+127.361475452 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:24 crc kubenswrapper[4943]: I0307 14:41:24.909529 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:24 crc kubenswrapper[4943]: E0307 14:41:24.909951 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:25.40994439 +0000 UTC m=+127.362080888 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:24 crc kubenswrapper[4943]: I0307 14:41:24.917352 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-sqsdb" podStartSLOduration=52.917322021 podStartE2EDuration="52.917322021s" podCreationTimestamp="2026-03-07 14:40:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:24.894858908 +0000 UTC m=+126.846995406" watchObservedRunningTime="2026-03-07 14:41:24.917322021 +0000 UTC m=+126.869458519" Mar 07 14:41:24 crc kubenswrapper[4943]: I0307 14:41:24.930402 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-8jjq4" event={"ID":"93cd8321-367e-4f02-9f8e-2e044e0f1d89","Type":"ContainerStarted","Data":"3f2d18803f1ab21b2f02b6bf981a660024f2511413d9db4e8953dda465724dba"} Mar 07 14:41:24 crc kubenswrapper[4943]: I0307 14:41:24.932328 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-n2mhr"] Mar 07 14:41:24 crc kubenswrapper[4943]: I0307 14:41:24.974708 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kmxzj" podStartSLOduration=52.97469131 podStartE2EDuration="52.97469131s" podCreationTimestamp="2026-03-07 14:40:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:24.97160611 +0000 UTC m=+126.923742608" watchObservedRunningTime="2026-03-07 14:41:24.97469131 +0000 UTC m=+126.926827808" Mar 07 14:41:24 crc kubenswrapper[4943]: I0307 14:41:24.976947 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-w5mcz" event={"ID":"4d86f7af-32df-419e-a833-d596c2a79ba7","Type":"ContainerStarted","Data":"3b91db8b3e6fb01822f6c49a169f48836bb5656402b0e5342a5d2c38bced5961"} Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.010874 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:25 crc kubenswrapper[4943]: E0307 14:41:25.011194 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:25.511159186 +0000 UTC m=+127.463295684 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.011373 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:25 crc kubenswrapper[4943]: E0307 14:41:25.011830 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:25.511821283 +0000 UTC m=+127.463957771 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.026051 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9gzdv"] Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.065718 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jpq82" podStartSLOduration=53.065697801 podStartE2EDuration="53.065697801s" podCreationTimestamp="2026-03-07 14:40:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:25.043210677 +0000 UTC m=+126.995347175" watchObservedRunningTime="2026-03-07 14:41:25.065697801 +0000 UTC m=+127.017834299" Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.066565 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-h2f4f" podStartSLOduration=53.066558953 podStartE2EDuration="53.066558953s" podCreationTimestamp="2026-03-07 14:40:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:25.06414288 +0000 UTC m=+127.016279378" watchObservedRunningTime="2026-03-07 14:41:25.066558953 +0000 UTC m=+127.018695451" Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.075395 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lk96z"] Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.076952 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-j89vd" event={"ID":"2a45a1ae-558f-4600-b354-dea0169b175f","Type":"ContainerStarted","Data":"49cfdfc47c5b847475a471e0c440d7a6e272edf604e9be9783cb377dbe6ebed5"} Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.083240 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-zgx7p" event={"ID":"698827d2-21c9-4856-a9cd-17bb3df88bce","Type":"ContainerStarted","Data":"71286e965f216d8e057aed8234d1d5eec6ce3c3a6d48e487cb865abb78c92bbd"} Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.089243 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hr5px" event={"ID":"dc37c9ce-2d22-4a80-be9b-0428621f7ca5","Type":"ContainerStarted","Data":"5fc42bf91de734fcad7849c33353e9df4f5bacb8db1907013181799e2d8ecf69"} Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.093247 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pv7b4" event={"ID":"fdcfa33a-5367-46db-8270-662184116079","Type":"ContainerStarted","Data":"a032845a1b67f8d0344860e029b855c74bcdafa797c7de2fdda73a84196fe84b"} Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.095451 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-ztm29" event={"ID":"0b6ab22d-b401-4727-83f2-94b527c4a366","Type":"ContainerStarted","Data":"e25be36316b31f5f6ddafff818d8078f707b9fff7cd6a6dc4bbb5aacee4cd826"} Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.096668 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qml2f" event={"ID":"f94ec039-60b3-4ecb-b784-04a7ebaf4848","Type":"ContainerStarted","Data":"112c3eeee59124dfede871bc4c5056eae28022e88442b3abe8506efd92d2f850"} Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.097874 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-2rd7t" event={"ID":"23baba8c-4f55-4188-8967-61093c57e913","Type":"ContainerStarted","Data":"3dab5bfb786b9da015f8ca715ead7e4198d667a3fd316bb8fcbf3c935f6c1b2b"} Mar 07 14:41:25 crc kubenswrapper[4943]: W0307 14:41:25.098720 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda04d88f1_a308_42cf_9462_1a06f6338b7b.slice/crio-ea887591d1d2ecfc91eac3bc39f3b122370b87c85929715f777938c5eeb85d26 WatchSource:0}: Error finding container ea887591d1d2ecfc91eac3bc39f3b122370b87c85929715f777938c5eeb85d26: Status 404 returned error can't find the container with id ea887591d1d2ecfc91eac3bc39f3b122370b87c85929715f777938c5eeb85d26 Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.098885 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-lg6mk" event={"ID":"b7960fae-df40-482f-a9ab-790fa11aaf8f","Type":"ContainerStarted","Data":"801db490944e604cfd7d11facef322b6af699f5620f369d8f5ea38e2d2004a33"} Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.100764 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6qmrb" event={"ID":"51fb375a-53bd-420a-930f-6b6cdb69f793","Type":"ContainerStarted","Data":"c9e2aff918e1390f32dab0458bb06c2afc74281f3d7feb2fc40f602018a9c95b"} Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.116495 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:25 crc kubenswrapper[4943]: E0307 14:41:25.117562 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:25.617545916 +0000 UTC m=+127.569682414 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.125405 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-cnbpx"] Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.141036 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-nk422" podStartSLOduration=53.141018395 podStartE2EDuration="53.141018395s" podCreationTimestamp="2026-03-07 14:40:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:25.137697829 +0000 UTC m=+127.089834327" watchObservedRunningTime="2026-03-07 14:41:25.141018395 +0000 UTC m=+127.093154893" Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.149503 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-58n5g" event={"ID":"214a6dbd-445a-4b41-a3b2-f3efc93f6ed4","Type":"ContainerStarted","Data":"fe391974f1929c3008c8a45b1b09fb8f8a9da1a5e72792de337160380cc7b1b9"} Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.191557 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=29.191534575 podStartE2EDuration="29.191534575s" podCreationTimestamp="2026-03-07 14:40:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:25.190834087 +0000 UTC m=+127.142970585" watchObservedRunningTime="2026-03-07 14:41:25.191534575 +0000 UTC m=+127.143671073" Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.214410 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-2lnfd" podStartSLOduration=53.214384328 podStartE2EDuration="53.214384328s" podCreationTimestamp="2026-03-07 14:40:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:25.212792297 +0000 UTC m=+127.164928795" watchObservedRunningTime="2026-03-07 14:41:25.214384328 +0000 UTC m=+127.166520826" Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.217958 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:25 crc kubenswrapper[4943]: E0307 14:41:25.218979 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:25.718965327 +0000 UTC m=+127.671101825 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.226988 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d622z" event={"ID":"415baeb7-0937-47e1-9d7e-8c2078911c3c","Type":"ContainerStarted","Data":"4461539aefd8fe0b515dc4d0d7a88590dba66aeec7ad9a1b0b4a3f407d099b89"} Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.289509 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x8mpt" event={"ID":"618e4221-7bc8-4721-ac10-1a1373d0fade","Type":"ContainerStarted","Data":"0a9a65af3317190014aabbbda5069f2f599ec1b6c16566915536c49d5ccc6996"} Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.290121 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7p4vl" podStartSLOduration=53.290104983 podStartE2EDuration="53.290104983s" podCreationTimestamp="2026-03-07 14:40:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:25.289034275 +0000 UTC m=+127.241170773" watchObservedRunningTime="2026-03-07 14:41:25.290104983 +0000 UTC m=+127.242241481" Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.311895 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-ltd8c"] Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.320861 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:25 crc kubenswrapper[4943]: E0307 14:41:25.321039 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:25.821018625 +0000 UTC m=+127.773155123 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.321171 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:25 crc kubenswrapper[4943]: E0307 14:41:25.322610 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:25.822597936 +0000 UTC m=+127.774734434 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.342204 4943 patch_prober.go:28] interesting pod/router-default-5444994796-nk422 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 07 14:41:25 crc kubenswrapper[4943]: [-]has-synced failed: reason withheld Mar 07 14:41:25 crc kubenswrapper[4943]: [+]process-running ok Mar 07 14:41:25 crc kubenswrapper[4943]: healthz check failed Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.342256 4943 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-nk422" podUID="7ca6333f-d5ff-4c06-a140-3b9c29710528" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.346980 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dvjg6"] Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.385374 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hzqrp" event={"ID":"e8c3aeaf-47bf-48ab-b029-2f19e51dae4a","Type":"ContainerStarted","Data":"a730859719832ddf9968d030d376b685f1ed37eda6aebc8aecaac489b0f24eb2"} Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.422760 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:25 crc kubenswrapper[4943]: E0307 14:41:25.426244 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:25.926195063 +0000 UTC m=+127.878331561 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.430018 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kdrbq" event={"ID":"8d97c7b9-9de3-40a6-ba65-db9c730604c0","Type":"ContainerStarted","Data":"77b9f097675994412b1188f51ae903504f3b06c742c1fdaa545172523061a156"} Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.442175 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-7p7x5" event={"ID":"ff8cf71a-e914-4244-b1aa-ede95edd30e8","Type":"ContainerStarted","Data":"59c1c12a7bc88b940a2167708e6facf5fb3e37e43f3b79c39d6cfc60d96450f9"} Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.448577 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d622z" podStartSLOduration=53.448546153 podStartE2EDuration="53.448546153s" podCreationTimestamp="2026-03-07 14:40:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:25.424768366 +0000 UTC m=+127.376904864" watchObservedRunningTime="2026-03-07 14:41:25.448546153 +0000 UTC m=+127.400682641" Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.458571 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-l7g67" event={"ID":"2886e9c5-32c9-440a-8d58-d8c00d7742de","Type":"ContainerStarted","Data":"6758a9fa9620211d630006390c5d5f1c29bd5e94585e450c86236ad420c6a3d1"} Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.458621 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-l7g67" event={"ID":"2886e9c5-32c9-440a-8d58-d8c00d7742de","Type":"ContainerStarted","Data":"f3942b73b3d890802617f3dfa7a094c77b8cde86598f0ea87c82af0e3b26efcd"} Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.460518 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-l7g67" Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.469700 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-bk5xc" event={"ID":"c400fdf2-390c-4c48-8c3e-4854d2e0b58f","Type":"ContainerStarted","Data":"28f81324483c2348d733f2794b8e8ee15df74413124ef9bddf71e55ccae1deaa"} Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.469765 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-bk5xc" event={"ID":"c400fdf2-390c-4c48-8c3e-4854d2e0b58f","Type":"ContainerStarted","Data":"5a28243755245faa2f23181dba49e12c8b5ea9af9446cd37fed2fb4cfe93a0df"} Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.497237 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-dqpk4" event={"ID":"030c2b00-0d20-447c-b7a5-55d86f1674d8","Type":"ContainerStarted","Data":"402b61cb519ab507be3ce71fff7960f5db8fc563f487a9e49203e0a9096c940d"} Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.508399 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-sqsdb" Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.509574 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jpq82" Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.515411 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kmxzj" Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.519803 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-q85kb" Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.525713 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:25 crc kubenswrapper[4943]: E0307 14:41:25.527374 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:26.027357588 +0000 UTC m=+127.979494086 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.529131 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4vkx6" Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.565889 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/cni-sysctl-allowlist-ds-htv6d" podStartSLOduration=7.565870987 podStartE2EDuration="7.565870987s" podCreationTimestamp="2026-03-07 14:41:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:25.54748669 +0000 UTC m=+127.499623198" watchObservedRunningTime="2026-03-07 14:41:25.565870987 +0000 UTC m=+127.518007485" Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.568511 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-x5qjf"] Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.569772 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dfqzr"] Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.572465 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dfqzr" Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.585567 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-w286n" podStartSLOduration=7.5855462970000005 podStartE2EDuration="7.585546297s" podCreationTimestamp="2026-03-07 14:41:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:25.581113552 +0000 UTC m=+127.533250050" watchObservedRunningTime="2026-03-07 14:41:25.585546297 +0000 UTC m=+127.537682785" Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.587057 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.592177 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dfqzr"] Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.626649 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:25 crc kubenswrapper[4943]: E0307 14:41:25.628274 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:26.128259236 +0000 UTC m=+128.080395734 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.644502 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-q85kb" podStartSLOduration=53.644482836 podStartE2EDuration="53.644482836s" podCreationTimestamp="2026-03-07 14:40:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:25.641042597 +0000 UTC m=+127.593179095" watchObservedRunningTime="2026-03-07 14:41:25.644482836 +0000 UTC m=+127.596619334" Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.710207 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-8x4b2"] Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.736941 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d1657d7-a787-4648-91f6-e0ff987cede2-utilities\") pod \"certified-operators-dfqzr\" (UID: \"0d1657d7-a787-4648-91f6-e0ff987cede2\") " pod="openshift-marketplace/certified-operators-dfqzr" Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.737011 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-582l8\" (UniqueName: \"kubernetes.io/projected/0d1657d7-a787-4648-91f6-e0ff987cede2-kube-api-access-582l8\") pod \"certified-operators-dfqzr\" (UID: \"0d1657d7-a787-4648-91f6-e0ff987cede2\") " pod="openshift-marketplace/certified-operators-dfqzr" Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.737048 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d1657d7-a787-4648-91f6-e0ff987cede2-catalog-content\") pod \"certified-operators-dfqzr\" (UID: \"0d1657d7-a787-4648-91f6-e0ff987cede2\") " pod="openshift-marketplace/certified-operators-dfqzr" Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.737080 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:25 crc kubenswrapper[4943]: E0307 14:41:25.737382 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:26.237368886 +0000 UTC m=+128.189505384 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.749895 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-htv6d"] Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.781599 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jrc6h"] Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.782457 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jrc6h" Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.799853 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.805907 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jrc6h"] Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.807069 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-ztm29" podStartSLOduration=53.807037594 podStartE2EDuration="53.807037594s" podCreationTimestamp="2026-03-07 14:40:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:25.795951786 +0000 UTC m=+127.748088284" watchObservedRunningTime="2026-03-07 14:41:25.807037594 +0000 UTC m=+127.759174112" Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.838348 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:25 crc kubenswrapper[4943]: E0307 14:41:25.838641 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:26.338625793 +0000 UTC m=+128.290762281 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.838672 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d1657d7-a787-4648-91f6-e0ff987cede2-catalog-content\") pod \"certified-operators-dfqzr\" (UID: \"0d1657d7-a787-4648-91f6-e0ff987cede2\") " pod="openshift-marketplace/certified-operators-dfqzr" Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.838701 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f10fa8f5-f504-40c9-81a5-d1658c1ef268-utilities\") pod \"community-operators-jrc6h\" (UID: \"f10fa8f5-f504-40c9-81a5-d1658c1ef268\") " pod="openshift-marketplace/community-operators-jrc6h" Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.838720 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.838744 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f10fa8f5-f504-40c9-81a5-d1658c1ef268-catalog-content\") pod \"community-operators-jrc6h\" (UID: \"f10fa8f5-f504-40c9-81a5-d1658c1ef268\") " pod="openshift-marketplace/community-operators-jrc6h" Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.838787 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d1657d7-a787-4648-91f6-e0ff987cede2-utilities\") pod \"certified-operators-dfqzr\" (UID: \"0d1657d7-a787-4648-91f6-e0ff987cede2\") " pod="openshift-marketplace/certified-operators-dfqzr" Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.838838 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wthgk\" (UniqueName: \"kubernetes.io/projected/f10fa8f5-f504-40c9-81a5-d1658c1ef268-kube-api-access-wthgk\") pod \"community-operators-jrc6h\" (UID: \"f10fa8f5-f504-40c9-81a5-d1658c1ef268\") " pod="openshift-marketplace/community-operators-jrc6h" Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.838855 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-582l8\" (UniqueName: \"kubernetes.io/projected/0d1657d7-a787-4648-91f6-e0ff987cede2-kube-api-access-582l8\") pod \"certified-operators-dfqzr\" (UID: \"0d1657d7-a787-4648-91f6-e0ff987cede2\") " pod="openshift-marketplace/certified-operators-dfqzr" Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.839304 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d1657d7-a787-4648-91f6-e0ff987cede2-catalog-content\") pod \"certified-operators-dfqzr\" (UID: \"0d1657d7-a787-4648-91f6-e0ff987cede2\") " pod="openshift-marketplace/certified-operators-dfqzr" Mar 07 14:41:25 crc kubenswrapper[4943]: E0307 14:41:25.839517 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:26.339508886 +0000 UTC m=+128.291645374 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.839833 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d1657d7-a787-4648-91f6-e0ff987cede2-utilities\") pod \"certified-operators-dfqzr\" (UID: \"0d1657d7-a787-4648-91f6-e0ff987cede2\") " pod="openshift-marketplace/certified-operators-dfqzr" Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.907995 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-582l8\" (UniqueName: \"kubernetes.io/projected/0d1657d7-a787-4648-91f6-e0ff987cede2-kube-api-access-582l8\") pod \"certified-operators-dfqzr\" (UID: \"0d1657d7-a787-4648-91f6-e0ff987cede2\") " pod="openshift-marketplace/certified-operators-dfqzr" Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.943840 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.944117 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wthgk\" (UniqueName: \"kubernetes.io/projected/f10fa8f5-f504-40c9-81a5-d1658c1ef268-kube-api-access-wthgk\") pod \"community-operators-jrc6h\" (UID: \"f10fa8f5-f504-40c9-81a5-d1658c1ef268\") " pod="openshift-marketplace/community-operators-jrc6h" Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.944179 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f10fa8f5-f504-40c9-81a5-d1658c1ef268-utilities\") pod \"community-operators-jrc6h\" (UID: \"f10fa8f5-f504-40c9-81a5-d1658c1ef268\") " pod="openshift-marketplace/community-operators-jrc6h" Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.944215 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f10fa8f5-f504-40c9-81a5-d1658c1ef268-catalog-content\") pod \"community-operators-jrc6h\" (UID: \"f10fa8f5-f504-40c9-81a5-d1658c1ef268\") " pod="openshift-marketplace/community-operators-jrc6h" Mar 07 14:41:25 crc kubenswrapper[4943]: E0307 14:41:25.944249 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:26.444223683 +0000 UTC m=+128.396360181 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.944659 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f10fa8f5-f504-40c9-81a5-d1658c1ef268-catalog-content\") pod \"community-operators-jrc6h\" (UID: \"f10fa8f5-f504-40c9-81a5-d1658c1ef268\") " pod="openshift-marketplace/community-operators-jrc6h" Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.944730 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f10fa8f5-f504-40c9-81a5-d1658c1ef268-utilities\") pod \"community-operators-jrc6h\" (UID: \"f10fa8f5-f504-40c9-81a5-d1658c1ef268\") " pod="openshift-marketplace/community-operators-jrc6h" Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.945985 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=1.9459575180000002 podStartE2EDuration="1.945957518s" podCreationTimestamp="2026-03-07 14:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:25.867988105 +0000 UTC m=+127.820124603" watchObservedRunningTime="2026-03-07 14:41:25.945957518 +0000 UTC m=+127.898094016" Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.977154 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pv7b4" podStartSLOduration=53.977129757 podStartE2EDuration="53.977129757s" podCreationTimestamp="2026-03-07 14:40:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:25.976684985 +0000 UTC m=+127.928821493" watchObservedRunningTime="2026-03-07 14:41:25.977129757 +0000 UTC m=+127.929266255" Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.983475 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-j89vd" podStartSLOduration=53.983455181 podStartE2EDuration="53.983455181s" podCreationTimestamp="2026-03-07 14:40:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:25.926213606 +0000 UTC m=+127.878350114" watchObservedRunningTime="2026-03-07 14:41:25.983455181 +0000 UTC m=+127.935591679" Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.990052 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-vkcxl"] Mar 07 14:41:25 crc kubenswrapper[4943]: I0307 14:41:25.991920 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vkcxl" Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.028383 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vkcxl"] Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.034009 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dfqzr" Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.047984 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:26 crc kubenswrapper[4943]: E0307 14:41:26.049102 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:26.549078843 +0000 UTC m=+128.501215341 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.050842 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wthgk\" (UniqueName: \"kubernetes.io/projected/f10fa8f5-f504-40c9-81a5-d1658c1ef268-kube-api-access-wthgk\") pod \"community-operators-jrc6h\" (UID: \"f10fa8f5-f504-40c9-81a5-d1658c1ef268\") " pod="openshift-marketplace/community-operators-jrc6h" Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.079103 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-l7g67" podStartSLOduration=54.079076522 podStartE2EDuration="54.079076522s" podCreationTimestamp="2026-03-07 14:40:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:26.031652011 +0000 UTC m=+127.983788519" watchObservedRunningTime="2026-03-07 14:41:26.079076522 +0000 UTC m=+128.031213020" Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.145418 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kdrbq" podStartSLOduration=54.145401022 podStartE2EDuration="54.145401022s" podCreationTimestamp="2026-03-07 14:40:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:26.103002922 +0000 UTC m=+128.055139420" watchObservedRunningTime="2026-03-07 14:41:26.145401022 +0000 UTC m=+128.097537520" Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.156947 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x8mpt" podStartSLOduration=54.156909141 podStartE2EDuration="54.156909141s" podCreationTimestamp="2026-03-07 14:40:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:26.146363587 +0000 UTC m=+128.098500085" watchObservedRunningTime="2026-03-07 14:41:26.156909141 +0000 UTC m=+128.109045639" Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.158736 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9sf6k"] Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.163947 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9sf6k" Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.166268 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.166598 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74a7b99c-3412-438e-a835-1e153a56047b-utilities\") pod \"certified-operators-vkcxl\" (UID: \"74a7b99c-3412-438e-a835-1e153a56047b\") " pod="openshift-marketplace/certified-operators-vkcxl" Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.166634 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mq6g7\" (UniqueName: \"kubernetes.io/projected/74a7b99c-3412-438e-a835-1e153a56047b-kube-api-access-mq6g7\") pod \"certified-operators-vkcxl\" (UID: \"74a7b99c-3412-438e-a835-1e153a56047b\") " pod="openshift-marketplace/certified-operators-vkcxl" Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.166658 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74a7b99c-3412-438e-a835-1e153a56047b-catalog-content\") pod \"certified-operators-vkcxl\" (UID: \"74a7b99c-3412-438e-a835-1e153a56047b\") " pod="openshift-marketplace/certified-operators-vkcxl" Mar 07 14:41:26 crc kubenswrapper[4943]: E0307 14:41:26.166849 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:26.666831008 +0000 UTC m=+128.618967506 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.173584 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jrc6h" Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.178303 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9sf6k"] Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.267655 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-dqpk4" podStartSLOduration=54.267636153 podStartE2EDuration="54.267636153s" podCreationTimestamp="2026-03-07 14:40:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:26.26636445 +0000 UTC m=+128.218500948" watchObservedRunningTime="2026-03-07 14:41:26.267636153 +0000 UTC m=+128.219772641" Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.268128 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mq6g7\" (UniqueName: \"kubernetes.io/projected/74a7b99c-3412-438e-a835-1e153a56047b-kube-api-access-mq6g7\") pod \"certified-operators-vkcxl\" (UID: \"74a7b99c-3412-438e-a835-1e153a56047b\") " pod="openshift-marketplace/certified-operators-vkcxl" Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.269152 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74a7b99c-3412-438e-a835-1e153a56047b-catalog-content\") pod \"certified-operators-vkcxl\" (UID: \"74a7b99c-3412-438e-a835-1e153a56047b\") " pod="openshift-marketplace/certified-operators-vkcxl" Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.269225 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.269419 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8qh5\" (UniqueName: \"kubernetes.io/projected/23fa83e7-df77-47e2-bb98-3344e48ab6bd-kube-api-access-f8qh5\") pod \"community-operators-9sf6k\" (UID: \"23fa83e7-df77-47e2-bb98-3344e48ab6bd\") " pod="openshift-marketplace/community-operators-9sf6k" Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.269604 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23fa83e7-df77-47e2-bb98-3344e48ab6bd-utilities\") pod \"community-operators-9sf6k\" (UID: \"23fa83e7-df77-47e2-bb98-3344e48ab6bd\") " pod="openshift-marketplace/community-operators-9sf6k" Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.269687 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23fa83e7-df77-47e2-bb98-3344e48ab6bd-catalog-content\") pod \"community-operators-9sf6k\" (UID: \"23fa83e7-df77-47e2-bb98-3344e48ab6bd\") " pod="openshift-marketplace/community-operators-9sf6k" Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.269726 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74a7b99c-3412-438e-a835-1e153a56047b-utilities\") pod \"certified-operators-vkcxl\" (UID: \"74a7b99c-3412-438e-a835-1e153a56047b\") " pod="openshift-marketplace/certified-operators-vkcxl" Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.270370 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74a7b99c-3412-438e-a835-1e153a56047b-utilities\") pod \"certified-operators-vkcxl\" (UID: \"74a7b99c-3412-438e-a835-1e153a56047b\") " pod="openshift-marketplace/certified-operators-vkcxl" Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.270606 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74a7b99c-3412-438e-a835-1e153a56047b-catalog-content\") pod \"certified-operators-vkcxl\" (UID: \"74a7b99c-3412-438e-a835-1e153a56047b\") " pod="openshift-marketplace/certified-operators-vkcxl" Mar 07 14:41:26 crc kubenswrapper[4943]: E0307 14:41:26.270943 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:26.770914608 +0000 UTC m=+128.723051106 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.316056 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mq6g7\" (UniqueName: \"kubernetes.io/projected/74a7b99c-3412-438e-a835-1e153a56047b-kube-api-access-mq6g7\") pod \"certified-operators-vkcxl\" (UID: \"74a7b99c-3412-438e-a835-1e153a56047b\") " pod="openshift-marketplace/certified-operators-vkcxl" Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.337992 4943 patch_prober.go:28] interesting pod/router-default-5444994796-nk422 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 07 14:41:26 crc kubenswrapper[4943]: [-]has-synced failed: reason withheld Mar 07 14:41:26 crc kubenswrapper[4943]: [+]process-running ok Mar 07 14:41:26 crc kubenswrapper[4943]: healthz check failed Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.338060 4943 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-nk422" podUID="7ca6333f-d5ff-4c06-a140-3b9c29710528" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.368685 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vkcxl" Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.369796 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hzqrp" podStartSLOduration=54.369775013 podStartE2EDuration="54.369775013s" podCreationTimestamp="2026-03-07 14:40:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:26.325263709 +0000 UTC m=+128.277400207" watchObservedRunningTime="2026-03-07 14:41:26.369775013 +0000 UTC m=+128.321911511" Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.375502 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:26 crc kubenswrapper[4943]: E0307 14:41:26.375746 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:26.875702237 +0000 UTC m=+128.827838735 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.375823 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23fa83e7-df77-47e2-bb98-3344e48ab6bd-utilities\") pod \"community-operators-9sf6k\" (UID: \"23fa83e7-df77-47e2-bb98-3344e48ab6bd\") " pod="openshift-marketplace/community-operators-9sf6k" Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.376043 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23fa83e7-df77-47e2-bb98-3344e48ab6bd-catalog-content\") pod \"community-operators-9sf6k\" (UID: \"23fa83e7-df77-47e2-bb98-3344e48ab6bd\") " pod="openshift-marketplace/community-operators-9sf6k" Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.376233 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.376440 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8qh5\" (UniqueName: \"kubernetes.io/projected/23fa83e7-df77-47e2-bb98-3344e48ab6bd-kube-api-access-f8qh5\") pod \"community-operators-9sf6k\" (UID: \"23fa83e7-df77-47e2-bb98-3344e48ab6bd\") " pod="openshift-marketplace/community-operators-9sf6k" Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.376972 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23fa83e7-df77-47e2-bb98-3344e48ab6bd-utilities\") pod \"community-operators-9sf6k\" (UID: \"23fa83e7-df77-47e2-bb98-3344e48ab6bd\") " pod="openshift-marketplace/community-operators-9sf6k" Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.377082 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23fa83e7-df77-47e2-bb98-3344e48ab6bd-catalog-content\") pod \"community-operators-9sf6k\" (UID: \"23fa83e7-df77-47e2-bb98-3344e48ab6bd\") " pod="openshift-marketplace/community-operators-9sf6k" Mar 07 14:41:26 crc kubenswrapper[4943]: E0307 14:41:26.377722 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:26.877711659 +0000 UTC m=+128.829848377 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.435615 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8qh5\" (UniqueName: \"kubernetes.io/projected/23fa83e7-df77-47e2-bb98-3344e48ab6bd-kube-api-access-f8qh5\") pod \"community-operators-9sf6k\" (UID: \"23fa83e7-df77-47e2-bb98-3344e48ab6bd\") " pod="openshift-marketplace/community-operators-9sf6k" Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.477489 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:26 crc kubenswrapper[4943]: E0307 14:41:26.477702 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:26.977646862 +0000 UTC m=+128.929783360 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.477784 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:26 crc kubenswrapper[4943]: E0307 14:41:26.478331 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:26.978321529 +0000 UTC m=+128.930458027 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.525853 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9gzdv" event={"ID":"b33c7fd7-2d0f-48b7-9932-e7740577280f","Type":"ContainerStarted","Data":"6c71624956fb813761b2cc984f90262721297c03ab4011ecfe9379abe1fc8f0e"} Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.525916 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9gzdv" event={"ID":"b33c7fd7-2d0f-48b7-9932-e7740577280f","Type":"ContainerStarted","Data":"2c84b37fd449eb1e278eff48c7e062255c786df8c860b5de457ecbe841e6cd21"} Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.526266 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9sf6k" Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.556291 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jrc6h"] Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.569391 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9gzdv" podStartSLOduration=54.56936742 podStartE2EDuration="54.56936742s" podCreationTimestamp="2026-03-07 14:40:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:26.547838842 +0000 UTC m=+128.499975350" watchObservedRunningTime="2026-03-07 14:41:26.56936742 +0000 UTC m=+128.521503918" Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.579227 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:26 crc kubenswrapper[4943]: E0307 14:41:26.579760 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:27.07974456 +0000 UTC m=+129.031881058 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.584779 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-n2mhr" event={"ID":"a04d88f1-a308-42cf-9462-1a06f6338b7b","Type":"ContainerStarted","Data":"8920264c887159db1fdfb8866a2ddfd8a0f23a39695f2da71b07cf00d40e8608"} Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.584826 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-n2mhr" event={"ID":"a04d88f1-a308-42cf-9462-1a06f6338b7b","Type":"ContainerStarted","Data":"ea887591d1d2ecfc91eac3bc39f3b122370b87c85929715f777938c5eeb85d26"} Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.585185 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-n2mhr" Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.600373 4943 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-n2mhr container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.600450 4943 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-n2mhr" podUID="a04d88f1-a308-42cf-9462-1a06f6338b7b" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.607893 4943 generic.go:334] "Generic (PLEG): container finished" podID="4d86f7af-32df-419e-a833-d596c2a79ba7" containerID="475ce8f4136a6c8e90ec6b60f45f5723fd1a8cfb5b8023cf9ee62c62531fe39d" exitCode=0 Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.608050 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-w5mcz" event={"ID":"4d86f7af-32df-419e-a833-d596c2a79ba7","Type":"ContainerDied","Data":"475ce8f4136a6c8e90ec6b60f45f5723fd1a8cfb5b8023cf9ee62c62531fe39d"} Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.609085 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-n2mhr" podStartSLOduration=54.60906324 podStartE2EDuration="54.60906324s" podCreationTimestamp="2026-03-07 14:40:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:26.608947967 +0000 UTC m=+128.561084485" watchObservedRunningTime="2026-03-07 14:41:26.60906324 +0000 UTC m=+128.561199738" Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.637572 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dfqzr"] Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.651317 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dvjg6" event={"ID":"0da2f3db-9ac4-42bf-b4b8-a4d538d9f4e6","Type":"ContainerStarted","Data":"34b3ba2db5f691aec5c0e71dffe4d547fd669d86f11ff99b4ddd17e2c1bdd15b"} Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.651369 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dvjg6" event={"ID":"0da2f3db-9ac4-42bf-b4b8-a4d538d9f4e6","Type":"ContainerStarted","Data":"2d104ba5f9ecc2eef346805c97db65ea0db3c890805345ec52acf0c74630f7f8"} Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.660873 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-x5qjf" event={"ID":"00cc0b59-be8e-402b-b6c9-56e9e20c4b9b","Type":"ContainerStarted","Data":"207fa287d9e93db913356202698084e3367708ffa81b77580720303adc05ce1e"} Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.660940 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-x5qjf" event={"ID":"00cc0b59-be8e-402b-b6c9-56e9e20c4b9b","Type":"ContainerStarted","Data":"8ee665428c5b7ea2c85c36b4c831024343c4100a8482b1c77e3ff7b4957c658e"} Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.664508 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-58n5g" event={"ID":"214a6dbd-445a-4b41-a3b2-f3efc93f6ed4","Type":"ContainerStarted","Data":"efa0d7f02552be1d97ca45f1db33406fd18be84017b4bd2f1b5813e8b30c9b5b"} Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.666987 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-ntc42" event={"ID":"e6f60f83-1b2e-4f75-9065-9971654e878a","Type":"ContainerStarted","Data":"58dd7d86ba1a3ae9ba13113a77d0e546253674d16630ce2e3e2bc747dc096eb4"} Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.671029 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-ltd8c" event={"ID":"b541cfee-3ec5-4be3-9acd-d58ebb79f0a2","Type":"ContainerStarted","Data":"8a571b3dc14d4a7bdf3c408fcc68de2c5f6e38208af39e1612322db8497f9c56"} Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.671058 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-ltd8c" event={"ID":"b541cfee-3ec5-4be3-9acd-d58ebb79f0a2","Type":"ContainerStarted","Data":"4e0d8aacce7ff96fa09815c9a35698a18b465ff42d2cd56e73a81d97267a7cf9"} Mar 07 14:41:26 crc kubenswrapper[4943]: W0307 14:41:26.676218 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0d1657d7_a787_4648_91f6_e0ff987cede2.slice/crio-170f3f2e2722a619cc0ea7dcfe8df679d89d9665f25432341c3067e7f0ebb1c0 WatchSource:0}: Error finding container 170f3f2e2722a619cc0ea7dcfe8df679d89d9665f25432341c3067e7f0ebb1c0: Status 404 returned error can't find the container with id 170f3f2e2722a619cc0ea7dcfe8df679d89d9665f25432341c3067e7f0ebb1c0 Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.676850 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-zgx7p" event={"ID":"698827d2-21c9-4856-a9cd-17bb3df88bce","Type":"ContainerStarted","Data":"ca33c4e724afdffe666bc1580b122143e7e9a40aa9ab446c9f12111b8d91fdff"} Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.680881 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:26 crc kubenswrapper[4943]: E0307 14:41:26.682178 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:27.182159207 +0000 UTC m=+129.134295705 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.684333 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8x4b2" event={"ID":"80ba5362-78a1-4ed1-a6be-fbda0dbe72df","Type":"ContainerStarted","Data":"43fba8e6d4a7f61e8c520db38d697f20977e1c61031ecebd354e79ac39f5f63a"} Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.715363 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lk96z" event={"ID":"df59699d-5df8-401e-8bb0-0afa130ac06b","Type":"ContainerStarted","Data":"9002952714c551932e6d58b8074240200b240b93691855b6ae9be87a7ef3e8a7"} Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.715409 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lk96z" event={"ID":"df59699d-5df8-401e-8bb0-0afa130ac06b","Type":"ContainerStarted","Data":"a793836bb78ee4ecf7eaa31d6178127e3da29397dfdd32ef7fbdef145ba38d6e"} Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.738117 4943 generic.go:334] "Generic (PLEG): container finished" podID="7601d311-698b-4389-a9a2-e4e24fae23ea" containerID="77e2176d82c678c6306ab7d0f7ca6234ef2c1b3bce73ba85d718298ef1acdc57" exitCode=0 Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.738206 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29548230-zr6fd" event={"ID":"7601d311-698b-4389-a9a2-e4e24fae23ea","Type":"ContainerDied","Data":"77e2176d82c678c6306ab7d0f7ca6234ef2c1b3bce73ba85d718298ef1acdc57"} Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.763138 4943 generic.go:334] "Generic (PLEG): container finished" podID="dc37c9ce-2d22-4a80-be9b-0428621f7ca5" containerID="3395b10931b62e876afc9d492918ff8a5b58f0910190a925ae63c638d014a4e9" exitCode=0 Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.763282 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hr5px" event={"ID":"dc37c9ce-2d22-4a80-be9b-0428621f7ca5","Type":"ContainerDied","Data":"3395b10931b62e876afc9d492918ff8a5b58f0910190a925ae63c638d014a4e9"} Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.763819 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dvjg6" podStartSLOduration=54.763797505 podStartE2EDuration="54.763797505s" podCreationTimestamp="2026-03-07 14:40:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:26.716566879 +0000 UTC m=+128.668703377" watchObservedRunningTime="2026-03-07 14:41:26.763797505 +0000 UTC m=+128.715934003" Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.800260 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:26 crc kubenswrapper[4943]: E0307 14:41:26.803130 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:27.303098804 +0000 UTC m=+129.255235302 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.854327 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qml2f" event={"ID":"f94ec039-60b3-4ecb-b784-04a7ebaf4848","Type":"ContainerStarted","Data":"d464e4b6d8dfe3fe80d8d1184321e48a64b76181bb28150b4da40b92ccbd38a1"} Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.854861 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6qmrb" Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.854876 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6qmrb" event={"ID":"51fb375a-53bd-420a-930f-6b6cdb69f793","Type":"ContainerStarted","Data":"7cac3eb0c33c81398a5af9fee7fb73796c70a66039922429a8b1985192c49659"} Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.860793 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dr8bw" event={"ID":"28995414-2bd3-4b50-90ec-0fbb63a15ef6","Type":"ContainerStarted","Data":"3e09692d8330fd85d83c3f08631de133ded041161fcc4e44ef023f875cf30b3b"} Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.860898 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dr8bw" event={"ID":"28995414-2bd3-4b50-90ec-0fbb63a15ef6","Type":"ContainerStarted","Data":"84f7eb58d9f7eeb4117d4b300c7ba5ec1d13bd74946340733a44675b524f9af0"} Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.860918 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dr8bw" event={"ID":"28995414-2bd3-4b50-90ec-0fbb63a15ef6","Type":"ContainerStarted","Data":"e4d1ebdab59c77d1749e306f2833dfb05fedb9a266686c80820d228f6e441035"} Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.874404 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-ltd8c" podStartSLOduration=54.874377114 podStartE2EDuration="54.874377114s" podCreationTimestamp="2026-03-07 14:40:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:26.765363345 +0000 UTC m=+128.717499853" watchObservedRunningTime="2026-03-07 14:41:26.874377114 +0000 UTC m=+128.826513612" Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.876153 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vkcxl"] Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.880774 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-8jjq4" event={"ID":"93cd8321-367e-4f02-9f8e-2e044e0f1d89","Type":"ContainerStarted","Data":"4aa2e322950aa91befa17268256df7247feba6b205c3b5831f6660188a0b79da"} Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.881875 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-8jjq4" Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.889301 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-zgx7p" podStartSLOduration=54.88927577 podStartE2EDuration="54.88927577s" podCreationTimestamp="2026-03-07 14:40:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:26.852048764 +0000 UTC m=+128.804185262" watchObservedRunningTime="2026-03-07 14:41:26.88927577 +0000 UTC m=+128.841412278" Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.892534 4943 patch_prober.go:28] interesting pod/downloads-7954f5f757-8jjq4 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.892619 4943 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-8jjq4" podUID="93cd8321-367e-4f02-9f8e-2e044e0f1d89" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.893054 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-58n5g" podStartSLOduration=54.893029087 podStartE2EDuration="54.893029087s" podCreationTimestamp="2026-03-07 14:40:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:26.889978258 +0000 UTC m=+128.842114766" watchObservedRunningTime="2026-03-07 14:41:26.893029087 +0000 UTC m=+128.845165585" Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.895751 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hzqrp" event={"ID":"e8c3aeaf-47bf-48ab-b029-2f19e51dae4a","Type":"ContainerStarted","Data":"c75059203bb3fc5c7ddcd7f7fa809cb188bc7b8ef7eca08a097a3ea4502b6e79"} Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.906884 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:26 crc kubenswrapper[4943]: E0307 14:41:26.907661 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:27.407645687 +0000 UTC m=+129.359782185 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.909126 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-lg6mk" event={"ID":"b7960fae-df40-482f-a9ab-790fa11aaf8f","Type":"ContainerStarted","Data":"0dedd2640efc66ee6b6f7894cb24f37c2235a01f2c9c92dafb0d9e44ff6da49b"} Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.909215 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-lg6mk" Mar 07 14:41:26 crc kubenswrapper[4943]: W0307 14:41:26.909475 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod74a7b99c_3412_438e_a835_1e153a56047b.slice/crio-3deec9910ce9875b285512643f55abea6acc1ca773058fdcc8e2d214d5bdbc21 WatchSource:0}: Error finding container 3deec9910ce9875b285512643f55abea6acc1ca773058fdcc8e2d214d5bdbc21: Status 404 returned error can't find the container with id 3deec9910ce9875b285512643f55abea6acc1ca773058fdcc8e2d214d5bdbc21 Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.917286 4943 patch_prober.go:28] interesting pod/console-operator-58897d9998-lg6mk container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.23:8443/readyz\": dial tcp 10.217.0.23:8443: connect: connection refused" start-of-body= Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.917337 4943 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-lg6mk" podUID="b7960fae-df40-482f-a9ab-790fa11aaf8f" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.23:8443/readyz\": dial tcp 10.217.0.23:8443: connect: connection refused" Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.930596 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-x5qjf" podStartSLOduration=54.930578142 podStartE2EDuration="54.930578142s" podCreationTimestamp="2026-03-07 14:40:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:26.929310789 +0000 UTC m=+128.881447287" watchObservedRunningTime="2026-03-07 14:41:26.930578142 +0000 UTC m=+128.882714640" Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.930822 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-bk5xc" event={"ID":"c400fdf2-390c-4c48-8c3e-4854d2e0b58f","Type":"ContainerStarted","Data":"00804e84bb8e7a8dc0732385206bd3738e9ef1164354fda5339cb471b42c691a"} Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.940300 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-2rd7t" event={"ID":"23baba8c-4f55-4188-8967-61093c57e913","Type":"ContainerStarted","Data":"2bb4011aafe424d4573b01ec1b5de1ebb7d4cb1e69eb6251e3c7a992a8744026"} Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.941116 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-2rd7t" Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.946580 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-multus/cni-sysctl-allowlist-ds-htv6d" podUID="0de7b0c4-4acf-40de-867d-1fc33069fa51" containerName="kube-multus-additional-cni-plugins" containerID="cri-o://5ad27bce7efee616ce74cde00b2b89d44c20dc06fb06394eddde07f604433c89" gracePeriod=30 Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.947778 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" event={"ID":"31ce0220-0f53-4d63-aa7b-99357e9c6b04","Type":"ContainerStarted","Data":"6158936056aea525e30acbce078252b346960a76806c3229d50779a81c5820bc"} Mar 07 14:41:26 crc kubenswrapper[4943]: I0307 14:41:26.986367 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6qmrb" Mar 07 14:41:27 crc kubenswrapper[4943]: I0307 14:41:27.007699 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:27 crc kubenswrapper[4943]: E0307 14:41:27.008937 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:27.508906594 +0000 UTC m=+129.461043092 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:27 crc kubenswrapper[4943]: I0307 14:41:27.010434 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lk96z" podStartSLOduration=55.010396152 podStartE2EDuration="55.010396152s" podCreationTimestamp="2026-03-07 14:40:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:26.98101655 +0000 UTC m=+128.933153058" watchObservedRunningTime="2026-03-07 14:41:27.010396152 +0000 UTC m=+128.962532810" Mar 07 14:41:27 crc kubenswrapper[4943]: I0307 14:41:27.039423 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-ntc42" podStartSLOduration=55.039394715 podStartE2EDuration="55.039394715s" podCreationTimestamp="2026-03-07 14:40:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:27.014133559 +0000 UTC m=+128.966270057" watchObservedRunningTime="2026-03-07 14:41:27.039394715 +0000 UTC m=+128.991531213" Mar 07 14:41:27 crc kubenswrapper[4943]: I0307 14:41:27.074107 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-bk5xc" podStartSLOduration=55.074079655 podStartE2EDuration="55.074079655s" podCreationTimestamp="2026-03-07 14:40:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:27.073228223 +0000 UTC m=+129.025364721" watchObservedRunningTime="2026-03-07 14:41:27.074079655 +0000 UTC m=+129.026216153" Mar 07 14:41:27 crc kubenswrapper[4943]: I0307 14:41:27.108501 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dr8bw" podStartSLOduration=55.108467717 podStartE2EDuration="55.108467717s" podCreationTimestamp="2026-03-07 14:40:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:27.099435582 +0000 UTC m=+129.051572080" watchObservedRunningTime="2026-03-07 14:41:27.108467717 +0000 UTC m=+129.060604215" Mar 07 14:41:27 crc kubenswrapper[4943]: I0307 14:41:27.109809 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:27 crc kubenswrapper[4943]: E0307 14:41:27.115990 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:27.615975862 +0000 UTC m=+129.568112360 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:27 crc kubenswrapper[4943]: I0307 14:41:27.124959 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-2rd7t" podStartSLOduration=9.124939784 podStartE2EDuration="9.124939784s" podCreationTimestamp="2026-03-07 14:41:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:27.119103843 +0000 UTC m=+129.071240351" watchObservedRunningTime="2026-03-07 14:41:27.124939784 +0000 UTC m=+129.077076282" Mar 07 14:41:27 crc kubenswrapper[4943]: I0307 14:41:27.174688 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6qmrb" podStartSLOduration=55.174667804 podStartE2EDuration="55.174667804s" podCreationTimestamp="2026-03-07 14:40:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:27.14406115 +0000 UTC m=+129.096197668" watchObservedRunningTime="2026-03-07 14:41:27.174667804 +0000 UTC m=+129.126804302" Mar 07 14:41:27 crc kubenswrapper[4943]: I0307 14:41:27.176284 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qml2f" podStartSLOduration=55.176277506 podStartE2EDuration="55.176277506s" podCreationTimestamp="2026-03-07 14:40:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:27.175254799 +0000 UTC m=+129.127391297" watchObservedRunningTime="2026-03-07 14:41:27.176277506 +0000 UTC m=+129.128414004" Mar 07 14:41:27 crc kubenswrapper[4943]: I0307 14:41:27.213059 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:27 crc kubenswrapper[4943]: E0307 14:41:27.213213 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:27.713182143 +0000 UTC m=+129.665318641 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:27 crc kubenswrapper[4943]: I0307 14:41:27.213716 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:27 crc kubenswrapper[4943]: E0307 14:41:27.214368 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:27.714350384 +0000 UTC m=+129.666486882 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:27 crc kubenswrapper[4943]: I0307 14:41:27.218324 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-8jjq4" podStartSLOduration=55.218297886 podStartE2EDuration="55.218297886s" podCreationTimestamp="2026-03-07 14:40:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:27.212036084 +0000 UTC m=+129.164172582" watchObservedRunningTime="2026-03-07 14:41:27.218297886 +0000 UTC m=+129.170434384" Mar 07 14:41:27 crc kubenswrapper[4943]: I0307 14:41:27.219300 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9sf6k"] Mar 07 14:41:27 crc kubenswrapper[4943]: I0307 14:41:27.290770 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-lg6mk" podStartSLOduration=55.290749276 podStartE2EDuration="55.290749276s" podCreationTimestamp="2026-03-07 14:40:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:27.279560726 +0000 UTC m=+129.231697224" watchObservedRunningTime="2026-03-07 14:41:27.290749276 +0000 UTC m=+129.242885774" Mar 07 14:41:27 crc kubenswrapper[4943]: I0307 14:41:27.315383 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:27 crc kubenswrapper[4943]: E0307 14:41:27.315809 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:27.815780055 +0000 UTC m=+129.767916553 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:27 crc kubenswrapper[4943]: I0307 14:41:27.347860 4943 patch_prober.go:28] interesting pod/router-default-5444994796-nk422 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 07 14:41:27 crc kubenswrapper[4943]: [-]has-synced failed: reason withheld Mar 07 14:41:27 crc kubenswrapper[4943]: [+]process-running ok Mar 07 14:41:27 crc kubenswrapper[4943]: healthz check failed Mar 07 14:41:27 crc kubenswrapper[4943]: I0307 14:41:27.347952 4943 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-nk422" podUID="7ca6333f-d5ff-4c06-a140-3b9c29710528" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 07 14:41:27 crc kubenswrapper[4943]: W0307 14:41:27.356495 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod23fa83e7_df77_47e2_bb98_3344e48ab6bd.slice/crio-9139f60cef576a3909eb49dc2cb92a68a6bb168e3dc55e778d8644ab62cd64f2 WatchSource:0}: Error finding container 9139f60cef576a3909eb49dc2cb92a68a6bb168e3dc55e778d8644ab62cd64f2: Status 404 returned error can't find the container with id 9139f60cef576a3909eb49dc2cb92a68a6bb168e3dc55e778d8644ab62cd64f2 Mar 07 14:41:27 crc kubenswrapper[4943]: I0307 14:41:27.425679 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:27 crc kubenswrapper[4943]: E0307 14:41:27.426357 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:27.926343114 +0000 UTC m=+129.878479612 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:27 crc kubenswrapper[4943]: I0307 14:41:27.527400 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:27 crc kubenswrapper[4943]: E0307 14:41:27.527710 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:28.027691653 +0000 UTC m=+129.979828151 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:27 crc kubenswrapper[4943]: I0307 14:41:27.540589 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xwdpr"] Mar 07 14:41:27 crc kubenswrapper[4943]: I0307 14:41:27.543997 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xwdpr" Mar 07 14:41:27 crc kubenswrapper[4943]: I0307 14:41:27.550446 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 07 14:41:27 crc kubenswrapper[4943]: I0307 14:41:27.558212 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xwdpr"] Mar 07 14:41:27 crc kubenswrapper[4943]: I0307 14:41:27.633006 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:27 crc kubenswrapper[4943]: E0307 14:41:27.633501 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:28.133485098 +0000 UTC m=+130.085621586 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:27 crc kubenswrapper[4943]: I0307 14:41:27.734755 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:27 crc kubenswrapper[4943]: I0307 14:41:27.734914 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmd9h\" (UniqueName: \"kubernetes.io/projected/254f0c0f-ee54-434e-bf7b-bf2e5274fdc9-kube-api-access-kmd9h\") pod \"redhat-marketplace-xwdpr\" (UID: \"254f0c0f-ee54-434e-bf7b-bf2e5274fdc9\") " pod="openshift-marketplace/redhat-marketplace-xwdpr" Mar 07 14:41:27 crc kubenswrapper[4943]: I0307 14:41:27.734970 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/254f0c0f-ee54-434e-bf7b-bf2e5274fdc9-catalog-content\") pod \"redhat-marketplace-xwdpr\" (UID: \"254f0c0f-ee54-434e-bf7b-bf2e5274fdc9\") " pod="openshift-marketplace/redhat-marketplace-xwdpr" Mar 07 14:41:27 crc kubenswrapper[4943]: I0307 14:41:27.735056 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/254f0c0f-ee54-434e-bf7b-bf2e5274fdc9-utilities\") pod \"redhat-marketplace-xwdpr\" (UID: \"254f0c0f-ee54-434e-bf7b-bf2e5274fdc9\") " pod="openshift-marketplace/redhat-marketplace-xwdpr" Mar 07 14:41:27 crc kubenswrapper[4943]: E0307 14:41:27.735148 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:28.235133785 +0000 UTC m=+130.187270283 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:27 crc kubenswrapper[4943]: I0307 14:41:27.783752 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 07 14:41:27 crc kubenswrapper[4943]: I0307 14:41:27.784413 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 07 14:41:27 crc kubenswrapper[4943]: I0307 14:41:27.786578 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Mar 07 14:41:27 crc kubenswrapper[4943]: I0307 14:41:27.787407 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Mar 07 14:41:27 crc kubenswrapper[4943]: I0307 14:41:27.804918 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 07 14:41:27 crc kubenswrapper[4943]: I0307 14:41:27.835992 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/254f0c0f-ee54-434e-bf7b-bf2e5274fdc9-utilities\") pod \"redhat-marketplace-xwdpr\" (UID: \"254f0c0f-ee54-434e-bf7b-bf2e5274fdc9\") " pod="openshift-marketplace/redhat-marketplace-xwdpr" Mar 07 14:41:27 crc kubenswrapper[4943]: I0307 14:41:27.836040 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmd9h\" (UniqueName: \"kubernetes.io/projected/254f0c0f-ee54-434e-bf7b-bf2e5274fdc9-kube-api-access-kmd9h\") pod \"redhat-marketplace-xwdpr\" (UID: \"254f0c0f-ee54-434e-bf7b-bf2e5274fdc9\") " pod="openshift-marketplace/redhat-marketplace-xwdpr" Mar 07 14:41:27 crc kubenswrapper[4943]: I0307 14:41:27.836067 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/254f0c0f-ee54-434e-bf7b-bf2e5274fdc9-catalog-content\") pod \"redhat-marketplace-xwdpr\" (UID: \"254f0c0f-ee54-434e-bf7b-bf2e5274fdc9\") " pod="openshift-marketplace/redhat-marketplace-xwdpr" Mar 07 14:41:27 crc kubenswrapper[4943]: I0307 14:41:27.836089 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:27 crc kubenswrapper[4943]: E0307 14:41:27.836400 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:28.336388942 +0000 UTC m=+130.288525440 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:27 crc kubenswrapper[4943]: I0307 14:41:27.836891 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/254f0c0f-ee54-434e-bf7b-bf2e5274fdc9-catalog-content\") pod \"redhat-marketplace-xwdpr\" (UID: \"254f0c0f-ee54-434e-bf7b-bf2e5274fdc9\") " pod="openshift-marketplace/redhat-marketplace-xwdpr" Mar 07 14:41:27 crc kubenswrapper[4943]: I0307 14:41:27.837248 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/254f0c0f-ee54-434e-bf7b-bf2e5274fdc9-utilities\") pod \"redhat-marketplace-xwdpr\" (UID: \"254f0c0f-ee54-434e-bf7b-bf2e5274fdc9\") " pod="openshift-marketplace/redhat-marketplace-xwdpr" Mar 07 14:41:27 crc kubenswrapper[4943]: I0307 14:41:27.869293 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmd9h\" (UniqueName: \"kubernetes.io/projected/254f0c0f-ee54-434e-bf7b-bf2e5274fdc9-kube-api-access-kmd9h\") pod \"redhat-marketplace-xwdpr\" (UID: \"254f0c0f-ee54-434e-bf7b-bf2e5274fdc9\") " pod="openshift-marketplace/redhat-marketplace-xwdpr" Mar 07 14:41:27 crc kubenswrapper[4943]: I0307 14:41:27.937822 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:27 crc kubenswrapper[4943]: E0307 14:41:27.938260 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:28.438226554 +0000 UTC m=+130.390363042 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:27 crc kubenswrapper[4943]: I0307 14:41:27.938896 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f0772d18-4073-4e7f-b4e8-1515daa81992-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"f0772d18-4073-4e7f-b4e8-1515daa81992\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 07 14:41:27 crc kubenswrapper[4943]: I0307 14:41:27.939017 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:27 crc kubenswrapper[4943]: I0307 14:41:27.939256 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f0772d18-4073-4e7f-b4e8-1515daa81992-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"f0772d18-4073-4e7f-b4e8-1515daa81992\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 07 14:41:27 crc kubenswrapper[4943]: E0307 14:41:27.939379 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:28.439370784 +0000 UTC m=+130.391507282 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:27 crc kubenswrapper[4943]: I0307 14:41:27.952682 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-p767n"] Mar 07 14:41:27 crc kubenswrapper[4943]: I0307 14:41:27.954306 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p767n" Mar 07 14:41:27 crc kubenswrapper[4943]: I0307 14:41:27.970225 4943 generic.go:334] "Generic (PLEG): container finished" podID="74a7b99c-3412-438e-a835-1e153a56047b" containerID="b693f2dfa58f9dd4c4a864a088610221ba9a30f644ee4c4624a68a73422f8056" exitCode=0 Mar 07 14:41:27 crc kubenswrapper[4943]: I0307 14:41:27.970366 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vkcxl" event={"ID":"74a7b99c-3412-438e-a835-1e153a56047b","Type":"ContainerDied","Data":"b693f2dfa58f9dd4c4a864a088610221ba9a30f644ee4c4624a68a73422f8056"} Mar 07 14:41:27 crc kubenswrapper[4943]: I0307 14:41:27.970402 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vkcxl" event={"ID":"74a7b99c-3412-438e-a835-1e153a56047b","Type":"ContainerStarted","Data":"3deec9910ce9875b285512643f55abea6acc1ca773058fdcc8e2d214d5bdbc21"} Mar 07 14:41:27 crc kubenswrapper[4943]: I0307 14:41:27.971200 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-p767n"] Mar 07 14:41:27 crc kubenswrapper[4943]: I0307 14:41:27.977675 4943 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 07 14:41:27 crc kubenswrapper[4943]: I0307 14:41:27.984479 4943 generic.go:334] "Generic (PLEG): container finished" podID="0d1657d7-a787-4648-91f6-e0ff987cede2" containerID="a07aa093e1d3c936f3f368f109c8afbcd14d4091f79d0b410b6d4258aa5fe893" exitCode=0 Mar 07 14:41:27 crc kubenswrapper[4943]: I0307 14:41:27.984644 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dfqzr" event={"ID":"0d1657d7-a787-4648-91f6-e0ff987cede2","Type":"ContainerDied","Data":"a07aa093e1d3c936f3f368f109c8afbcd14d4091f79d0b410b6d4258aa5fe893"} Mar 07 14:41:27 crc kubenswrapper[4943]: I0307 14:41:27.984681 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dfqzr" event={"ID":"0d1657d7-a787-4648-91f6-e0ff987cede2","Type":"ContainerStarted","Data":"170f3f2e2722a619cc0ea7dcfe8df679d89d9665f25432341c3067e7f0ebb1c0"} Mar 07 14:41:27 crc kubenswrapper[4943]: I0307 14:41:27.989014 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hr5px" event={"ID":"dc37c9ce-2d22-4a80-be9b-0428621f7ca5","Type":"ContainerStarted","Data":"66e763da7a983358e4631e1e73976fed00648dcdea46fa2c10380e4421ca342c"} Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.006248 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qml2f" event={"ID":"f94ec039-60b3-4ecb-b784-04a7ebaf4848","Type":"ContainerStarted","Data":"c2da7dbb5eee769c99b9cb0ff042ff676604f01cb37a3ae6d9e5c2239d92646b"} Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.012749 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" event={"ID":"31ce0220-0f53-4d63-aa7b-99357e9c6b04","Type":"ContainerStarted","Data":"64735be97cb4b56fae79aae00df6781ca6b2bc8a37a772cbedb7b9239d45514d"} Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.013327 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.015701 4943 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-cnbpx container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.21:6443/healthz\": dial tcp 10.217.0.21:6443: connect: connection refused" start-of-body= Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.015744 4943 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" podUID="31ce0220-0f53-4d63-aa7b-99357e9c6b04" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.21:6443/healthz\": dial tcp 10.217.0.21:6443: connect: connection refused" Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.016303 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8x4b2" event={"ID":"80ba5362-78a1-4ed1-a6be-fbda0dbe72df","Type":"ContainerStarted","Data":"d55b9f043325d8de42d8949f339c0ea0a2420bf9c868f8745396bf003aef721a"} Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.016350 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8x4b2" event={"ID":"80ba5362-78a1-4ed1-a6be-fbda0dbe72df","Type":"ContainerStarted","Data":"8b0fc793c8260f9a4198ed5c8c4e83f340bf3eae4ce4921641508bdd03396243"} Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.025024 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-7p7x5" event={"ID":"ff8cf71a-e914-4244-b1aa-ede95edd30e8","Type":"ContainerStarted","Data":"6ae43edd0124372c4def35779b3a22aba264438887147fd8c7966871eb1aa9e0"} Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.037951 4943 ???:1] "http: TLS handshake error from 192.168.126.11:43074: no serving certificate available for the kubelet" Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.040705 4943 generic.go:334] "Generic (PLEG): container finished" podID="f10fa8f5-f504-40c9-81a5-d1658c1ef268" containerID="eb4adba6696048589e1e42c2302ca9cfc67db345f210720f61ecec85240427e2" exitCode=0 Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.040826 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jrc6h" event={"ID":"f10fa8f5-f504-40c9-81a5-d1658c1ef268","Type":"ContainerDied","Data":"eb4adba6696048589e1e42c2302ca9cfc67db345f210720f61ecec85240427e2"} Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.040863 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jrc6h" event={"ID":"f10fa8f5-f504-40c9-81a5-d1658c1ef268","Type":"ContainerStarted","Data":"3c06a48028f715519a06fd42092c90004ec700a2701f57713cf67b76f542b7a9"} Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.041082 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:28 crc kubenswrapper[4943]: E0307 14:41:28.041509 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:28.541481893 +0000 UTC m=+130.493618391 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.041550 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f0772d18-4073-4e7f-b4e8-1515daa81992-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"f0772d18-4073-4e7f-b4e8-1515daa81992\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.041709 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f0772d18-4073-4e7f-b4e8-1515daa81992-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"f0772d18-4073-4e7f-b4e8-1515daa81992\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.042242 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f0772d18-4073-4e7f-b4e8-1515daa81992-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"f0772d18-4073-4e7f-b4e8-1515daa81992\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.042464 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:28 crc kubenswrapper[4943]: E0307 14:41:28.042883 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:28.542871279 +0000 UTC m=+130.495007777 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.050682 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-w5mcz" event={"ID":"4d86f7af-32df-419e-a833-d596c2a79ba7","Type":"ContainerStarted","Data":"a3b2164e6a8c3c4e495c9b84f9c970cf747b32054f891c47fd4bca066b083f56"} Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.050744 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-w5mcz" event={"ID":"4d86f7af-32df-419e-a833-d596c2a79ba7","Type":"ContainerStarted","Data":"66fe270952e237518dbd448185bb73a28e660412d51176b495b289b6c9aadaf3"} Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.066593 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f0772d18-4073-4e7f-b4e8-1515daa81992-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"f0772d18-4073-4e7f-b4e8-1515daa81992\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.071836 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hr5px" podStartSLOduration=56.07182109 podStartE2EDuration="56.07182109s" podCreationTimestamp="2026-03-07 14:40:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:28.07145328 +0000 UTC m=+130.023589788" watchObservedRunningTime="2026-03-07 14:41:28.07182109 +0000 UTC m=+130.023957588" Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.103657 4943 generic.go:334] "Generic (PLEG): container finished" podID="23fa83e7-df77-47e2-bb98-3344e48ab6bd" containerID="7d1b2a44c52125e4bda1275594a9ef1849cbe7f97f88ce6b456365004f8a327d" exitCode=0 Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.103701 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.103872 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9sf6k" event={"ID":"23fa83e7-df77-47e2-bb98-3344e48ab6bd","Type":"ContainerDied","Data":"7d1b2a44c52125e4bda1275594a9ef1849cbe7f97f88ce6b456365004f8a327d"} Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.103945 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9sf6k" event={"ID":"23fa83e7-df77-47e2-bb98-3344e48ab6bd","Type":"ContainerStarted","Data":"9139f60cef576a3909eb49dc2cb92a68a6bb168e3dc55e778d8644ab62cd64f2"} Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.106439 4943 patch_prober.go:28] interesting pod/downloads-7954f5f757-8jjq4 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.106503 4943 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-8jjq4" podUID="93cd8321-367e-4f02-9f8e-2e044e0f1d89" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.124615 4943 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.125844 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-w5mcz" Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.126982 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-n2mhr" Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.127619 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-w5mcz" Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.127006 4943 patch_prober.go:28] interesting pod/apiserver-76f77b778f-w5mcz container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="Get \"https://10.217.0.6:8443/livez\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.127875 4943 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-w5mcz" podUID="4d86f7af-32df-419e-a833-d596c2a79ba7" containerName="openshift-apiserver" probeResult="failure" output="Get \"https://10.217.0.6:8443/livez\": dial tcp 10.217.0.6:8443: connect: connection refused" Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.144555 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.145266 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a05c7ae-f2a1-4116-9038-ac838fa88af7-catalog-content\") pod \"redhat-marketplace-p767n\" (UID: \"6a05c7ae-f2a1-4116-9038-ac838fa88af7\") " pod="openshift-marketplace/redhat-marketplace-p767n" Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.145358 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rf8ld\" (UniqueName: \"kubernetes.io/projected/6a05c7ae-f2a1-4116-9038-ac838fa88af7-kube-api-access-rf8ld\") pod \"redhat-marketplace-p767n\" (UID: \"6a05c7ae-f2a1-4116-9038-ac838fa88af7\") " pod="openshift-marketplace/redhat-marketplace-p767n" Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.145429 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a05c7ae-f2a1-4116-9038-ac838fa88af7-utilities\") pod \"redhat-marketplace-p767n\" (UID: \"6a05c7ae-f2a1-4116-9038-ac838fa88af7\") " pod="openshift-marketplace/redhat-marketplace-p767n" Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.151762 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hr5px" Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.152129 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hr5px" Mar 07 14:41:28 crc kubenswrapper[4943]: E0307 14:41:28.153496 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:28.653457228 +0000 UTC m=+130.605593726 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.163580 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xwdpr" Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.170531 4943 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-hr5px container/oauth-apiserver namespace/openshift-oauth-apiserver: Startup probe status=failure output="Get \"https://10.217.0.7:8443/livez\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.170597 4943 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hr5px" podUID="dc37c9ce-2d22-4a80-be9b-0428621f7ca5" containerName="oauth-apiserver" probeResult="failure" output="Get \"https://10.217.0.7:8443/livez\": dial tcp 10.217.0.7:8443: connect: connection refused" Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.191146 4943 ???:1] "http: TLS handshake error from 192.168.126.11:43082: no serving certificate available for the kubelet" Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.216556 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-lg6mk" Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.250006 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a05c7ae-f2a1-4116-9038-ac838fa88af7-catalog-content\") pod \"redhat-marketplace-p767n\" (UID: \"6a05c7ae-f2a1-4116-9038-ac838fa88af7\") " pod="openshift-marketplace/redhat-marketplace-p767n" Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.250110 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rf8ld\" (UniqueName: \"kubernetes.io/projected/6a05c7ae-f2a1-4116-9038-ac838fa88af7-kube-api-access-rf8ld\") pod \"redhat-marketplace-p767n\" (UID: \"6a05c7ae-f2a1-4116-9038-ac838fa88af7\") " pod="openshift-marketplace/redhat-marketplace-p767n" Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.250154 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a05c7ae-f2a1-4116-9038-ac838fa88af7-utilities\") pod \"redhat-marketplace-p767n\" (UID: \"6a05c7ae-f2a1-4116-9038-ac838fa88af7\") " pod="openshift-marketplace/redhat-marketplace-p767n" Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.250257 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:28 crc kubenswrapper[4943]: E0307 14:41:28.266878 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:28.76686202 +0000 UTC m=+130.718998518 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.270770 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a05c7ae-f2a1-4116-9038-ac838fa88af7-catalog-content\") pod \"redhat-marketplace-p767n\" (UID: \"6a05c7ae-f2a1-4116-9038-ac838fa88af7\") " pod="openshift-marketplace/redhat-marketplace-p767n" Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.271862 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a05c7ae-f2a1-4116-9038-ac838fa88af7-utilities\") pod \"redhat-marketplace-p767n\" (UID: \"6a05c7ae-f2a1-4116-9038-ac838fa88af7\") " pod="openshift-marketplace/redhat-marketplace-p767n" Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.334300 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" podStartSLOduration=56.334283379 podStartE2EDuration="56.334283379s" podCreationTimestamp="2026-03-07 14:40:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:28.333842738 +0000 UTC m=+130.285979236" watchObservedRunningTime="2026-03-07 14:41:28.334283379 +0000 UTC m=+130.286419877" Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.355122 4943 patch_prober.go:28] interesting pod/router-default-5444994796-nk422 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 07 14:41:28 crc kubenswrapper[4943]: [-]has-synced failed: reason withheld Mar 07 14:41:28 crc kubenswrapper[4943]: [+]process-running ok Mar 07 14:41:28 crc kubenswrapper[4943]: healthz check failed Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.355180 4943 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-nk422" podUID="7ca6333f-d5ff-4c06-a140-3b9c29710528" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.355372 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:28 crc kubenswrapper[4943]: E0307 14:41:28.355762 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:28.855749146 +0000 UTC m=+130.807885644 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.382094 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rf8ld\" (UniqueName: \"kubernetes.io/projected/6a05c7ae-f2a1-4116-9038-ac838fa88af7-kube-api-access-rf8ld\") pod \"redhat-marketplace-p767n\" (UID: \"6a05c7ae-f2a1-4116-9038-ac838fa88af7\") " pod="openshift-marketplace/redhat-marketplace-p767n" Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.421279 4943 ???:1] "http: TLS handshake error from 192.168.126.11:43098: no serving certificate available for the kubelet" Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.457772 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:28 crc kubenswrapper[4943]: E0307 14:41:28.458133 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:28.958120192 +0000 UTC m=+130.910256690 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.467274 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8x4b2" podStartSLOduration=56.467256349 podStartE2EDuration="56.467256349s" podCreationTimestamp="2026-03-07 14:40:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:28.466080698 +0000 UTC m=+130.418217196" watchObservedRunningTime="2026-03-07 14:41:28.467256349 +0000 UTC m=+130.419392847" Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.468334 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-w5mcz" podStartSLOduration=56.468329977 podStartE2EDuration="56.468329977s" podCreationTimestamp="2026-03-07 14:40:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:28.416065291 +0000 UTC m=+130.368201789" watchObservedRunningTime="2026-03-07 14:41:28.468329977 +0000 UTC m=+130.420466475" Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.503266 4943 ???:1] "http: TLS handshake error from 192.168.126.11:43102: no serving certificate available for the kubelet" Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.559184 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:28 crc kubenswrapper[4943]: E0307 14:41:28.559469 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:29.059454641 +0000 UTC m=+131.011591139 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.570089 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p767n" Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.625346 4943 ???:1] "http: TLS handshake error from 192.168.126.11:43104: no serving certificate available for the kubelet" Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.661005 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:28 crc kubenswrapper[4943]: E0307 14:41:28.661343 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:29.161331744 +0000 UTC m=+131.113468242 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.704378 4943 ???:1] "http: TLS handshake error from 192.168.126.11:43114: no serving certificate available for the kubelet" Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.761595 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:28 crc kubenswrapper[4943]: E0307 14:41:28.761869 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-07 14:41:29.261858032 +0000 UTC m=+131.213994530 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.806619 4943 ???:1] "http: TLS handshake error from 192.168.126.11:43128: no serving certificate available for the kubelet" Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.807850 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-f95bw"] Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.808850 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f95bw" Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.836839 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.851761 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-f95bw"] Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.868740 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:28 crc kubenswrapper[4943]: E0307 14:41:28.869188 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-07 14:41:29.369172896 +0000 UTC m=+131.321309394 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-r82fw" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.919060 4943 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-03-07T14:41:28.12465423Z","Handler":null,"Name":""} Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.924870 4943 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.924908 4943 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.933057 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29548230-zr6fd" Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.933891 4943 ???:1] "http: TLS handshake error from 192.168.126.11:43132: no serving certificate available for the kubelet" Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.973943 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.974784 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a41f89c1-5f6d-46bc-801b-a29ebebf4468-catalog-content\") pod \"redhat-operators-f95bw\" (UID: \"a41f89c1-5f6d-46bc-801b-a29ebebf4468\") " pod="openshift-marketplace/redhat-operators-f95bw" Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.974820 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mt9r\" (UniqueName: \"kubernetes.io/projected/a41f89c1-5f6d-46bc-801b-a29ebebf4468-kube-api-access-7mt9r\") pod \"redhat-operators-f95bw\" (UID: \"a41f89c1-5f6d-46bc-801b-a29ebebf4468\") " pod="openshift-marketplace/redhat-operators-f95bw" Mar 07 14:41:28 crc kubenswrapper[4943]: I0307 14:41:28.974951 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a41f89c1-5f6d-46bc-801b-a29ebebf4468-utilities\") pod \"redhat-operators-f95bw\" (UID: \"a41f89c1-5f6d-46bc-801b-a29ebebf4468\") " pod="openshift-marketplace/redhat-operators-f95bw" Mar 07 14:41:29 crc kubenswrapper[4943]: I0307 14:41:29.034236 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 07 14:41:29 crc kubenswrapper[4943]: I0307 14:41:29.076598 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7601d311-698b-4389-a9a2-e4e24fae23ea-secret-volume\") pod \"7601d311-698b-4389-a9a2-e4e24fae23ea\" (UID: \"7601d311-698b-4389-a9a2-e4e24fae23ea\") " Mar 07 14:41:29 crc kubenswrapper[4943]: I0307 14:41:29.076694 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7601d311-698b-4389-a9a2-e4e24fae23ea-config-volume\") pod \"7601d311-698b-4389-a9a2-e4e24fae23ea\" (UID: \"7601d311-698b-4389-a9a2-e4e24fae23ea\") " Mar 07 14:41:29 crc kubenswrapper[4943]: I0307 14:41:29.076731 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jlplm\" (UniqueName: \"kubernetes.io/projected/7601d311-698b-4389-a9a2-e4e24fae23ea-kube-api-access-jlplm\") pod \"7601d311-698b-4389-a9a2-e4e24fae23ea\" (UID: \"7601d311-698b-4389-a9a2-e4e24fae23ea\") " Mar 07 14:41:29 crc kubenswrapper[4943]: I0307 14:41:29.077049 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a41f89c1-5f6d-46bc-801b-a29ebebf4468-catalog-content\") pod \"redhat-operators-f95bw\" (UID: \"a41f89c1-5f6d-46bc-801b-a29ebebf4468\") " pod="openshift-marketplace/redhat-operators-f95bw" Mar 07 14:41:29 crc kubenswrapper[4943]: I0307 14:41:29.077084 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mt9r\" (UniqueName: \"kubernetes.io/projected/a41f89c1-5f6d-46bc-801b-a29ebebf4468-kube-api-access-7mt9r\") pod \"redhat-operators-f95bw\" (UID: \"a41f89c1-5f6d-46bc-801b-a29ebebf4468\") " pod="openshift-marketplace/redhat-operators-f95bw" Mar 07 14:41:29 crc kubenswrapper[4943]: I0307 14:41:29.077117 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:29 crc kubenswrapper[4943]: I0307 14:41:29.077204 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a41f89c1-5f6d-46bc-801b-a29ebebf4468-utilities\") pod \"redhat-operators-f95bw\" (UID: \"a41f89c1-5f6d-46bc-801b-a29ebebf4468\") " pod="openshift-marketplace/redhat-operators-f95bw" Mar 07 14:41:29 crc kubenswrapper[4943]: I0307 14:41:29.077677 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a41f89c1-5f6d-46bc-801b-a29ebebf4468-utilities\") pod \"redhat-operators-f95bw\" (UID: \"a41f89c1-5f6d-46bc-801b-a29ebebf4468\") " pod="openshift-marketplace/redhat-operators-f95bw" Mar 07 14:41:29 crc kubenswrapper[4943]: I0307 14:41:29.078507 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a41f89c1-5f6d-46bc-801b-a29ebebf4468-catalog-content\") pod \"redhat-operators-f95bw\" (UID: \"a41f89c1-5f6d-46bc-801b-a29ebebf4468\") " pod="openshift-marketplace/redhat-operators-f95bw" Mar 07 14:41:29 crc kubenswrapper[4943]: I0307 14:41:29.079108 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7601d311-698b-4389-a9a2-e4e24fae23ea-config-volume" (OuterVolumeSpecName: "config-volume") pod "7601d311-698b-4389-a9a2-e4e24fae23ea" (UID: "7601d311-698b-4389-a9a2-e4e24fae23ea"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:41:29 crc kubenswrapper[4943]: I0307 14:41:29.085094 4943 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 07 14:41:29 crc kubenswrapper[4943]: I0307 14:41:29.085212 4943 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:29 crc kubenswrapper[4943]: I0307 14:41:29.131113 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mt9r\" (UniqueName: \"kubernetes.io/projected/a41f89c1-5f6d-46bc-801b-a29ebebf4468-kube-api-access-7mt9r\") pod \"redhat-operators-f95bw\" (UID: \"a41f89c1-5f6d-46bc-801b-a29ebebf4468\") " pod="openshift-marketplace/redhat-operators-f95bw" Mar 07 14:41:29 crc kubenswrapper[4943]: I0307 14:41:29.131536 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7601d311-698b-4389-a9a2-e4e24fae23ea-kube-api-access-jlplm" (OuterVolumeSpecName: "kube-api-access-jlplm") pod "7601d311-698b-4389-a9a2-e4e24fae23ea" (UID: "7601d311-698b-4389-a9a2-e4e24fae23ea"). InnerVolumeSpecName "kube-api-access-jlplm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:41:29 crc kubenswrapper[4943]: I0307 14:41:29.134112 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7601d311-698b-4389-a9a2-e4e24fae23ea-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "7601d311-698b-4389-a9a2-e4e24fae23ea" (UID: "7601d311-698b-4389-a9a2-e4e24fae23ea"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:41:29 crc kubenswrapper[4943]: I0307 14:41:29.151957 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f95bw" Mar 07 14:41:29 crc kubenswrapper[4943]: I0307 14:41:29.158343 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rvkgz"] Mar 07 14:41:29 crc kubenswrapper[4943]: E0307 14:41:29.158580 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7601d311-698b-4389-a9a2-e4e24fae23ea" containerName="collect-profiles" Mar 07 14:41:29 crc kubenswrapper[4943]: I0307 14:41:29.158592 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="7601d311-698b-4389-a9a2-e4e24fae23ea" containerName="collect-profiles" Mar 07 14:41:29 crc kubenswrapper[4943]: I0307 14:41:29.158692 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="7601d311-698b-4389-a9a2-e4e24fae23ea" containerName="collect-profiles" Mar 07 14:41:29 crc kubenswrapper[4943]: I0307 14:41:29.159478 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rvkgz" Mar 07 14:41:29 crc kubenswrapper[4943]: I0307 14:41:29.173207 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-7p7x5" event={"ID":"ff8cf71a-e914-4244-b1aa-ede95edd30e8","Type":"ContainerStarted","Data":"a0365282629934182316571dc5f2766ebd851d42fe18acd8450f3f2eb8c267de"} Mar 07 14:41:29 crc kubenswrapper[4943]: I0307 14:41:29.180196 4943 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7601d311-698b-4389-a9a2-e4e24fae23ea-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 07 14:41:29 crc kubenswrapper[4943]: I0307 14:41:29.180221 4943 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7601d311-698b-4389-a9a2-e4e24fae23ea-config-volume\") on node \"crc\" DevicePath \"\"" Mar 07 14:41:29 crc kubenswrapper[4943]: I0307 14:41:29.180231 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jlplm\" (UniqueName: \"kubernetes.io/projected/7601d311-698b-4389-a9a2-e4e24fae23ea-kube-api-access-jlplm\") on node \"crc\" DevicePath \"\"" Mar 07 14:41:29 crc kubenswrapper[4943]: I0307 14:41:29.181326 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rvkgz"] Mar 07 14:41:29 crc kubenswrapper[4943]: I0307 14:41:29.228287 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 07 14:41:29 crc kubenswrapper[4943]: I0307 14:41:29.242400 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-r82fw\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:29 crc kubenswrapper[4943]: I0307 14:41:29.242852 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29548230-zr6fd" Mar 07 14:41:29 crc kubenswrapper[4943]: I0307 14:41:29.246663 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29548230-zr6fd" event={"ID":"7601d311-698b-4389-a9a2-e4e24fae23ea","Type":"ContainerDied","Data":"d9cdc1fb988957b10fce7af01e8029c6f8ede0b5a7a6a7856492bfff0b6c326f"} Mar 07 14:41:29 crc kubenswrapper[4943]: I0307 14:41:29.246710 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d9cdc1fb988957b10fce7af01e8029c6f8ede0b5a7a6a7856492bfff0b6c326f" Mar 07 14:41:29 crc kubenswrapper[4943]: I0307 14:41:29.261453 4943 patch_prober.go:28] interesting pod/downloads-7954f5f757-8jjq4 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Mar 07 14:41:29 crc kubenswrapper[4943]: I0307 14:41:29.261502 4943 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-8jjq4" podUID="93cd8321-367e-4f02-9f8e-2e044e0f1d89" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Mar 07 14:41:29 crc kubenswrapper[4943]: I0307 14:41:29.268005 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xwdpr"] Mar 07 14:41:29 crc kubenswrapper[4943]: I0307 14:41:29.282196 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:41:29 crc kubenswrapper[4943]: I0307 14:41:29.283818 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8p98b\" (UniqueName: \"kubernetes.io/projected/4ae1cf01-1400-4500-9b84-8435a6541612-kube-api-access-8p98b\") pod \"redhat-operators-rvkgz\" (UID: \"4ae1cf01-1400-4500-9b84-8435a6541612\") " pod="openshift-marketplace/redhat-operators-rvkgz" Mar 07 14:41:29 crc kubenswrapper[4943]: I0307 14:41:29.283904 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ae1cf01-1400-4500-9b84-8435a6541612-utilities\") pod \"redhat-operators-rvkgz\" (UID: \"4ae1cf01-1400-4500-9b84-8435a6541612\") " pod="openshift-marketplace/redhat-operators-rvkgz" Mar 07 14:41:29 crc kubenswrapper[4943]: I0307 14:41:29.283954 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ae1cf01-1400-4500-9b84-8435a6541612-catalog-content\") pod \"redhat-operators-rvkgz\" (UID: \"4ae1cf01-1400-4500-9b84-8435a6541612\") " pod="openshift-marketplace/redhat-operators-rvkgz" Mar 07 14:41:29 crc kubenswrapper[4943]: W0307 14:41:29.304129 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod254f0c0f_ee54_434e_bf7b_bf2e5274fdc9.slice/crio-001de696f44d832458359bbd56358a8936643cfa6074a217dee572b906259c5c WatchSource:0}: Error finding container 001de696f44d832458359bbd56358a8936643cfa6074a217dee572b906259c5c: Status 404 returned error can't find the container with id 001de696f44d832458359bbd56358a8936643cfa6074a217dee572b906259c5c Mar 07 14:41:29 crc kubenswrapper[4943]: I0307 14:41:29.355044 4943 patch_prober.go:28] interesting pod/router-default-5444994796-nk422 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 07 14:41:29 crc kubenswrapper[4943]: [-]has-synced failed: reason withheld Mar 07 14:41:29 crc kubenswrapper[4943]: [+]process-running ok Mar 07 14:41:29 crc kubenswrapper[4943]: healthz check failed Mar 07 14:41:29 crc kubenswrapper[4943]: I0307 14:41:29.355096 4943 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-nk422" podUID="7ca6333f-d5ff-4c06-a140-3b9c29710528" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 07 14:41:29 crc kubenswrapper[4943]: I0307 14:41:29.386275 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8p98b\" (UniqueName: \"kubernetes.io/projected/4ae1cf01-1400-4500-9b84-8435a6541612-kube-api-access-8p98b\") pod \"redhat-operators-rvkgz\" (UID: \"4ae1cf01-1400-4500-9b84-8435a6541612\") " pod="openshift-marketplace/redhat-operators-rvkgz" Mar 07 14:41:29 crc kubenswrapper[4943]: I0307 14:41:29.386736 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ae1cf01-1400-4500-9b84-8435a6541612-utilities\") pod \"redhat-operators-rvkgz\" (UID: \"4ae1cf01-1400-4500-9b84-8435a6541612\") " pod="openshift-marketplace/redhat-operators-rvkgz" Mar 07 14:41:29 crc kubenswrapper[4943]: I0307 14:41:29.386802 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ae1cf01-1400-4500-9b84-8435a6541612-catalog-content\") pod \"redhat-operators-rvkgz\" (UID: \"4ae1cf01-1400-4500-9b84-8435a6541612\") " pod="openshift-marketplace/redhat-operators-rvkgz" Mar 07 14:41:29 crc kubenswrapper[4943]: I0307 14:41:29.392860 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ae1cf01-1400-4500-9b84-8435a6541612-utilities\") pod \"redhat-operators-rvkgz\" (UID: \"4ae1cf01-1400-4500-9b84-8435a6541612\") " pod="openshift-marketplace/redhat-operators-rvkgz" Mar 07 14:41:29 crc kubenswrapper[4943]: I0307 14:41:29.393661 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ae1cf01-1400-4500-9b84-8435a6541612-catalog-content\") pod \"redhat-operators-rvkgz\" (UID: \"4ae1cf01-1400-4500-9b84-8435a6541612\") " pod="openshift-marketplace/redhat-operators-rvkgz" Mar 07 14:41:29 crc kubenswrapper[4943]: I0307 14:41:29.492349 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 07 14:41:29 crc kubenswrapper[4943]: I0307 14:41:29.493176 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:29 crc kubenswrapper[4943]: E0307 14:41:29.528358 4943 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7601d311_698b_4389_a9a2_e4e24fae23ea.slice/crio-d9cdc1fb988957b10fce7af01e8029c6f8ede0b5a7a6a7856492bfff0b6c326f\": RecentStats: unable to find data in memory cache]" Mar 07 14:41:29 crc kubenswrapper[4943]: I0307 14:41:29.536638 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-n2mhr"] Mar 07 14:41:29 crc kubenswrapper[4943]: I0307 14:41:29.578124 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8p98b\" (UniqueName: \"kubernetes.io/projected/4ae1cf01-1400-4500-9b84-8435a6541612-kube-api-access-8p98b\") pod \"redhat-operators-rvkgz\" (UID: \"4ae1cf01-1400-4500-9b84-8435a6541612\") " pod="openshift-marketplace/redhat-operators-rvkgz" Mar 07 14:41:29 crc kubenswrapper[4943]: I0307 14:41:29.594750 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-6qmrb"] Mar 07 14:41:29 crc kubenswrapper[4943]: I0307 14:41:29.651357 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-f95bw"] Mar 07 14:41:29 crc kubenswrapper[4943]: I0307 14:41:29.680612 4943 ???:1] "http: TLS handshake error from 192.168.126.11:43136: no serving certificate available for the kubelet" Mar 07 14:41:29 crc kubenswrapper[4943]: I0307 14:41:29.763427 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-p767n"] Mar 07 14:41:29 crc kubenswrapper[4943]: I0307 14:41:29.816646 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rvkgz" Mar 07 14:41:29 crc kubenswrapper[4943]: W0307 14:41:29.910045 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6a05c7ae_f2a1_4116_9038_ac838fa88af7.slice/crio-79d319958a42f72270e5dd7104572de0fb9a91005002b405489433c291e60cca WatchSource:0}: Error finding container 79d319958a42f72270e5dd7104572de0fb9a91005002b405489433c291e60cca: Status 404 returned error can't find the container with id 79d319958a42f72270e5dd7104572de0fb9a91005002b405489433c291e60cca Mar 07 14:41:30 crc kubenswrapper[4943]: I0307 14:41:30.262283 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p767n" event={"ID":"6a05c7ae-f2a1-4116-9038-ac838fa88af7","Type":"ContainerStarted","Data":"79d319958a42f72270e5dd7104572de0fb9a91005002b405489433c291e60cca"} Mar 07 14:41:30 crc kubenswrapper[4943]: I0307 14:41:30.266351 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-7p7x5" event={"ID":"ff8cf71a-e914-4244-b1aa-ede95edd30e8","Type":"ContainerStarted","Data":"525a3fc4908ba68952e45cf7aca25d9ebfc32209829588f1a780e6c5f3096d9f"} Mar 07 14:41:30 crc kubenswrapper[4943]: I0307 14:41:30.270600 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"f0772d18-4073-4e7f-b4e8-1515daa81992","Type":"ContainerStarted","Data":"e6260962295a8e577ba859a93c24bc932886529c704dace7d2e8481f0fb5073d"} Mar 07 14:41:30 crc kubenswrapper[4943]: I0307 14:41:30.341090 4943 patch_prober.go:28] interesting pod/router-default-5444994796-nk422 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 07 14:41:30 crc kubenswrapper[4943]: [-]has-synced failed: reason withheld Mar 07 14:41:30 crc kubenswrapper[4943]: [+]process-running ok Mar 07 14:41:30 crc kubenswrapper[4943]: healthz check failed Mar 07 14:41:30 crc kubenswrapper[4943]: I0307 14:41:30.341158 4943 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-nk422" podUID="7ca6333f-d5ff-4c06-a140-3b9c29710528" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 07 14:41:30 crc kubenswrapper[4943]: I0307 14:41:30.343426 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f95bw" event={"ID":"a41f89c1-5f6d-46bc-801b-a29ebebf4468","Type":"ContainerStarted","Data":"7652bbe2473079e8df9950f4a114b06f40bffb90917d1251c67fd891bb0374a0"} Mar 07 14:41:30 crc kubenswrapper[4943]: I0307 14:41:30.361247 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-7p7x5" podStartSLOduration=12.36122563 podStartE2EDuration="12.36122563s" podCreationTimestamp="2026-03-07 14:41:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:30.319183924 +0000 UTC m=+132.271320422" watchObservedRunningTime="2026-03-07 14:41:30.36122563 +0000 UTC m=+132.313362128" Mar 07 14:41:30 crc kubenswrapper[4943]: I0307 14:41:30.364138 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-r82fw"] Mar 07 14:41:30 crc kubenswrapper[4943]: I0307 14:41:30.395439 4943 generic.go:334] "Generic (PLEG): container finished" podID="254f0c0f-ee54-434e-bf7b-bf2e5274fdc9" containerID="85734361061e123d9cb4c89ee7923bcd0395bfaef11c025edf86d71024ca1137" exitCode=0 Mar 07 14:41:30 crc kubenswrapper[4943]: I0307 14:41:30.400573 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xwdpr" event={"ID":"254f0c0f-ee54-434e-bf7b-bf2e5274fdc9","Type":"ContainerDied","Data":"85734361061e123d9cb4c89ee7923bcd0395bfaef11c025edf86d71024ca1137"} Mar 07 14:41:30 crc kubenswrapper[4943]: I0307 14:41:30.400691 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xwdpr" event={"ID":"254f0c0f-ee54-434e-bf7b-bf2e5274fdc9","Type":"ContainerStarted","Data":"001de696f44d832458359bbd56358a8936643cfa6074a217dee572b906259c5c"} Mar 07 14:41:30 crc kubenswrapper[4943]: I0307 14:41:30.400909 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-n2mhr" podUID="a04d88f1-a308-42cf-9462-1a06f6338b7b" containerName="controller-manager" containerID="cri-o://8920264c887159db1fdfb8866a2ddfd8a0f23a39695f2da71b07cf00d40e8608" gracePeriod=30 Mar 07 14:41:30 crc kubenswrapper[4943]: I0307 14:41:30.402623 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6qmrb" podUID="51fb375a-53bd-420a-930f-6b6cdb69f793" containerName="route-controller-manager" containerID="cri-o://7cac3eb0c33c81398a5af9fee7fb73796c70a66039922429a8b1985192c49659" gracePeriod=30 Mar 07 14:41:30 crc kubenswrapper[4943]: I0307 14:41:30.456776 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rvkgz"] Mar 07 14:41:30 crc kubenswrapper[4943]: I0307 14:41:30.783233 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Mar 07 14:41:30 crc kubenswrapper[4943]: I0307 14:41:30.908972 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6qmrb" Mar 07 14:41:30 crc kubenswrapper[4943]: I0307 14:41:30.952188 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/51fb375a-53bd-420a-930f-6b6cdb69f793-serving-cert\") pod \"51fb375a-53bd-420a-930f-6b6cdb69f793\" (UID: \"51fb375a-53bd-420a-930f-6b6cdb69f793\") " Mar 07 14:41:30 crc kubenswrapper[4943]: I0307 14:41:30.952548 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qx4fv\" (UniqueName: \"kubernetes.io/projected/51fb375a-53bd-420a-930f-6b6cdb69f793-kube-api-access-qx4fv\") pod \"51fb375a-53bd-420a-930f-6b6cdb69f793\" (UID: \"51fb375a-53bd-420a-930f-6b6cdb69f793\") " Mar 07 14:41:30 crc kubenswrapper[4943]: I0307 14:41:30.952591 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51fb375a-53bd-420a-930f-6b6cdb69f793-config\") pod \"51fb375a-53bd-420a-930f-6b6cdb69f793\" (UID: \"51fb375a-53bd-420a-930f-6b6cdb69f793\") " Mar 07 14:41:30 crc kubenswrapper[4943]: I0307 14:41:30.952614 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/51fb375a-53bd-420a-930f-6b6cdb69f793-client-ca\") pod \"51fb375a-53bd-420a-930f-6b6cdb69f793\" (UID: \"51fb375a-53bd-420a-930f-6b6cdb69f793\") " Mar 07 14:41:30 crc kubenswrapper[4943]: I0307 14:41:30.953244 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51fb375a-53bd-420a-930f-6b6cdb69f793-client-ca" (OuterVolumeSpecName: "client-ca") pod "51fb375a-53bd-420a-930f-6b6cdb69f793" (UID: "51fb375a-53bd-420a-930f-6b6cdb69f793"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:41:30 crc kubenswrapper[4943]: I0307 14:41:30.953390 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51fb375a-53bd-420a-930f-6b6cdb69f793-config" (OuterVolumeSpecName: "config") pod "51fb375a-53bd-420a-930f-6b6cdb69f793" (UID: "51fb375a-53bd-420a-930f-6b6cdb69f793"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:41:30 crc kubenswrapper[4943]: I0307 14:41:30.967703 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51fb375a-53bd-420a-930f-6b6cdb69f793-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "51fb375a-53bd-420a-930f-6b6cdb69f793" (UID: "51fb375a-53bd-420a-930f-6b6cdb69f793"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:41:30 crc kubenswrapper[4943]: I0307 14:41:30.995149 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51fb375a-53bd-420a-930f-6b6cdb69f793-kube-api-access-qx4fv" (OuterVolumeSpecName: "kube-api-access-qx4fv") pod "51fb375a-53bd-420a-930f-6b6cdb69f793" (UID: "51fb375a-53bd-420a-930f-6b6cdb69f793"). InnerVolumeSpecName "kube-api-access-qx4fv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.008826 4943 ???:1] "http: TLS handshake error from 192.168.126.11:43142: no serving certificate available for the kubelet" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.054146 4943 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/51fb375a-53bd-420a-930f-6b6cdb69f793-client-ca\") on node \"crc\" DevicePath \"\"" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.054179 4943 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/51fb375a-53bd-420a-930f-6b6cdb69f793-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.054190 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qx4fv\" (UniqueName: \"kubernetes.io/projected/51fb375a-53bd-420a-930f-6b6cdb69f793-kube-api-access-qx4fv\") on node \"crc\" DevicePath \"\"" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.054201 4943 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51fb375a-53bd-420a-930f-6b6cdb69f793-config\") on node \"crc\" DevicePath \"\"" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.101418 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-n2mhr" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.154741 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a04d88f1-a308-42cf-9462-1a06f6338b7b-config\") pod \"a04d88f1-a308-42cf-9462-1a06f6338b7b\" (UID: \"a04d88f1-a308-42cf-9462-1a06f6338b7b\") " Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.154814 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a04d88f1-a308-42cf-9462-1a06f6338b7b-client-ca\") pod \"a04d88f1-a308-42cf-9462-1a06f6338b7b\" (UID: \"a04d88f1-a308-42cf-9462-1a06f6338b7b\") " Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.154845 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a04d88f1-a308-42cf-9462-1a06f6338b7b-serving-cert\") pod \"a04d88f1-a308-42cf-9462-1a06f6338b7b\" (UID: \"a04d88f1-a308-42cf-9462-1a06f6338b7b\") " Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.154909 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x46k8\" (UniqueName: \"kubernetes.io/projected/a04d88f1-a308-42cf-9462-1a06f6338b7b-kube-api-access-x46k8\") pod \"a04d88f1-a308-42cf-9462-1a06f6338b7b\" (UID: \"a04d88f1-a308-42cf-9462-1a06f6338b7b\") " Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.154961 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a04d88f1-a308-42cf-9462-1a06f6338b7b-proxy-ca-bundles\") pod \"a04d88f1-a308-42cf-9462-1a06f6338b7b\" (UID: \"a04d88f1-a308-42cf-9462-1a06f6338b7b\") " Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.156161 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a04d88f1-a308-42cf-9462-1a06f6338b7b-client-ca" (OuterVolumeSpecName: "client-ca") pod "a04d88f1-a308-42cf-9462-1a06f6338b7b" (UID: "a04d88f1-a308-42cf-9462-1a06f6338b7b"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.156179 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a04d88f1-a308-42cf-9462-1a06f6338b7b-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "a04d88f1-a308-42cf-9462-1a06f6338b7b" (UID: "a04d88f1-a308-42cf-9462-1a06f6338b7b"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.156257 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a04d88f1-a308-42cf-9462-1a06f6338b7b-config" (OuterVolumeSpecName: "config") pod "a04d88f1-a308-42cf-9462-1a06f6338b7b" (UID: "a04d88f1-a308-42cf-9462-1a06f6338b7b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.160486 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a04d88f1-a308-42cf-9462-1a06f6338b7b-kube-api-access-x46k8" (OuterVolumeSpecName: "kube-api-access-x46k8") pod "a04d88f1-a308-42cf-9462-1a06f6338b7b" (UID: "a04d88f1-a308-42cf-9462-1a06f6338b7b"). InnerVolumeSpecName "kube-api-access-x46k8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.164628 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a04d88f1-a308-42cf-9462-1a06f6338b7b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "a04d88f1-a308-42cf-9462-1a06f6338b7b" (UID: "a04d88f1-a308-42cf-9462-1a06f6338b7b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.255702 4943 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a04d88f1-a308-42cf-9462-1a06f6338b7b-client-ca\") on node \"crc\" DevicePath \"\"" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.255734 4943 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a04d88f1-a308-42cf-9462-1a06f6338b7b-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.255743 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x46k8\" (UniqueName: \"kubernetes.io/projected/a04d88f1-a308-42cf-9462-1a06f6338b7b-kube-api-access-x46k8\") on node \"crc\" DevicePath \"\"" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.255753 4943 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a04d88f1-a308-42cf-9462-1a06f6338b7b-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.255762 4943 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a04d88f1-a308-42cf-9462-1a06f6338b7b-config\") on node \"crc\" DevicePath \"\"" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.338635 4943 patch_prober.go:28] interesting pod/router-default-5444994796-nk422 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 07 14:41:31 crc kubenswrapper[4943]: [-]has-synced failed: reason withheld Mar 07 14:41:31 crc kubenswrapper[4943]: [+]process-running ok Mar 07 14:41:31 crc kubenswrapper[4943]: healthz check failed Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.338726 4943 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-nk422" podUID="7ca6333f-d5ff-4c06-a140-3b9c29710528" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.468708 4943 generic.go:334] "Generic (PLEG): container finished" podID="f0772d18-4073-4e7f-b4e8-1515daa81992" containerID="f24984094c381362448e053d4d490c31c8fa9b010d0c68dad431d1d24f3ae99d" exitCode=0 Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.468786 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"f0772d18-4073-4e7f-b4e8-1515daa81992","Type":"ContainerDied","Data":"f24984094c381362448e053d4d490c31c8fa9b010d0c68dad431d1d24f3ae99d"} Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.486603 4943 generic.go:334] "Generic (PLEG): container finished" podID="4ae1cf01-1400-4500-9b84-8435a6541612" containerID="c348d8eae504da56c00175d59f8486fed8406de64de45371f3818a9673b00289" exitCode=0 Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.486736 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rvkgz" event={"ID":"4ae1cf01-1400-4500-9b84-8435a6541612","Type":"ContainerDied","Data":"c348d8eae504da56c00175d59f8486fed8406de64de45371f3818a9673b00289"} Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.486775 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rvkgz" event={"ID":"4ae1cf01-1400-4500-9b84-8435a6541612","Type":"ContainerStarted","Data":"808f86db1ff79b593ed5ac29dd9cf0bab9698e7d67dec02a4362737858863de2"} Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.492133 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" event={"ID":"a9fcbefd-dfa7-4afd-b198-d872137a9f51","Type":"ContainerStarted","Data":"a179c3c12d12b89207c401aa3d7da0a6ff2852cb58c1e775d073ebdda8006c95"} Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.492190 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" event={"ID":"a9fcbefd-dfa7-4afd-b198-d872137a9f51","Type":"ContainerStarted","Data":"72ebd4d6c80a14ea3189fd90adacea2d525b1ffc057ba0a83e9c5aaf8b1eeb98"} Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.493131 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.503428 4943 generic.go:334] "Generic (PLEG): container finished" podID="a41f89c1-5f6d-46bc-801b-a29ebebf4468" containerID="7e1875f1478912123de7be2c99d34eca0ef570d508fb004db0cd5ab210429a91" exitCode=0 Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.503716 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f95bw" event={"ID":"a41f89c1-5f6d-46bc-801b-a29ebebf4468","Type":"ContainerDied","Data":"7e1875f1478912123de7be2c99d34eca0ef570d508fb004db0cd5ab210429a91"} Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.531042 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" podStartSLOduration=59.53101844 podStartE2EDuration="59.53101844s" podCreationTimestamp="2026-03-07 14:40:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:31.524516375 +0000 UTC m=+133.476652883" watchObservedRunningTime="2026-03-07 14:41:31.53101844 +0000 UTC m=+133.483154948" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.534555 4943 generic.go:334] "Generic (PLEG): container finished" podID="a04d88f1-a308-42cf-9462-1a06f6338b7b" containerID="8920264c887159db1fdfb8866a2ddfd8a0f23a39695f2da71b07cf00d40e8608" exitCode=0 Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.534787 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-n2mhr" event={"ID":"a04d88f1-a308-42cf-9462-1a06f6338b7b","Type":"ContainerDied","Data":"8920264c887159db1fdfb8866a2ddfd8a0f23a39695f2da71b07cf00d40e8608"} Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.534844 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-n2mhr" event={"ID":"a04d88f1-a308-42cf-9462-1a06f6338b7b","Type":"ContainerDied","Data":"ea887591d1d2ecfc91eac3bc39f3b122370b87c85929715f777938c5eeb85d26"} Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.534869 4943 scope.go:117] "RemoveContainer" containerID="8920264c887159db1fdfb8866a2ddfd8a0f23a39695f2da71b07cf00d40e8608" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.535217 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-n2mhr" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.541292 4943 generic.go:334] "Generic (PLEG): container finished" podID="6a05c7ae-f2a1-4116-9038-ac838fa88af7" containerID="5716e592ac5b4912810aeda7b6fc5b837c8c5c98afe83537ed3b4bea0e1fe36d" exitCode=0 Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.541877 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p767n" event={"ID":"6a05c7ae-f2a1-4116-9038-ac838fa88af7","Type":"ContainerDied","Data":"5716e592ac5b4912810aeda7b6fc5b837c8c5c98afe83537ed3b4bea0e1fe36d"} Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.550040 4943 generic.go:334] "Generic (PLEG): container finished" podID="51fb375a-53bd-420a-930f-6b6cdb69f793" containerID="7cac3eb0c33c81398a5af9fee7fb73796c70a66039922429a8b1985192c49659" exitCode=0 Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.550131 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6qmrb" event={"ID":"51fb375a-53bd-420a-930f-6b6cdb69f793","Type":"ContainerDied","Data":"7cac3eb0c33c81398a5af9fee7fb73796c70a66039922429a8b1985192c49659"} Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.550184 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6qmrb" event={"ID":"51fb375a-53bd-420a-930f-6b6cdb69f793","Type":"ContainerDied","Data":"c9e2aff918e1390f32dab0458bb06c2afc74281f3d7feb2fc40f602018a9c95b"} Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.552295 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6qmrb" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.591394 4943 scope.go:117] "RemoveContainer" containerID="8920264c887159db1fdfb8866a2ddfd8a0f23a39695f2da71b07cf00d40e8608" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.608069 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-n2mhr"] Mar 07 14:41:31 crc kubenswrapper[4943]: E0307 14:41:31.608199 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8920264c887159db1fdfb8866a2ddfd8a0f23a39695f2da71b07cf00d40e8608\": container with ID starting with 8920264c887159db1fdfb8866a2ddfd8a0f23a39695f2da71b07cf00d40e8608 not found: ID does not exist" containerID="8920264c887159db1fdfb8866a2ddfd8a0f23a39695f2da71b07cf00d40e8608" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.608513 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8920264c887159db1fdfb8866a2ddfd8a0f23a39695f2da71b07cf00d40e8608"} err="failed to get container status \"8920264c887159db1fdfb8866a2ddfd8a0f23a39695f2da71b07cf00d40e8608\": rpc error: code = NotFound desc = could not find container \"8920264c887159db1fdfb8866a2ddfd8a0f23a39695f2da71b07cf00d40e8608\": container with ID starting with 8920264c887159db1fdfb8866a2ddfd8a0f23a39695f2da71b07cf00d40e8608 not found: ID does not exist" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.608552 4943 scope.go:117] "RemoveContainer" containerID="7cac3eb0c33c81398a5af9fee7fb73796c70a66039922429a8b1985192c49659" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.618763 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-n2mhr"] Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.622327 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-6qmrb"] Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.623670 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-6qmrb"] Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.643584 4943 scope.go:117] "RemoveContainer" containerID="7cac3eb0c33c81398a5af9fee7fb73796c70a66039922429a8b1985192c49659" Mar 07 14:41:31 crc kubenswrapper[4943]: E0307 14:41:31.644783 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7cac3eb0c33c81398a5af9fee7fb73796c70a66039922429a8b1985192c49659\": container with ID starting with 7cac3eb0c33c81398a5af9fee7fb73796c70a66039922429a8b1985192c49659 not found: ID does not exist" containerID="7cac3eb0c33c81398a5af9fee7fb73796c70a66039922429a8b1985192c49659" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.644821 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7cac3eb0c33c81398a5af9fee7fb73796c70a66039922429a8b1985192c49659"} err="failed to get container status \"7cac3eb0c33c81398a5af9fee7fb73796c70a66039922429a8b1985192c49659\": rpc error: code = NotFound desc = could not find container \"7cac3eb0c33c81398a5af9fee7fb73796c70a66039922429a8b1985192c49659\": container with ID starting with 7cac3eb0c33c81398a5af9fee7fb73796c70a66039922429a8b1985192c49659 not found: ID does not exist" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.752739 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-8686b6766c-cqjdw"] Mar 07 14:41:31 crc kubenswrapper[4943]: E0307 14:41:31.753038 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a04d88f1-a308-42cf-9462-1a06f6338b7b" containerName="controller-manager" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.753054 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="a04d88f1-a308-42cf-9462-1a06f6338b7b" containerName="controller-manager" Mar 07 14:41:31 crc kubenswrapper[4943]: E0307 14:41:31.753063 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51fb375a-53bd-420a-930f-6b6cdb69f793" containerName="route-controller-manager" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.753072 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="51fb375a-53bd-420a-930f-6b6cdb69f793" containerName="route-controller-manager" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.753153 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="a04d88f1-a308-42cf-9462-1a06f6338b7b" containerName="controller-manager" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.753166 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="51fb375a-53bd-420a-930f-6b6cdb69f793" containerName="route-controller-manager" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.753589 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-8686b6766c-cqjdw" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.755210 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.755453 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.755612 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.755714 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.755812 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.758159 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-587c5f97b-smswn"] Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.759488 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-587c5f97b-smswn" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.764404 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.764526 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.764626 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.764749 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.764562 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.771497 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.772252 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.779904 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.780737 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-587c5f97b-smswn"] Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.782094 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-8686b6766c-cqjdw"] Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.865542 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/782f34ac-63b1-452d-9157-344361ae3723-serving-cert\") pod \"route-controller-manager-587c5f97b-smswn\" (UID: \"782f34ac-63b1-452d-9157-344361ae3723\") " pod="openshift-route-controller-manager/route-controller-manager-587c5f97b-smswn" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.865600 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3374c43b-61ed-4e63-87d0-2b212fc70a67-proxy-ca-bundles\") pod \"controller-manager-8686b6766c-cqjdw\" (UID: \"3374c43b-61ed-4e63-87d0-2b212fc70a67\") " pod="openshift-controller-manager/controller-manager-8686b6766c-cqjdw" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.865626 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3374c43b-61ed-4e63-87d0-2b212fc70a67-config\") pod \"controller-manager-8686b6766c-cqjdw\" (UID: \"3374c43b-61ed-4e63-87d0-2b212fc70a67\") " pod="openshift-controller-manager/controller-manager-8686b6766c-cqjdw" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.865676 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bcnl\" (UniqueName: \"kubernetes.io/projected/3374c43b-61ed-4e63-87d0-2b212fc70a67-kube-api-access-8bcnl\") pod \"controller-manager-8686b6766c-cqjdw\" (UID: \"3374c43b-61ed-4e63-87d0-2b212fc70a67\") " pod="openshift-controller-manager/controller-manager-8686b6766c-cqjdw" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.865706 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3374c43b-61ed-4e63-87d0-2b212fc70a67-client-ca\") pod \"controller-manager-8686b6766c-cqjdw\" (UID: \"3374c43b-61ed-4e63-87d0-2b212fc70a67\") " pod="openshift-controller-manager/controller-manager-8686b6766c-cqjdw" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.865725 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3374c43b-61ed-4e63-87d0-2b212fc70a67-serving-cert\") pod \"controller-manager-8686b6766c-cqjdw\" (UID: \"3374c43b-61ed-4e63-87d0-2b212fc70a67\") " pod="openshift-controller-manager/controller-manager-8686b6766c-cqjdw" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.865743 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgmwj\" (UniqueName: \"kubernetes.io/projected/782f34ac-63b1-452d-9157-344361ae3723-kube-api-access-kgmwj\") pod \"route-controller-manager-587c5f97b-smswn\" (UID: \"782f34ac-63b1-452d-9157-344361ae3723\") " pod="openshift-route-controller-manager/route-controller-manager-587c5f97b-smswn" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.865781 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/782f34ac-63b1-452d-9157-344361ae3723-client-ca\") pod \"route-controller-manager-587c5f97b-smswn\" (UID: \"782f34ac-63b1-452d-9157-344361ae3723\") " pod="openshift-route-controller-manager/route-controller-manager-587c5f97b-smswn" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.865810 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/782f34ac-63b1-452d-9157-344361ae3723-config\") pod \"route-controller-manager-587c5f97b-smswn\" (UID: \"782f34ac-63b1-452d-9157-344361ae3723\") " pod="openshift-route-controller-manager/route-controller-manager-587c5f97b-smswn" Mar 07 14:41:31 crc kubenswrapper[4943]: E0307 14:41:31.902599 4943 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5ad27bce7efee616ce74cde00b2b89d44c20dc06fb06394eddde07f604433c89" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 07 14:41:31 crc kubenswrapper[4943]: E0307 14:41:31.905501 4943 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5ad27bce7efee616ce74cde00b2b89d44c20dc06fb06394eddde07f604433c89" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 07 14:41:31 crc kubenswrapper[4943]: E0307 14:41:31.907431 4943 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5ad27bce7efee616ce74cde00b2b89d44c20dc06fb06394eddde07f604433c89" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 07 14:41:31 crc kubenswrapper[4943]: E0307 14:41:31.907474 4943 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-htv6d" podUID="0de7b0c4-4acf-40de-867d-1fc33069fa51" containerName="kube-multus-additional-cni-plugins" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.967532 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/782f34ac-63b1-452d-9157-344361ae3723-client-ca\") pod \"route-controller-manager-587c5f97b-smswn\" (UID: \"782f34ac-63b1-452d-9157-344361ae3723\") " pod="openshift-route-controller-manager/route-controller-manager-587c5f97b-smswn" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.967587 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/782f34ac-63b1-452d-9157-344361ae3723-config\") pod \"route-controller-manager-587c5f97b-smswn\" (UID: \"782f34ac-63b1-452d-9157-344361ae3723\") " pod="openshift-route-controller-manager/route-controller-manager-587c5f97b-smswn" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.967640 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/782f34ac-63b1-452d-9157-344361ae3723-serving-cert\") pod \"route-controller-manager-587c5f97b-smswn\" (UID: \"782f34ac-63b1-452d-9157-344361ae3723\") " pod="openshift-route-controller-manager/route-controller-manager-587c5f97b-smswn" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.967660 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3374c43b-61ed-4e63-87d0-2b212fc70a67-proxy-ca-bundles\") pod \"controller-manager-8686b6766c-cqjdw\" (UID: \"3374c43b-61ed-4e63-87d0-2b212fc70a67\") " pod="openshift-controller-manager/controller-manager-8686b6766c-cqjdw" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.967690 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3374c43b-61ed-4e63-87d0-2b212fc70a67-config\") pod \"controller-manager-8686b6766c-cqjdw\" (UID: \"3374c43b-61ed-4e63-87d0-2b212fc70a67\") " pod="openshift-controller-manager/controller-manager-8686b6766c-cqjdw" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.967723 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bcnl\" (UniqueName: \"kubernetes.io/projected/3374c43b-61ed-4e63-87d0-2b212fc70a67-kube-api-access-8bcnl\") pod \"controller-manager-8686b6766c-cqjdw\" (UID: \"3374c43b-61ed-4e63-87d0-2b212fc70a67\") " pod="openshift-controller-manager/controller-manager-8686b6766c-cqjdw" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.967747 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3374c43b-61ed-4e63-87d0-2b212fc70a67-client-ca\") pod \"controller-manager-8686b6766c-cqjdw\" (UID: \"3374c43b-61ed-4e63-87d0-2b212fc70a67\") " pod="openshift-controller-manager/controller-manager-8686b6766c-cqjdw" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.967764 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3374c43b-61ed-4e63-87d0-2b212fc70a67-serving-cert\") pod \"controller-manager-8686b6766c-cqjdw\" (UID: \"3374c43b-61ed-4e63-87d0-2b212fc70a67\") " pod="openshift-controller-manager/controller-manager-8686b6766c-cqjdw" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.967784 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgmwj\" (UniqueName: \"kubernetes.io/projected/782f34ac-63b1-452d-9157-344361ae3723-kube-api-access-kgmwj\") pod \"route-controller-manager-587c5f97b-smswn\" (UID: \"782f34ac-63b1-452d-9157-344361ae3723\") " pod="openshift-route-controller-manager/route-controller-manager-587c5f97b-smswn" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.969149 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/782f34ac-63b1-452d-9157-344361ae3723-client-ca\") pod \"route-controller-manager-587c5f97b-smswn\" (UID: \"782f34ac-63b1-452d-9157-344361ae3723\") " pod="openshift-route-controller-manager/route-controller-manager-587c5f97b-smswn" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.970396 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/782f34ac-63b1-452d-9157-344361ae3723-config\") pod \"route-controller-manager-587c5f97b-smswn\" (UID: \"782f34ac-63b1-452d-9157-344361ae3723\") " pod="openshift-route-controller-manager/route-controller-manager-587c5f97b-smswn" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.970492 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3374c43b-61ed-4e63-87d0-2b212fc70a67-client-ca\") pod \"controller-manager-8686b6766c-cqjdw\" (UID: \"3374c43b-61ed-4e63-87d0-2b212fc70a67\") " pod="openshift-controller-manager/controller-manager-8686b6766c-cqjdw" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.971678 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3374c43b-61ed-4e63-87d0-2b212fc70a67-config\") pod \"controller-manager-8686b6766c-cqjdw\" (UID: \"3374c43b-61ed-4e63-87d0-2b212fc70a67\") " pod="openshift-controller-manager/controller-manager-8686b6766c-cqjdw" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.972577 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3374c43b-61ed-4e63-87d0-2b212fc70a67-proxy-ca-bundles\") pod \"controller-manager-8686b6766c-cqjdw\" (UID: \"3374c43b-61ed-4e63-87d0-2b212fc70a67\") " pod="openshift-controller-manager/controller-manager-8686b6766c-cqjdw" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.978769 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3374c43b-61ed-4e63-87d0-2b212fc70a67-serving-cert\") pod \"controller-manager-8686b6766c-cqjdw\" (UID: \"3374c43b-61ed-4e63-87d0-2b212fc70a67\") " pod="openshift-controller-manager/controller-manager-8686b6766c-cqjdw" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.984578 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgmwj\" (UniqueName: \"kubernetes.io/projected/782f34ac-63b1-452d-9157-344361ae3723-kube-api-access-kgmwj\") pod \"route-controller-manager-587c5f97b-smswn\" (UID: \"782f34ac-63b1-452d-9157-344361ae3723\") " pod="openshift-route-controller-manager/route-controller-manager-587c5f97b-smswn" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.984996 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/782f34ac-63b1-452d-9157-344361ae3723-serving-cert\") pod \"route-controller-manager-587c5f97b-smswn\" (UID: \"782f34ac-63b1-452d-9157-344361ae3723\") " pod="openshift-route-controller-manager/route-controller-manager-587c5f97b-smswn" Mar 07 14:41:31 crc kubenswrapper[4943]: I0307 14:41:31.988889 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bcnl\" (UniqueName: \"kubernetes.io/projected/3374c43b-61ed-4e63-87d0-2b212fc70a67-kube-api-access-8bcnl\") pod \"controller-manager-8686b6766c-cqjdw\" (UID: \"3374c43b-61ed-4e63-87d0-2b212fc70a67\") " pod="openshift-controller-manager/controller-manager-8686b6766c-cqjdw" Mar 07 14:41:32 crc kubenswrapper[4943]: I0307 14:41:32.087337 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-8686b6766c-cqjdw" Mar 07 14:41:32 crc kubenswrapper[4943]: I0307 14:41:32.111517 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-587c5f97b-smswn" Mar 07 14:41:32 crc kubenswrapper[4943]: I0307 14:41:32.338077 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-nk422" Mar 07 14:41:32 crc kubenswrapper[4943]: I0307 14:41:32.353433 4943 patch_prober.go:28] interesting pod/router-default-5444994796-nk422 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 07 14:41:32 crc kubenswrapper[4943]: [-]has-synced failed: reason withheld Mar 07 14:41:32 crc kubenswrapper[4943]: [+]process-running ok Mar 07 14:41:32 crc kubenswrapper[4943]: healthz check failed Mar 07 14:41:32 crc kubenswrapper[4943]: I0307 14:41:32.353781 4943 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-nk422" podUID="7ca6333f-d5ff-4c06-a140-3b9c29710528" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 07 14:41:32 crc kubenswrapper[4943]: I0307 14:41:32.485902 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-8686b6766c-cqjdw"] Mar 07 14:41:32 crc kubenswrapper[4943]: W0307 14:41:32.545462 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3374c43b_61ed_4e63_87d0_2b212fc70a67.slice/crio-15fbdfa9b89b26d60c2c03c2b5c071605fc8464c267769de33d9bc88d7286d44 WatchSource:0}: Error finding container 15fbdfa9b89b26d60c2c03c2b5c071605fc8464c267769de33d9bc88d7286d44: Status 404 returned error can't find the container with id 15fbdfa9b89b26d60c2c03c2b5c071605fc8464c267769de33d9bc88d7286d44 Mar 07 14:41:32 crc kubenswrapper[4943]: I0307 14:41:32.572789 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-8686b6766c-cqjdw" event={"ID":"3374c43b-61ed-4e63-87d0-2b212fc70a67","Type":"ContainerStarted","Data":"15fbdfa9b89b26d60c2c03c2b5c071605fc8464c267769de33d9bc88d7286d44"} Mar 07 14:41:32 crc kubenswrapper[4943]: I0307 14:41:32.775349 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51fb375a-53bd-420a-930f-6b6cdb69f793" path="/var/lib/kubelet/pods/51fb375a-53bd-420a-930f-6b6cdb69f793/volumes" Mar 07 14:41:32 crc kubenswrapper[4943]: I0307 14:41:32.776888 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a04d88f1-a308-42cf-9462-1a06f6338b7b" path="/var/lib/kubelet/pods/a04d88f1-a308-42cf-9462-1a06f6338b7b/volumes" Mar 07 14:41:32 crc kubenswrapper[4943]: I0307 14:41:32.778157 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-587c5f97b-smswn"] Mar 07 14:41:32 crc kubenswrapper[4943]: I0307 14:41:32.889224 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 07 14:41:32 crc kubenswrapper[4943]: I0307 14:41:32.995369 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f0772d18-4073-4e7f-b4e8-1515daa81992-kube-api-access\") pod \"f0772d18-4073-4e7f-b4e8-1515daa81992\" (UID: \"f0772d18-4073-4e7f-b4e8-1515daa81992\") " Mar 07 14:41:32 crc kubenswrapper[4943]: I0307 14:41:32.995477 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f0772d18-4073-4e7f-b4e8-1515daa81992-kubelet-dir\") pod \"f0772d18-4073-4e7f-b4e8-1515daa81992\" (UID: \"f0772d18-4073-4e7f-b4e8-1515daa81992\") " Mar 07 14:41:32 crc kubenswrapper[4943]: I0307 14:41:32.995699 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f0772d18-4073-4e7f-b4e8-1515daa81992-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "f0772d18-4073-4e7f-b4e8-1515daa81992" (UID: "f0772d18-4073-4e7f-b4e8-1515daa81992"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 07 14:41:32 crc kubenswrapper[4943]: I0307 14:41:32.995903 4943 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f0772d18-4073-4e7f-b4e8-1515daa81992-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 07 14:41:33 crc kubenswrapper[4943]: I0307 14:41:33.001889 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0772d18-4073-4e7f-b4e8-1515daa81992-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "f0772d18-4073-4e7f-b4e8-1515daa81992" (UID: "f0772d18-4073-4e7f-b4e8-1515daa81992"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:41:33 crc kubenswrapper[4943]: I0307 14:41:33.097211 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f0772d18-4073-4e7f-b4e8-1515daa81992-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 07 14:41:33 crc kubenswrapper[4943]: I0307 14:41:33.129358 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-w5mcz" Mar 07 14:41:33 crc kubenswrapper[4943]: I0307 14:41:33.136488 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-w5mcz" Mar 07 14:41:33 crc kubenswrapper[4943]: I0307 14:41:33.183048 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-zgx7p" Mar 07 14:41:33 crc kubenswrapper[4943]: I0307 14:41:33.183569 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-zgx7p" Mar 07 14:41:33 crc kubenswrapper[4943]: I0307 14:41:33.183667 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hr5px" Mar 07 14:41:33 crc kubenswrapper[4943]: I0307 14:41:33.191106 4943 patch_prober.go:28] interesting pod/console-f9d7485db-zgx7p container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Mar 07 14:41:33 crc kubenswrapper[4943]: I0307 14:41:33.191193 4943 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-zgx7p" podUID="698827d2-21c9-4856-a9cd-17bb3df88bce" containerName="console" probeResult="failure" output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" Mar 07 14:41:33 crc kubenswrapper[4943]: I0307 14:41:33.200652 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hr5px" Mar 07 14:41:33 crc kubenswrapper[4943]: I0307 14:41:33.347061 4943 patch_prober.go:28] interesting pod/router-default-5444994796-nk422 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 07 14:41:33 crc kubenswrapper[4943]: [-]has-synced failed: reason withheld Mar 07 14:41:33 crc kubenswrapper[4943]: [+]process-running ok Mar 07 14:41:33 crc kubenswrapper[4943]: healthz check failed Mar 07 14:41:33 crc kubenswrapper[4943]: I0307 14:41:33.347129 4943 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-nk422" podUID="7ca6333f-d5ff-4c06-a140-3b9c29710528" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 07 14:41:33 crc kubenswrapper[4943]: I0307 14:41:33.504119 4943 patch_prober.go:28] interesting pod/downloads-7954f5f757-8jjq4 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Mar 07 14:41:33 crc kubenswrapper[4943]: I0307 14:41:33.504192 4943 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-8jjq4" podUID="93cd8321-367e-4f02-9f8e-2e044e0f1d89" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Mar 07 14:41:33 crc kubenswrapper[4943]: I0307 14:41:33.504326 4943 patch_prober.go:28] interesting pod/downloads-7954f5f757-8jjq4 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Mar 07 14:41:33 crc kubenswrapper[4943]: I0307 14:41:33.504384 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-8jjq4" podUID="93cd8321-367e-4f02-9f8e-2e044e0f1d89" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Mar 07 14:41:33 crc kubenswrapper[4943]: I0307 14:41:33.629859 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 07 14:41:33 crc kubenswrapper[4943]: E0307 14:41:33.630384 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0772d18-4073-4e7f-b4e8-1515daa81992" containerName="pruner" Mar 07 14:41:33 crc kubenswrapper[4943]: I0307 14:41:33.630453 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0772d18-4073-4e7f-b4e8-1515daa81992" containerName="pruner" Mar 07 14:41:33 crc kubenswrapper[4943]: I0307 14:41:33.630608 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0772d18-4073-4e7f-b4e8-1515daa81992" containerName="pruner" Mar 07 14:41:33 crc kubenswrapper[4943]: I0307 14:41:33.631063 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 07 14:41:33 crc kubenswrapper[4943]: I0307 14:41:33.639386 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 07 14:41:33 crc kubenswrapper[4943]: I0307 14:41:33.639663 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 07 14:41:33 crc kubenswrapper[4943]: I0307 14:41:33.671012 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-587c5f97b-smswn" event={"ID":"782f34ac-63b1-452d-9157-344361ae3723","Type":"ContainerStarted","Data":"146679eb9aab8c30ac1f2d5598049bbd31f3f7404bca7d92a51c3daa13a1857f"} Mar 07 14:41:33 crc kubenswrapper[4943]: I0307 14:41:33.671096 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-587c5f97b-smswn" event={"ID":"782f34ac-63b1-452d-9157-344361ae3723","Type":"ContainerStarted","Data":"20872e7e7beca543cbd741c4b7eb0b6515ae4e35ac1d0d3bc127a7fc283d8205"} Mar 07 14:41:33 crc kubenswrapper[4943]: I0307 14:41:33.672211 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-587c5f97b-smswn" Mar 07 14:41:33 crc kubenswrapper[4943]: I0307 14:41:33.676031 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 07 14:41:33 crc kubenswrapper[4943]: I0307 14:41:33.678466 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-8686b6766c-cqjdw" event={"ID":"3374c43b-61ed-4e63-87d0-2b212fc70a67","Type":"ContainerStarted","Data":"382a1c9af8e00bcfdb9388ee197b918061c2b683a078650129340e5e1683a761"} Mar 07 14:41:33 crc kubenswrapper[4943]: I0307 14:41:33.679123 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-8686b6766c-cqjdw" Mar 07 14:41:33 crc kubenswrapper[4943]: I0307 14:41:33.703045 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 07 14:41:33 crc kubenswrapper[4943]: I0307 14:41:33.705576 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"f0772d18-4073-4e7f-b4e8-1515daa81992","Type":"ContainerDied","Data":"e6260962295a8e577ba859a93c24bc932886529c704dace7d2e8481f0fb5073d"} Mar 07 14:41:33 crc kubenswrapper[4943]: I0307 14:41:33.705631 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e6260962295a8e577ba859a93c24bc932886529c704dace7d2e8481f0fb5073d" Mar 07 14:41:33 crc kubenswrapper[4943]: I0307 14:41:33.728377 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-8686b6766c-cqjdw" Mar 07 14:41:33 crc kubenswrapper[4943]: I0307 14:41:33.728828 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-587c5f97b-smswn" podStartSLOduration=4.728810943 podStartE2EDuration="4.728810943s" podCreationTimestamp="2026-03-07 14:41:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:33.727845258 +0000 UTC m=+135.679981756" watchObservedRunningTime="2026-03-07 14:41:33.728810943 +0000 UTC m=+135.680947441" Mar 07 14:41:33 crc kubenswrapper[4943]: I0307 14:41:33.731731 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/34a311b8-c902-40d4-ae68-4915984edfb7-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"34a311b8-c902-40d4-ae68-4915984edfb7\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 07 14:41:33 crc kubenswrapper[4943]: I0307 14:41:33.731767 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/34a311b8-c902-40d4-ae68-4915984edfb7-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"34a311b8-c902-40d4-ae68-4915984edfb7\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 07 14:41:33 crc kubenswrapper[4943]: I0307 14:41:33.738953 4943 ???:1] "http: TLS handshake error from 192.168.126.11:43144: no serving certificate available for the kubelet" Mar 07 14:41:33 crc kubenswrapper[4943]: I0307 14:41:33.798565 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-8686b6766c-cqjdw" podStartSLOduration=4.798546014 podStartE2EDuration="4.798546014s" podCreationTimestamp="2026-03-07 14:41:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:33.797813143 +0000 UTC m=+135.749949651" watchObservedRunningTime="2026-03-07 14:41:33.798546014 +0000 UTC m=+135.750682512" Mar 07 14:41:33 crc kubenswrapper[4943]: I0307 14:41:33.834152 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/34a311b8-c902-40d4-ae68-4915984edfb7-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"34a311b8-c902-40d4-ae68-4915984edfb7\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 07 14:41:33 crc kubenswrapper[4943]: I0307 14:41:33.834201 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/34a311b8-c902-40d4-ae68-4915984edfb7-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"34a311b8-c902-40d4-ae68-4915984edfb7\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 07 14:41:33 crc kubenswrapper[4943]: I0307 14:41:33.834912 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/34a311b8-c902-40d4-ae68-4915984edfb7-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"34a311b8-c902-40d4-ae68-4915984edfb7\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 07 14:41:33 crc kubenswrapper[4943]: I0307 14:41:33.880972 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/34a311b8-c902-40d4-ae68-4915984edfb7-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"34a311b8-c902-40d4-ae68-4915984edfb7\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 07 14:41:33 crc kubenswrapper[4943]: I0307 14:41:33.968293 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 07 14:41:34 crc kubenswrapper[4943]: I0307 14:41:34.038983 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 07 14:41:34 crc kubenswrapper[4943]: I0307 14:41:34.102528 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-587c5f97b-smswn" Mar 07 14:41:34 crc kubenswrapper[4943]: I0307 14:41:34.341780 4943 patch_prober.go:28] interesting pod/router-default-5444994796-nk422 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 07 14:41:34 crc kubenswrapper[4943]: [+]has-synced ok Mar 07 14:41:34 crc kubenswrapper[4943]: [+]process-running ok Mar 07 14:41:34 crc kubenswrapper[4943]: healthz check failed Mar 07 14:41:34 crc kubenswrapper[4943]: I0307 14:41:34.341878 4943 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-nk422" podUID="7ca6333f-d5ff-4c06-a140-3b9c29710528" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 07 14:41:34 crc kubenswrapper[4943]: I0307 14:41:34.497146 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-2rd7t" Mar 07 14:41:34 crc kubenswrapper[4943]: I0307 14:41:34.593208 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 07 14:41:34 crc kubenswrapper[4943]: I0307 14:41:34.707303 4943 ???:1] "http: TLS handshake error from 192.168.126.11:43346: no serving certificate available for the kubelet" Mar 07 14:41:34 crc kubenswrapper[4943]: I0307 14:41:34.730184 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"34a311b8-c902-40d4-ae68-4915984edfb7","Type":"ContainerStarted","Data":"5e4eff538aba65a49cf6ba53041c0541c5302861360187901cba7b7049763d8b"} Mar 07 14:41:35 crc kubenswrapper[4943]: I0307 14:41:35.339653 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-nk422" Mar 07 14:41:35 crc kubenswrapper[4943]: I0307 14:41:35.343584 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-nk422" Mar 07 14:41:35 crc kubenswrapper[4943]: I0307 14:41:35.773582 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"34a311b8-c902-40d4-ae68-4915984edfb7","Type":"ContainerStarted","Data":"4149a00fd15f5c679c1d0e4c8065f0e1d69ac1a47c00702aac2b7109c3cbe0a3"} Mar 07 14:41:35 crc kubenswrapper[4943]: I0307 14:41:35.797896 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.797872241 podStartE2EDuration="2.797872241s" podCreationTimestamp="2026-03-07 14:41:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:35.793440826 +0000 UTC m=+137.745577324" watchObservedRunningTime="2026-03-07 14:41:35.797872241 +0000 UTC m=+137.750008739" Mar 07 14:41:36 crc kubenswrapper[4943]: I0307 14:41:36.220806 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:41:36 crc kubenswrapper[4943]: I0307 14:41:36.801299 4943 generic.go:334] "Generic (PLEG): container finished" podID="34a311b8-c902-40d4-ae68-4915984edfb7" containerID="4149a00fd15f5c679c1d0e4c8065f0e1d69ac1a47c00702aac2b7109c3cbe0a3" exitCode=0 Mar 07 14:41:36 crc kubenswrapper[4943]: I0307 14:41:36.801752 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"34a311b8-c902-40d4-ae68-4915984edfb7","Type":"ContainerDied","Data":"4149a00fd15f5c679c1d0e4c8065f0e1d69ac1a47c00702aac2b7109c3cbe0a3"} Mar 07 14:41:38 crc kubenswrapper[4943]: I0307 14:41:38.903750 4943 ???:1] "http: TLS handshake error from 192.168.126.11:43356: no serving certificate available for the kubelet" Mar 07 14:41:41 crc kubenswrapper[4943]: E0307 14:41:41.904191 4943 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5ad27bce7efee616ce74cde00b2b89d44c20dc06fb06394eddde07f604433c89" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 07 14:41:41 crc kubenswrapper[4943]: E0307 14:41:41.906573 4943 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5ad27bce7efee616ce74cde00b2b89d44c20dc06fb06394eddde07f604433c89" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 07 14:41:41 crc kubenswrapper[4943]: E0307 14:41:41.909405 4943 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5ad27bce7efee616ce74cde00b2b89d44c20dc06fb06394eddde07f604433c89" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 07 14:41:41 crc kubenswrapper[4943]: E0307 14:41:41.909458 4943 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-htv6d" podUID="0de7b0c4-4acf-40de-867d-1fc33069fa51" containerName="kube-multus-additional-cni-plugins" Mar 07 14:41:43 crc kubenswrapper[4943]: I0307 14:41:43.187584 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-zgx7p" Mar 07 14:41:43 crc kubenswrapper[4943]: I0307 14:41:43.191513 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-zgx7p" Mar 07 14:41:43 crc kubenswrapper[4943]: I0307 14:41:43.496774 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-8jjq4" Mar 07 14:41:43 crc kubenswrapper[4943]: I0307 14:41:43.730970 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 07 14:41:43 crc kubenswrapper[4943]: I0307 14:41:43.833174 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/34a311b8-c902-40d4-ae68-4915984edfb7-kube-api-access\") pod \"34a311b8-c902-40d4-ae68-4915984edfb7\" (UID: \"34a311b8-c902-40d4-ae68-4915984edfb7\") " Mar 07 14:41:43 crc kubenswrapper[4943]: I0307 14:41:43.833310 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/34a311b8-c902-40d4-ae68-4915984edfb7-kubelet-dir\") pod \"34a311b8-c902-40d4-ae68-4915984edfb7\" (UID: \"34a311b8-c902-40d4-ae68-4915984edfb7\") " Mar 07 14:41:43 crc kubenswrapper[4943]: I0307 14:41:43.833742 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/34a311b8-c902-40d4-ae68-4915984edfb7-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "34a311b8-c902-40d4-ae68-4915984edfb7" (UID: "34a311b8-c902-40d4-ae68-4915984edfb7"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 07 14:41:43 crc kubenswrapper[4943]: I0307 14:41:43.838857 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34a311b8-c902-40d4-ae68-4915984edfb7-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "34a311b8-c902-40d4-ae68-4915984edfb7" (UID: "34a311b8-c902-40d4-ae68-4915984edfb7"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:41:43 crc kubenswrapper[4943]: I0307 14:41:43.888278 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"34a311b8-c902-40d4-ae68-4915984edfb7","Type":"ContainerDied","Data":"5e4eff538aba65a49cf6ba53041c0541c5302861360187901cba7b7049763d8b"} Mar 07 14:41:43 crc kubenswrapper[4943]: I0307 14:41:43.888332 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 07 14:41:43 crc kubenswrapper[4943]: I0307 14:41:43.888344 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5e4eff538aba65a49cf6ba53041c0541c5302861360187901cba7b7049763d8b" Mar 07 14:41:43 crc kubenswrapper[4943]: I0307 14:41:43.935614 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/34a311b8-c902-40d4-ae68-4915984edfb7-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 07 14:41:43 crc kubenswrapper[4943]: I0307 14:41:43.935661 4943 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/34a311b8-c902-40d4-ae68-4915984edfb7-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 07 14:41:47 crc kubenswrapper[4943]: I0307 14:41:47.619364 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 07 14:41:47 crc kubenswrapper[4943]: I0307 14:41:47.619898 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 07 14:41:47 crc kubenswrapper[4943]: I0307 14:41:47.620609 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 07 14:41:47 crc kubenswrapper[4943]: I0307 14:41:47.626618 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 07 14:41:47 crc kubenswrapper[4943]: I0307 14:41:47.706897 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 07 14:41:47 crc kubenswrapper[4943]: I0307 14:41:47.720967 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 07 14:41:47 crc kubenswrapper[4943]: I0307 14:41:47.721022 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 07 14:41:47 crc kubenswrapper[4943]: I0307 14:41:47.727112 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 07 14:41:47 crc kubenswrapper[4943]: I0307 14:41:47.728637 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 07 14:41:47 crc kubenswrapper[4943]: I0307 14:41:47.968033 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 07 14:41:47 crc kubenswrapper[4943]: I0307 14:41:47.975285 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 07 14:41:48 crc kubenswrapper[4943]: I0307 14:41:48.792708 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Mar 07 14:41:48 crc kubenswrapper[4943]: I0307 14:41:48.950074 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-8686b6766c-cqjdw"] Mar 07 14:41:48 crc kubenswrapper[4943]: I0307 14:41:48.951872 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-8686b6766c-cqjdw" podUID="3374c43b-61ed-4e63-87d0-2b212fc70a67" containerName="controller-manager" containerID="cri-o://382a1c9af8e00bcfdb9388ee197b918061c2b683a078650129340e5e1683a761" gracePeriod=30 Mar 07 14:41:48 crc kubenswrapper[4943]: I0307 14:41:48.964288 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-587c5f97b-smswn"] Mar 07 14:41:48 crc kubenswrapper[4943]: I0307 14:41:48.964823 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-587c5f97b-smswn" podUID="782f34ac-63b1-452d-9157-344361ae3723" containerName="route-controller-manager" containerID="cri-o://146679eb9aab8c30ac1f2d5598049bbd31f3f7404bca7d92a51c3daa13a1857f" gracePeriod=30 Mar 07 14:41:48 crc kubenswrapper[4943]: I0307 14:41:48.981722 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=0.981698627 podStartE2EDuration="981.698627ms" podCreationTimestamp="2026-03-07 14:41:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:48.981035748 +0000 UTC m=+150.933172256" watchObservedRunningTime="2026-03-07 14:41:48.981698627 +0000 UTC m=+150.933835125" Mar 07 14:41:49 crc kubenswrapper[4943]: I0307 14:41:49.509092 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:41:49 crc kubenswrapper[4943]: I0307 14:41:49.928886 4943 generic.go:334] "Generic (PLEG): container finished" podID="3374c43b-61ed-4e63-87d0-2b212fc70a67" containerID="382a1c9af8e00bcfdb9388ee197b918061c2b683a078650129340e5e1683a761" exitCode=0 Mar 07 14:41:49 crc kubenswrapper[4943]: I0307 14:41:49.929022 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-8686b6766c-cqjdw" event={"ID":"3374c43b-61ed-4e63-87d0-2b212fc70a67","Type":"ContainerDied","Data":"382a1c9af8e00bcfdb9388ee197b918061c2b683a078650129340e5e1683a761"} Mar 07 14:41:49 crc kubenswrapper[4943]: I0307 14:41:49.932456 4943 generic.go:334] "Generic (PLEG): container finished" podID="782f34ac-63b1-452d-9157-344361ae3723" containerID="146679eb9aab8c30ac1f2d5598049bbd31f3f7404bca7d92a51c3daa13a1857f" exitCode=0 Mar 07 14:41:49 crc kubenswrapper[4943]: I0307 14:41:49.932584 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-587c5f97b-smswn" event={"ID":"782f34ac-63b1-452d-9157-344361ae3723","Type":"ContainerDied","Data":"146679eb9aab8c30ac1f2d5598049bbd31f3f7404bca7d92a51c3daa13a1857f"} Mar 07 14:41:51 crc kubenswrapper[4943]: E0307 14:41:51.904307 4943 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5ad27bce7efee616ce74cde00b2b89d44c20dc06fb06394eddde07f604433c89" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 07 14:41:51 crc kubenswrapper[4943]: E0307 14:41:51.906671 4943 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5ad27bce7efee616ce74cde00b2b89d44c20dc06fb06394eddde07f604433c89" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 07 14:41:51 crc kubenswrapper[4943]: E0307 14:41:51.908977 4943 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5ad27bce7efee616ce74cde00b2b89d44c20dc06fb06394eddde07f604433c89" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 07 14:41:51 crc kubenswrapper[4943]: E0307 14:41:51.909102 4943 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-htv6d" podUID="0de7b0c4-4acf-40de-867d-1fc33069fa51" containerName="kube-multus-additional-cni-plugins" Mar 07 14:41:52 crc kubenswrapper[4943]: I0307 14:41:52.088369 4943 patch_prober.go:28] interesting pod/controller-manager-8686b6766c-cqjdw container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.53:8443/healthz\": dial tcp 10.217.0.53:8443: connect: connection refused" start-of-body= Mar 07 14:41:52 crc kubenswrapper[4943]: I0307 14:41:52.088476 4943 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-8686b6766c-cqjdw" podUID="3374c43b-61ed-4e63-87d0-2b212fc70a67" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.53:8443/healthz\": dial tcp 10.217.0.53:8443: connect: connection refused" Mar 07 14:41:52 crc kubenswrapper[4943]: I0307 14:41:52.113206 4943 patch_prober.go:28] interesting pod/route-controller-manager-587c5f97b-smswn container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.54:8443/healthz\": dial tcp 10.217.0.54:8443: connect: connection refused" start-of-body= Mar 07 14:41:52 crc kubenswrapper[4943]: I0307 14:41:52.113307 4943 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-587c5f97b-smswn" podUID="782f34ac-63b1-452d-9157-344361ae3723" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.54:8443/healthz\": dial tcp 10.217.0.54:8443: connect: connection refused" Mar 07 14:41:53 crc kubenswrapper[4943]: E0307 14:41:53.558906 4943 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Mar 07 14:41:53 crc kubenswrapper[4943]: E0307 14:41:53.559312 4943 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-f8qh5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-9sf6k_openshift-marketplace(23fa83e7-df77-47e2-bb98-3344e48ab6bd): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 07 14:41:53 crc kubenswrapper[4943]: E0307 14:41:53.560625 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-9sf6k" podUID="23fa83e7-df77-47e2-bb98-3344e48ab6bd" Mar 07 14:41:57 crc kubenswrapper[4943]: E0307 14:41:57.601527 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-9sf6k" podUID="23fa83e7-df77-47e2-bb98-3344e48ab6bd" Mar 07 14:41:57 crc kubenswrapper[4943]: E0307 14:41:57.689017 4943 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Mar 07 14:41:57 crc kubenswrapper[4943]: E0307 14:41:57.689204 4943 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wthgk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-jrc6h_openshift-marketplace(f10fa8f5-f504-40c9-81a5-d1658c1ef268): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 07 14:41:57 crc kubenswrapper[4943]: E0307 14:41:57.690455 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-jrc6h" podUID="f10fa8f5-f504-40c9-81a5-d1658c1ef268" Mar 07 14:41:57 crc kubenswrapper[4943]: E0307 14:41:57.714998 4943 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Mar 07 14:41:57 crc kubenswrapper[4943]: E0307 14:41:57.715385 4943 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8p98b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-rvkgz_openshift-marketplace(4ae1cf01-1400-4500-9b84-8435a6541612): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 07 14:41:57 crc kubenswrapper[4943]: E0307 14:41:57.716600 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-rvkgz" podUID="4ae1cf01-1400-4500-9b84-8435a6541612" Mar 07 14:41:57 crc kubenswrapper[4943]: E0307 14:41:57.728960 4943 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Mar 07 14:41:57 crc kubenswrapper[4943]: E0307 14:41:57.729195 4943 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7mt9r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-f95bw_openshift-marketplace(a41f89c1-5f6d-46bc-801b-a29ebebf4468): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 07 14:41:57 crc kubenswrapper[4943]: E0307 14:41:57.730837 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-f95bw" podUID="a41f89c1-5f6d-46bc-801b-a29ebebf4468" Mar 07 14:41:57 crc kubenswrapper[4943]: I0307 14:41:57.769783 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Mar 07 14:41:57 crc kubenswrapper[4943]: I0307 14:41:57.989528 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_cni-sysctl-allowlist-ds-htv6d_0de7b0c4-4acf-40de-867d-1fc33069fa51/kube-multus-additional-cni-plugins/0.log" Mar 07 14:41:57 crc kubenswrapper[4943]: I0307 14:41:57.989963 4943 generic.go:334] "Generic (PLEG): container finished" podID="0de7b0c4-4acf-40de-867d-1fc33069fa51" containerID="5ad27bce7efee616ce74cde00b2b89d44c20dc06fb06394eddde07f604433c89" exitCode=137 Mar 07 14:41:57 crc kubenswrapper[4943]: I0307 14:41:57.990142 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-htv6d" event={"ID":"0de7b0c4-4acf-40de-867d-1fc33069fa51","Type":"ContainerDied","Data":"5ad27bce7efee616ce74cde00b2b89d44c20dc06fb06394eddde07f604433c89"} Mar 07 14:41:58 crc kubenswrapper[4943]: I0307 14:41:58.009506 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=1.00947745 podStartE2EDuration="1.00947745s" podCreationTimestamp="2026-03-07 14:41:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:41:58.005776531 +0000 UTC m=+159.957913029" watchObservedRunningTime="2026-03-07 14:41:58.00947745 +0000 UTC m=+159.961613968" Mar 07 14:41:59 crc kubenswrapper[4943]: I0307 14:41:59.404756 4943 ???:1] "http: TLS handshake error from 192.168.126.11:36898: no serving certificate available for the kubelet" Mar 07 14:41:59 crc kubenswrapper[4943]: E0307 14:41:59.571857 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-rvkgz" podUID="4ae1cf01-1400-4500-9b84-8435a6541612" Mar 07 14:41:59 crc kubenswrapper[4943]: E0307 14:41:59.571978 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-jrc6h" podUID="f10fa8f5-f504-40c9-81a5-d1658c1ef268" Mar 07 14:41:59 crc kubenswrapper[4943]: E0307 14:41:59.572078 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-f95bw" podUID="a41f89c1-5f6d-46bc-801b-a29ebebf4468" Mar 07 14:41:59 crc kubenswrapper[4943]: I0307 14:41:59.688469 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_cni-sysctl-allowlist-ds-htv6d_0de7b0c4-4acf-40de-867d-1fc33069fa51/kube-multus-additional-cni-plugins/0.log" Mar 07 14:41:59 crc kubenswrapper[4943]: I0307 14:41:59.688821 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-htv6d" Mar 07 14:41:59 crc kubenswrapper[4943]: E0307 14:41:59.694269 4943 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Mar 07 14:41:59 crc kubenswrapper[4943]: E0307 14:41:59.694463 4943 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Mar 07 14:41:59 crc kubenswrapper[4943]: E0307 14:41:59.694537 4943 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rf8ld,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-p767n_openshift-marketplace(6a05c7ae-f2a1-4116-9038-ac838fa88af7): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 07 14:41:59 crc kubenswrapper[4943]: E0307 14:41:59.694715 4943 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mq6g7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-vkcxl_openshift-marketplace(74a7b99c-3412-438e-a835-1e153a56047b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 07 14:41:59 crc kubenswrapper[4943]: E0307 14:41:59.695778 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-p767n" podUID="6a05c7ae-f2a1-4116-9038-ac838fa88af7" Mar 07 14:41:59 crc kubenswrapper[4943]: E0307 14:41:59.697988 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-vkcxl" podUID="74a7b99c-3412-438e-a835-1e153a56047b" Mar 07 14:41:59 crc kubenswrapper[4943]: E0307 14:41:59.719746 4943 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Mar 07 14:41:59 crc kubenswrapper[4943]: E0307 14:41:59.720057 4943 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-582l8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-dfqzr_openshift-marketplace(0d1657d7-a787-4648-91f6-e0ff987cede2): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 07 14:41:59 crc kubenswrapper[4943]: E0307 14:41:59.721394 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-dfqzr" podUID="0d1657d7-a787-4648-91f6-e0ff987cede2" Mar 07 14:41:59 crc kubenswrapper[4943]: I0307 14:41:59.802077 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dgzhw\" (UniqueName: \"kubernetes.io/projected/0de7b0c4-4acf-40de-867d-1fc33069fa51-kube-api-access-dgzhw\") pod \"0de7b0c4-4acf-40de-867d-1fc33069fa51\" (UID: \"0de7b0c4-4acf-40de-867d-1fc33069fa51\") " Mar 07 14:41:59 crc kubenswrapper[4943]: I0307 14:41:59.802601 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/0de7b0c4-4acf-40de-867d-1fc33069fa51-ready\") pod \"0de7b0c4-4acf-40de-867d-1fc33069fa51\" (UID: \"0de7b0c4-4acf-40de-867d-1fc33069fa51\") " Mar 07 14:41:59 crc kubenswrapper[4943]: I0307 14:41:59.802643 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/0de7b0c4-4acf-40de-867d-1fc33069fa51-tuning-conf-dir\") pod \"0de7b0c4-4acf-40de-867d-1fc33069fa51\" (UID: \"0de7b0c4-4acf-40de-867d-1fc33069fa51\") " Mar 07 14:41:59 crc kubenswrapper[4943]: I0307 14:41:59.802710 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/0de7b0c4-4acf-40de-867d-1fc33069fa51-cni-sysctl-allowlist\") pod \"0de7b0c4-4acf-40de-867d-1fc33069fa51\" (UID: \"0de7b0c4-4acf-40de-867d-1fc33069fa51\") " Mar 07 14:41:59 crc kubenswrapper[4943]: I0307 14:41:59.803076 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0de7b0c4-4acf-40de-867d-1fc33069fa51-ready" (OuterVolumeSpecName: "ready") pod "0de7b0c4-4acf-40de-867d-1fc33069fa51" (UID: "0de7b0c4-4acf-40de-867d-1fc33069fa51"). InnerVolumeSpecName "ready". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:41:59 crc kubenswrapper[4943]: I0307 14:41:59.803691 4943 reconciler_common.go:293] "Volume detached for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/0de7b0c4-4acf-40de-867d-1fc33069fa51-ready\") on node \"crc\" DevicePath \"\"" Mar 07 14:41:59 crc kubenswrapper[4943]: I0307 14:41:59.803694 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0de7b0c4-4acf-40de-867d-1fc33069fa51-tuning-conf-dir" (OuterVolumeSpecName: "tuning-conf-dir") pod "0de7b0c4-4acf-40de-867d-1fc33069fa51" (UID: "0de7b0c4-4acf-40de-867d-1fc33069fa51"). InnerVolumeSpecName "tuning-conf-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 07 14:41:59 crc kubenswrapper[4943]: I0307 14:41:59.803785 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0de7b0c4-4acf-40de-867d-1fc33069fa51-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "0de7b0c4-4acf-40de-867d-1fc33069fa51" (UID: "0de7b0c4-4acf-40de-867d-1fc33069fa51"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:41:59 crc kubenswrapper[4943]: I0307 14:41:59.812233 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0de7b0c4-4acf-40de-867d-1fc33069fa51-kube-api-access-dgzhw" (OuterVolumeSpecName: "kube-api-access-dgzhw") pod "0de7b0c4-4acf-40de-867d-1fc33069fa51" (UID: "0de7b0c4-4acf-40de-867d-1fc33069fa51"). InnerVolumeSpecName "kube-api-access-dgzhw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:41:59 crc kubenswrapper[4943]: I0307 14:41:59.904953 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dgzhw\" (UniqueName: \"kubernetes.io/projected/0de7b0c4-4acf-40de-867d-1fc33069fa51-kube-api-access-dgzhw\") on node \"crc\" DevicePath \"\"" Mar 07 14:41:59 crc kubenswrapper[4943]: I0307 14:41:59.904997 4943 reconciler_common.go:293] "Volume detached for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/0de7b0c4-4acf-40de-867d-1fc33069fa51-tuning-conf-dir\") on node \"crc\" DevicePath \"\"" Mar 07 14:41:59 crc kubenswrapper[4943]: I0307 14:41:59.905007 4943 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/0de7b0c4-4acf-40de-867d-1fc33069fa51-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.002370 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_cni-sysctl-allowlist-ds-htv6d_0de7b0c4-4acf-40de-867d-1fc33069fa51/kube-multus-additional-cni-plugins/0.log" Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.002434 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-htv6d" event={"ID":"0de7b0c4-4acf-40de-867d-1fc33069fa51","Type":"ContainerDied","Data":"a503af39597f4a531b75289a655cb79ea7f4c380985b8d0744618afd83475c86"} Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.002475 4943 scope.go:117] "RemoveContainer" containerID="5ad27bce7efee616ce74cde00b2b89d44c20dc06fb06394eddde07f604433c89" Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.002571 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-htv6d" Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.010041 4943 generic.go:334] "Generic (PLEG): container finished" podID="254f0c0f-ee54-434e-bf7b-bf2e5274fdc9" containerID="0163d96eb2ce8fb71907d1a0fc94f1275f85e77a54c20cb56ce46c9882453aec" exitCode=0 Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.010513 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xwdpr" event={"ID":"254f0c0f-ee54-434e-bf7b-bf2e5274fdc9","Type":"ContainerDied","Data":"0163d96eb2ce8fb71907d1a0fc94f1275f85e77a54c20cb56ce46c9882453aec"} Mar 07 14:42:00 crc kubenswrapper[4943]: E0307 14:42:00.014049 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-vkcxl" podUID="74a7b99c-3412-438e-a835-1e153a56047b" Mar 07 14:42:00 crc kubenswrapper[4943]: E0307 14:42:00.023223 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-p767n" podUID="6a05c7ae-f2a1-4116-9038-ac838fa88af7" Mar 07 14:42:00 crc kubenswrapper[4943]: E0307 14:42:00.023618 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-dfqzr" podUID="0d1657d7-a787-4648-91f6-e0ff987cede2" Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.068169 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-587c5f97b-smswn" Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.090293 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-8686b6766c-cqjdw" Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.116320 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-htv6d"] Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.119628 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-htv6d"] Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.132144 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29548242-scpk7"] Mar 07 14:42:00 crc kubenswrapper[4943]: E0307 14:42:00.135773 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="782f34ac-63b1-452d-9157-344361ae3723" containerName="route-controller-manager" Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.135901 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="782f34ac-63b1-452d-9157-344361ae3723" containerName="route-controller-manager" Mar 07 14:42:00 crc kubenswrapper[4943]: E0307 14:42:00.136424 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3374c43b-61ed-4e63-87d0-2b212fc70a67" containerName="controller-manager" Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.136658 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="3374c43b-61ed-4e63-87d0-2b212fc70a67" containerName="controller-manager" Mar 07 14:42:00 crc kubenswrapper[4943]: E0307 14:42:00.136812 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0de7b0c4-4acf-40de-867d-1fc33069fa51" containerName="kube-multus-additional-cni-plugins" Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.136894 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="0de7b0c4-4acf-40de-867d-1fc33069fa51" containerName="kube-multus-additional-cni-plugins" Mar 07 14:42:00 crc kubenswrapper[4943]: E0307 14:42:00.136991 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34a311b8-c902-40d4-ae68-4915984edfb7" containerName="pruner" Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.137816 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="34a311b8-c902-40d4-ae68-4915984edfb7" containerName="pruner" Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.138053 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="0de7b0c4-4acf-40de-867d-1fc33069fa51" containerName="kube-multus-additional-cni-plugins" Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.138140 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="782f34ac-63b1-452d-9157-344361ae3723" containerName="route-controller-manager" Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.138223 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="3374c43b-61ed-4e63-87d0-2b212fc70a67" containerName="controller-manager" Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.138309 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="34a311b8-c902-40d4-ae68-4915984edfb7" containerName="pruner" Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.141010 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548242-scpk7" Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.144737 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.145066 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.145284 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-qqlk5" Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.146468 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29548242-scpk7"] Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.207741 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8bcnl\" (UniqueName: \"kubernetes.io/projected/3374c43b-61ed-4e63-87d0-2b212fc70a67-kube-api-access-8bcnl\") pod \"3374c43b-61ed-4e63-87d0-2b212fc70a67\" (UID: \"3374c43b-61ed-4e63-87d0-2b212fc70a67\") " Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.207799 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3374c43b-61ed-4e63-87d0-2b212fc70a67-config\") pod \"3374c43b-61ed-4e63-87d0-2b212fc70a67\" (UID: \"3374c43b-61ed-4e63-87d0-2b212fc70a67\") " Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.207821 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kgmwj\" (UniqueName: \"kubernetes.io/projected/782f34ac-63b1-452d-9157-344361ae3723-kube-api-access-kgmwj\") pod \"782f34ac-63b1-452d-9157-344361ae3723\" (UID: \"782f34ac-63b1-452d-9157-344361ae3723\") " Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.207862 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/782f34ac-63b1-452d-9157-344361ae3723-serving-cert\") pod \"782f34ac-63b1-452d-9157-344361ae3723\" (UID: \"782f34ac-63b1-452d-9157-344361ae3723\") " Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.208007 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3374c43b-61ed-4e63-87d0-2b212fc70a67-serving-cert\") pod \"3374c43b-61ed-4e63-87d0-2b212fc70a67\" (UID: \"3374c43b-61ed-4e63-87d0-2b212fc70a67\") " Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.208030 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3374c43b-61ed-4e63-87d0-2b212fc70a67-proxy-ca-bundles\") pod \"3374c43b-61ed-4e63-87d0-2b212fc70a67\" (UID: \"3374c43b-61ed-4e63-87d0-2b212fc70a67\") " Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.208047 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/782f34ac-63b1-452d-9157-344361ae3723-client-ca\") pod \"782f34ac-63b1-452d-9157-344361ae3723\" (UID: \"782f34ac-63b1-452d-9157-344361ae3723\") " Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.208071 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/782f34ac-63b1-452d-9157-344361ae3723-config\") pod \"782f34ac-63b1-452d-9157-344361ae3723\" (UID: \"782f34ac-63b1-452d-9157-344361ae3723\") " Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.208092 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3374c43b-61ed-4e63-87d0-2b212fc70a67-client-ca\") pod \"3374c43b-61ed-4e63-87d0-2b212fc70a67\" (UID: \"3374c43b-61ed-4e63-87d0-2b212fc70a67\") " Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.208287 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6c2r\" (UniqueName: \"kubernetes.io/projected/14d07994-bd34-4802-8dd5-8fd8afe5e878-kube-api-access-n6c2r\") pod \"auto-csr-approver-29548242-scpk7\" (UID: \"14d07994-bd34-4802-8dd5-8fd8afe5e878\") " pod="openshift-infra/auto-csr-approver-29548242-scpk7" Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.211638 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/782f34ac-63b1-452d-9157-344361ae3723-client-ca" (OuterVolumeSpecName: "client-ca") pod "782f34ac-63b1-452d-9157-344361ae3723" (UID: "782f34ac-63b1-452d-9157-344361ae3723"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.212355 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/782f34ac-63b1-452d-9157-344361ae3723-config" (OuterVolumeSpecName: "config") pod "782f34ac-63b1-452d-9157-344361ae3723" (UID: "782f34ac-63b1-452d-9157-344361ae3723"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.213529 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3374c43b-61ed-4e63-87d0-2b212fc70a67-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "3374c43b-61ed-4e63-87d0-2b212fc70a67" (UID: "3374c43b-61ed-4e63-87d0-2b212fc70a67"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.213740 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3374c43b-61ed-4e63-87d0-2b212fc70a67-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "3374c43b-61ed-4e63-87d0-2b212fc70a67" (UID: "3374c43b-61ed-4e63-87d0-2b212fc70a67"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.213821 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3374c43b-61ed-4e63-87d0-2b212fc70a67-client-ca" (OuterVolumeSpecName: "client-ca") pod "3374c43b-61ed-4e63-87d0-2b212fc70a67" (UID: "3374c43b-61ed-4e63-87d0-2b212fc70a67"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.214399 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3374c43b-61ed-4e63-87d0-2b212fc70a67-config" (OuterVolumeSpecName: "config") pod "3374c43b-61ed-4e63-87d0-2b212fc70a67" (UID: "3374c43b-61ed-4e63-87d0-2b212fc70a67"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.214545 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3374c43b-61ed-4e63-87d0-2b212fc70a67-kube-api-access-8bcnl" (OuterVolumeSpecName: "kube-api-access-8bcnl") pod "3374c43b-61ed-4e63-87d0-2b212fc70a67" (UID: "3374c43b-61ed-4e63-87d0-2b212fc70a67"). InnerVolumeSpecName "kube-api-access-8bcnl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.219276 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/782f34ac-63b1-452d-9157-344361ae3723-kube-api-access-kgmwj" (OuterVolumeSpecName: "kube-api-access-kgmwj") pod "782f34ac-63b1-452d-9157-344361ae3723" (UID: "782f34ac-63b1-452d-9157-344361ae3723"). InnerVolumeSpecName "kube-api-access-kgmwj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.231168 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/782f34ac-63b1-452d-9157-344361ae3723-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "782f34ac-63b1-452d-9157-344361ae3723" (UID: "782f34ac-63b1-452d-9157-344361ae3723"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.309238 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6c2r\" (UniqueName: \"kubernetes.io/projected/14d07994-bd34-4802-8dd5-8fd8afe5e878-kube-api-access-n6c2r\") pod \"auto-csr-approver-29548242-scpk7\" (UID: \"14d07994-bd34-4802-8dd5-8fd8afe5e878\") " pod="openshift-infra/auto-csr-approver-29548242-scpk7" Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.309376 4943 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3374c43b-61ed-4e63-87d0-2b212fc70a67-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.309394 4943 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/782f34ac-63b1-452d-9157-344361ae3723-config\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.309407 4943 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3374c43b-61ed-4e63-87d0-2b212fc70a67-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.309423 4943 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/782f34ac-63b1-452d-9157-344361ae3723-client-ca\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.309436 4943 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3374c43b-61ed-4e63-87d0-2b212fc70a67-client-ca\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.309448 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8bcnl\" (UniqueName: \"kubernetes.io/projected/3374c43b-61ed-4e63-87d0-2b212fc70a67-kube-api-access-8bcnl\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.309460 4943 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3374c43b-61ed-4e63-87d0-2b212fc70a67-config\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.309471 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kgmwj\" (UniqueName: \"kubernetes.io/projected/782f34ac-63b1-452d-9157-344361ae3723-kube-api-access-kgmwj\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.309483 4943 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/782f34ac-63b1-452d-9157-344361ae3723-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.330690 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6c2r\" (UniqueName: \"kubernetes.io/projected/14d07994-bd34-4802-8dd5-8fd8afe5e878-kube-api-access-n6c2r\") pod \"auto-csr-approver-29548242-scpk7\" (UID: \"14d07994-bd34-4802-8dd5-8fd8afe5e878\") " pod="openshift-infra/auto-csr-approver-29548242-scpk7" Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.465027 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548242-scpk7" Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.764791 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0de7b0c4-4acf-40de-867d-1fc33069fa51" path="/var/lib/kubelet/pods/0de7b0c4-4acf-40de-867d-1fc33069fa51/volumes" Mar 07 14:42:00 crc kubenswrapper[4943]: I0307 14:42:00.923162 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29548242-scpk7"] Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.017048 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"f86de7fe0648447fdbec2675afb9e12b56bd9e57ae0670530b7686e0e0b626bf"} Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.017142 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"58e23186c2a2929d327fd8b689cc71cf8952146b2280e73a986ed56c1ba58bf7"} Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.021095 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-587c5f97b-smswn" Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.021737 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-587c5f97b-smswn" event={"ID":"782f34ac-63b1-452d-9157-344361ae3723","Type":"ContainerDied","Data":"20872e7e7beca543cbd741c4b7eb0b6515ae4e35ac1d0d3bc127a7fc283d8205"} Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.021778 4943 scope.go:117] "RemoveContainer" containerID="146679eb9aab8c30ac1f2d5598049bbd31f3f7404bca7d92a51c3daa13a1857f" Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.025904 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xwdpr" event={"ID":"254f0c0f-ee54-434e-bf7b-bf2e5274fdc9","Type":"ContainerStarted","Data":"786d27c9685fe23593de377ea7485e2d68bf44446fa5662aad6df02c8625e99b"} Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.027675 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-8686b6766c-cqjdw" Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.027683 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-8686b6766c-cqjdw" event={"ID":"3374c43b-61ed-4e63-87d0-2b212fc70a67","Type":"ContainerDied","Data":"15fbdfa9b89b26d60c2c03c2b5c071605fc8464c267769de33d9bc88d7286d44"} Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.033959 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"2a406d2d9cc20313ff14fd91ca22a3c0ab849f8e881f946d92c4d37c70779606"} Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.034139 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"4c88a38cfc38ecafcdc4b51b452bef69b9e0ba1f60a62657104a472569e9b4e4"} Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.037055 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548242-scpk7" event={"ID":"14d07994-bd34-4802-8dd5-8fd8afe5e878","Type":"ContainerStarted","Data":"2b619c7fa4f294863fb14b75a2ddfe9cb31a5bb989803badfb17c0ea218fed2d"} Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.042204 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"ada321029ccb617558878e95a6e003d85b6705324e40847c6c41df938a400a00"} Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.042336 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"7eda824ba10216e6978c1e27fc7e1d4dfdac0e0a88ae5661913572db295a0854"} Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.042575 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.053230 4943 scope.go:117] "RemoveContainer" containerID="382a1c9af8e00bcfdb9388ee197b918061c2b683a078650129340e5e1683a761" Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.080442 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xwdpr" podStartSLOduration=3.986905384 podStartE2EDuration="34.080422145s" podCreationTimestamp="2026-03-07 14:41:27 +0000 UTC" firstStartedPulling="2026-03-07 14:41:30.404708517 +0000 UTC m=+132.356845015" lastFinishedPulling="2026-03-07 14:42:00.498225228 +0000 UTC m=+162.450361776" observedRunningTime="2026-03-07 14:42:01.067961594 +0000 UTC m=+163.020098112" watchObservedRunningTime="2026-03-07 14:42:01.080422145 +0000 UTC m=+163.032558643" Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.082364 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-587c5f97b-smswn"] Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.086336 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-587c5f97b-smswn"] Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.144298 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-8686b6766c-cqjdw"] Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.144343 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-8686b6766c-cqjdw"] Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.778361 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5764d979d6-l2tqn"] Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.780140 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5764d979d6-l2tqn" Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.781761 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-69ff69cc58-vmx82"] Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.782720 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.782805 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.782997 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-69ff69cc58-vmx82" Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.784702 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.784724 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.785294 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.785529 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.785542 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.785708 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.785868 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.786042 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.786198 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.786203 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.803321 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-69ff69cc58-vmx82"] Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.806191 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5764d979d6-l2tqn"] Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.832713 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d41ebe43-4352-4732-bcc2-0ba95c3db967-serving-cert\") pod \"route-controller-manager-69ff69cc58-vmx82\" (UID: \"d41ebe43-4352-4732-bcc2-0ba95c3db967\") " pod="openshift-route-controller-manager/route-controller-manager-69ff69cc58-vmx82" Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.832760 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/efb24914-9e13-4aed-bfd6-f2f6291a2750-serving-cert\") pod \"controller-manager-5764d979d6-l2tqn\" (UID: \"efb24914-9e13-4aed-bfd6-f2f6291a2750\") " pod="openshift-controller-manager/controller-manager-5764d979d6-l2tqn" Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.832784 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nc7f9\" (UniqueName: \"kubernetes.io/projected/efb24914-9e13-4aed-bfd6-f2f6291a2750-kube-api-access-nc7f9\") pod \"controller-manager-5764d979d6-l2tqn\" (UID: \"efb24914-9e13-4aed-bfd6-f2f6291a2750\") " pod="openshift-controller-manager/controller-manager-5764d979d6-l2tqn" Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.832805 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d41ebe43-4352-4732-bcc2-0ba95c3db967-config\") pod \"route-controller-manager-69ff69cc58-vmx82\" (UID: \"d41ebe43-4352-4732-bcc2-0ba95c3db967\") " pod="openshift-route-controller-manager/route-controller-manager-69ff69cc58-vmx82" Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.832836 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/efb24914-9e13-4aed-bfd6-f2f6291a2750-config\") pod \"controller-manager-5764d979d6-l2tqn\" (UID: \"efb24914-9e13-4aed-bfd6-f2f6291a2750\") " pod="openshift-controller-manager/controller-manager-5764d979d6-l2tqn" Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.832859 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drp64\" (UniqueName: \"kubernetes.io/projected/d41ebe43-4352-4732-bcc2-0ba95c3db967-kube-api-access-drp64\") pod \"route-controller-manager-69ff69cc58-vmx82\" (UID: \"d41ebe43-4352-4732-bcc2-0ba95c3db967\") " pod="openshift-route-controller-manager/route-controller-manager-69ff69cc58-vmx82" Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.832887 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/efb24914-9e13-4aed-bfd6-f2f6291a2750-client-ca\") pod \"controller-manager-5764d979d6-l2tqn\" (UID: \"efb24914-9e13-4aed-bfd6-f2f6291a2750\") " pod="openshift-controller-manager/controller-manager-5764d979d6-l2tqn" Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.832910 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/efb24914-9e13-4aed-bfd6-f2f6291a2750-proxy-ca-bundles\") pod \"controller-manager-5764d979d6-l2tqn\" (UID: \"efb24914-9e13-4aed-bfd6-f2f6291a2750\") " pod="openshift-controller-manager/controller-manager-5764d979d6-l2tqn" Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.832945 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d41ebe43-4352-4732-bcc2-0ba95c3db967-client-ca\") pod \"route-controller-manager-69ff69cc58-vmx82\" (UID: \"d41ebe43-4352-4732-bcc2-0ba95c3db967\") " pod="openshift-route-controller-manager/route-controller-manager-69ff69cc58-vmx82" Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.845247 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.933851 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d41ebe43-4352-4732-bcc2-0ba95c3db967-serving-cert\") pod \"route-controller-manager-69ff69cc58-vmx82\" (UID: \"d41ebe43-4352-4732-bcc2-0ba95c3db967\") " pod="openshift-route-controller-manager/route-controller-manager-69ff69cc58-vmx82" Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.933901 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/efb24914-9e13-4aed-bfd6-f2f6291a2750-serving-cert\") pod \"controller-manager-5764d979d6-l2tqn\" (UID: \"efb24914-9e13-4aed-bfd6-f2f6291a2750\") " pod="openshift-controller-manager/controller-manager-5764d979d6-l2tqn" Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.933937 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nc7f9\" (UniqueName: \"kubernetes.io/projected/efb24914-9e13-4aed-bfd6-f2f6291a2750-kube-api-access-nc7f9\") pod \"controller-manager-5764d979d6-l2tqn\" (UID: \"efb24914-9e13-4aed-bfd6-f2f6291a2750\") " pod="openshift-controller-manager/controller-manager-5764d979d6-l2tqn" Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.933959 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d41ebe43-4352-4732-bcc2-0ba95c3db967-config\") pod \"route-controller-manager-69ff69cc58-vmx82\" (UID: \"d41ebe43-4352-4732-bcc2-0ba95c3db967\") " pod="openshift-route-controller-manager/route-controller-manager-69ff69cc58-vmx82" Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.933992 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/efb24914-9e13-4aed-bfd6-f2f6291a2750-config\") pod \"controller-manager-5764d979d6-l2tqn\" (UID: \"efb24914-9e13-4aed-bfd6-f2f6291a2750\") " pod="openshift-controller-manager/controller-manager-5764d979d6-l2tqn" Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.934016 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drp64\" (UniqueName: \"kubernetes.io/projected/d41ebe43-4352-4732-bcc2-0ba95c3db967-kube-api-access-drp64\") pod \"route-controller-manager-69ff69cc58-vmx82\" (UID: \"d41ebe43-4352-4732-bcc2-0ba95c3db967\") " pod="openshift-route-controller-manager/route-controller-manager-69ff69cc58-vmx82" Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.934052 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/efb24914-9e13-4aed-bfd6-f2f6291a2750-client-ca\") pod \"controller-manager-5764d979d6-l2tqn\" (UID: \"efb24914-9e13-4aed-bfd6-f2f6291a2750\") " pod="openshift-controller-manager/controller-manager-5764d979d6-l2tqn" Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.934074 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/efb24914-9e13-4aed-bfd6-f2f6291a2750-proxy-ca-bundles\") pod \"controller-manager-5764d979d6-l2tqn\" (UID: \"efb24914-9e13-4aed-bfd6-f2f6291a2750\") " pod="openshift-controller-manager/controller-manager-5764d979d6-l2tqn" Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.934097 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d41ebe43-4352-4732-bcc2-0ba95c3db967-client-ca\") pod \"route-controller-manager-69ff69cc58-vmx82\" (UID: \"d41ebe43-4352-4732-bcc2-0ba95c3db967\") " pod="openshift-route-controller-manager/route-controller-manager-69ff69cc58-vmx82" Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.935465 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d41ebe43-4352-4732-bcc2-0ba95c3db967-client-ca\") pod \"route-controller-manager-69ff69cc58-vmx82\" (UID: \"d41ebe43-4352-4732-bcc2-0ba95c3db967\") " pod="openshift-route-controller-manager/route-controller-manager-69ff69cc58-vmx82" Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.935805 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/efb24914-9e13-4aed-bfd6-f2f6291a2750-client-ca\") pod \"controller-manager-5764d979d6-l2tqn\" (UID: \"efb24914-9e13-4aed-bfd6-f2f6291a2750\") " pod="openshift-controller-manager/controller-manager-5764d979d6-l2tqn" Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.935864 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d41ebe43-4352-4732-bcc2-0ba95c3db967-config\") pod \"route-controller-manager-69ff69cc58-vmx82\" (UID: \"d41ebe43-4352-4732-bcc2-0ba95c3db967\") " pod="openshift-route-controller-manager/route-controller-manager-69ff69cc58-vmx82" Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.935993 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/efb24914-9e13-4aed-bfd6-f2f6291a2750-proxy-ca-bundles\") pod \"controller-manager-5764d979d6-l2tqn\" (UID: \"efb24914-9e13-4aed-bfd6-f2f6291a2750\") " pod="openshift-controller-manager/controller-manager-5764d979d6-l2tqn" Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.936323 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/efb24914-9e13-4aed-bfd6-f2f6291a2750-config\") pod \"controller-manager-5764d979d6-l2tqn\" (UID: \"efb24914-9e13-4aed-bfd6-f2f6291a2750\") " pod="openshift-controller-manager/controller-manager-5764d979d6-l2tqn" Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.947852 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d41ebe43-4352-4732-bcc2-0ba95c3db967-serving-cert\") pod \"route-controller-manager-69ff69cc58-vmx82\" (UID: \"d41ebe43-4352-4732-bcc2-0ba95c3db967\") " pod="openshift-route-controller-manager/route-controller-manager-69ff69cc58-vmx82" Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.949086 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/efb24914-9e13-4aed-bfd6-f2f6291a2750-serving-cert\") pod \"controller-manager-5764d979d6-l2tqn\" (UID: \"efb24914-9e13-4aed-bfd6-f2f6291a2750\") " pod="openshift-controller-manager/controller-manager-5764d979d6-l2tqn" Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.955076 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nc7f9\" (UniqueName: \"kubernetes.io/projected/efb24914-9e13-4aed-bfd6-f2f6291a2750-kube-api-access-nc7f9\") pod \"controller-manager-5764d979d6-l2tqn\" (UID: \"efb24914-9e13-4aed-bfd6-f2f6291a2750\") " pod="openshift-controller-manager/controller-manager-5764d979d6-l2tqn" Mar 07 14:42:01 crc kubenswrapper[4943]: I0307 14:42:01.957904 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drp64\" (UniqueName: \"kubernetes.io/projected/d41ebe43-4352-4732-bcc2-0ba95c3db967-kube-api-access-drp64\") pod \"route-controller-manager-69ff69cc58-vmx82\" (UID: \"d41ebe43-4352-4732-bcc2-0ba95c3db967\") " pod="openshift-route-controller-manager/route-controller-manager-69ff69cc58-vmx82" Mar 07 14:42:02 crc kubenswrapper[4943]: I0307 14:42:02.097895 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5764d979d6-l2tqn" Mar 07 14:42:02 crc kubenswrapper[4943]: I0307 14:42:02.107193 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-69ff69cc58-vmx82" Mar 07 14:42:02 crc kubenswrapper[4943]: I0307 14:42:02.378143 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5764d979d6-l2tqn"] Mar 07 14:42:02 crc kubenswrapper[4943]: I0307 14:42:02.422324 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-l7g67" Mar 07 14:42:02 crc kubenswrapper[4943]: I0307 14:42:02.559843 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-69ff69cc58-vmx82"] Mar 07 14:42:02 crc kubenswrapper[4943]: W0307 14:42:02.573950 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd41ebe43_4352_4732_bcc2_0ba95c3db967.slice/crio-b1b4acd4fda574b57723890cc3532dae99c2529470caa9fadbae1a8ea64968b3 WatchSource:0}: Error finding container b1b4acd4fda574b57723890cc3532dae99c2529470caa9fadbae1a8ea64968b3: Status 404 returned error can't find the container with id b1b4acd4fda574b57723890cc3532dae99c2529470caa9fadbae1a8ea64968b3 Mar 07 14:42:02 crc kubenswrapper[4943]: I0307 14:42:02.766462 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3374c43b-61ed-4e63-87d0-2b212fc70a67" path="/var/lib/kubelet/pods/3374c43b-61ed-4e63-87d0-2b212fc70a67/volumes" Mar 07 14:42:02 crc kubenswrapper[4943]: I0307 14:42:02.767425 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="782f34ac-63b1-452d-9157-344361ae3723" path="/var/lib/kubelet/pods/782f34ac-63b1-452d-9157-344361ae3723/volumes" Mar 07 14:42:03 crc kubenswrapper[4943]: I0307 14:42:03.061546 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-69ff69cc58-vmx82" event={"ID":"d41ebe43-4352-4732-bcc2-0ba95c3db967","Type":"ContainerStarted","Data":"0c9396ba0ef5b2a92e5c010959bc941492498db680facdb584f938fad62a1cfe"} Mar 07 14:42:03 crc kubenswrapper[4943]: I0307 14:42:03.061596 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-69ff69cc58-vmx82" event={"ID":"d41ebe43-4352-4732-bcc2-0ba95c3db967","Type":"ContainerStarted","Data":"b1b4acd4fda574b57723890cc3532dae99c2529470caa9fadbae1a8ea64968b3"} Mar 07 14:42:03 crc kubenswrapper[4943]: I0307 14:42:03.061859 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-69ff69cc58-vmx82" Mar 07 14:42:03 crc kubenswrapper[4943]: I0307 14:42:03.062877 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5764d979d6-l2tqn" event={"ID":"efb24914-9e13-4aed-bfd6-f2f6291a2750","Type":"ContainerStarted","Data":"72c7bb2e07d131c44d00530dffe92ee6bceb16f7bfd9b8974fe5b2a10b29c0aa"} Mar 07 14:42:03 crc kubenswrapper[4943]: I0307 14:42:03.062946 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5764d979d6-l2tqn" event={"ID":"efb24914-9e13-4aed-bfd6-f2f6291a2750","Type":"ContainerStarted","Data":"814258638d6cefa4892ddc4035575738a16852793279e7c9a109ea492126cc10"} Mar 07 14:42:03 crc kubenswrapper[4943]: I0307 14:42:03.063105 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5764d979d6-l2tqn" Mar 07 14:42:03 crc kubenswrapper[4943]: I0307 14:42:03.074269 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5764d979d6-l2tqn" Mar 07 14:42:03 crc kubenswrapper[4943]: I0307 14:42:03.081691 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-69ff69cc58-vmx82" podStartSLOduration=15.081673148 podStartE2EDuration="15.081673148s" podCreationTimestamp="2026-03-07 14:41:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:42:03.07882546 +0000 UTC m=+165.030961958" watchObservedRunningTime="2026-03-07 14:42:03.081673148 +0000 UTC m=+165.033809646" Mar 07 14:42:03 crc kubenswrapper[4943]: I0307 14:42:03.101984 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5764d979d6-l2tqn" podStartSLOduration=15.101963867 podStartE2EDuration="15.101963867s" podCreationTimestamp="2026-03-07 14:41:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:42:03.095089061 +0000 UTC m=+165.047225569" watchObservedRunningTime="2026-03-07 14:42:03.101963867 +0000 UTC m=+165.054100365" Mar 07 14:42:03 crc kubenswrapper[4943]: I0307 14:42:03.136499 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-69ff69cc58-vmx82" Mar 07 14:42:05 crc kubenswrapper[4943]: I0307 14:42:05.019535 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 07 14:42:05 crc kubenswrapper[4943]: I0307 14:42:05.020459 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 07 14:42:05 crc kubenswrapper[4943]: I0307 14:42:05.023798 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 07 14:42:05 crc kubenswrapper[4943]: I0307 14:42:05.024488 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 07 14:42:05 crc kubenswrapper[4943]: I0307 14:42:05.031159 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 07 14:42:05 crc kubenswrapper[4943]: I0307 14:42:05.086371 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/32ce6d8b-d46d-4d39-ab52-8fb8eed1643a-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"32ce6d8b-d46d-4d39-ab52-8fb8eed1643a\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 07 14:42:05 crc kubenswrapper[4943]: I0307 14:42:05.086416 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/32ce6d8b-d46d-4d39-ab52-8fb8eed1643a-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"32ce6d8b-d46d-4d39-ab52-8fb8eed1643a\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 07 14:42:05 crc kubenswrapper[4943]: I0307 14:42:05.187851 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/32ce6d8b-d46d-4d39-ab52-8fb8eed1643a-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"32ce6d8b-d46d-4d39-ab52-8fb8eed1643a\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 07 14:42:05 crc kubenswrapper[4943]: I0307 14:42:05.187945 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/32ce6d8b-d46d-4d39-ab52-8fb8eed1643a-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"32ce6d8b-d46d-4d39-ab52-8fb8eed1643a\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 07 14:42:05 crc kubenswrapper[4943]: I0307 14:42:05.188258 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/32ce6d8b-d46d-4d39-ab52-8fb8eed1643a-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"32ce6d8b-d46d-4d39-ab52-8fb8eed1643a\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 07 14:42:05 crc kubenswrapper[4943]: I0307 14:42:05.217005 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/32ce6d8b-d46d-4d39-ab52-8fb8eed1643a-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"32ce6d8b-d46d-4d39-ab52-8fb8eed1643a\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 07 14:42:05 crc kubenswrapper[4943]: I0307 14:42:05.350965 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 07 14:42:08 crc kubenswrapper[4943]: I0307 14:42:08.165828 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xwdpr" Mar 07 14:42:08 crc kubenswrapper[4943]: I0307 14:42:08.166518 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xwdpr" Mar 07 14:42:08 crc kubenswrapper[4943]: I0307 14:42:08.314104 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xwdpr" Mar 07 14:42:08 crc kubenswrapper[4943]: I0307 14:42:08.400637 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 07 14:42:08 crc kubenswrapper[4943]: W0307 14:42:08.408498 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod32ce6d8b_d46d_4d39_ab52_8fb8eed1643a.slice/crio-6d31df4647101e318d6b67d073a8e0ab7f8f7ee50be9d425475035138ebf5d68 WatchSource:0}: Error finding container 6d31df4647101e318d6b67d073a8e0ab7f8f7ee50be9d425475035138ebf5d68: Status 404 returned error can't find the container with id 6d31df4647101e318d6b67d073a8e0ab7f8f7ee50be9d425475035138ebf5d68 Mar 07 14:42:08 crc kubenswrapper[4943]: I0307 14:42:08.721909 4943 csr.go:261] certificate signing request csr-zz6h2 is approved, waiting to be issued Mar 07 14:42:08 crc kubenswrapper[4943]: I0307 14:42:08.731173 4943 csr.go:257] certificate signing request csr-zz6h2 is issued Mar 07 14:42:08 crc kubenswrapper[4943]: I0307 14:42:08.951175 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5764d979d6-l2tqn"] Mar 07 14:42:08 crc kubenswrapper[4943]: I0307 14:42:08.951413 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-5764d979d6-l2tqn" podUID="efb24914-9e13-4aed-bfd6-f2f6291a2750" containerName="controller-manager" containerID="cri-o://72c7bb2e07d131c44d00530dffe92ee6bceb16f7bfd9b8974fe5b2a10b29c0aa" gracePeriod=30 Mar 07 14:42:08 crc kubenswrapper[4943]: I0307 14:42:08.992568 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-69ff69cc58-vmx82"] Mar 07 14:42:08 crc kubenswrapper[4943]: I0307 14:42:08.992845 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-69ff69cc58-vmx82" podUID="d41ebe43-4352-4732-bcc2-0ba95c3db967" containerName="route-controller-manager" containerID="cri-o://0c9396ba0ef5b2a92e5c010959bc941492498db680facdb584f938fad62a1cfe" gracePeriod=30 Mar 07 14:42:09 crc kubenswrapper[4943]: I0307 14:42:09.116822 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"32ce6d8b-d46d-4d39-ab52-8fb8eed1643a","Type":"ContainerStarted","Data":"949bd4d511b7462a9f8d09f99c6e304851bc5e7dd78fd13568c872c9c4712ab1"} Mar 07 14:42:09 crc kubenswrapper[4943]: I0307 14:42:09.116884 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"32ce6d8b-d46d-4d39-ab52-8fb8eed1643a","Type":"ContainerStarted","Data":"6d31df4647101e318d6b67d073a8e0ab7f8f7ee50be9d425475035138ebf5d68"} Mar 07 14:42:09 crc kubenswrapper[4943]: I0307 14:42:09.119700 4943 generic.go:334] "Generic (PLEG): container finished" podID="efb24914-9e13-4aed-bfd6-f2f6291a2750" containerID="72c7bb2e07d131c44d00530dffe92ee6bceb16f7bfd9b8974fe5b2a10b29c0aa" exitCode=0 Mar 07 14:42:09 crc kubenswrapper[4943]: I0307 14:42:09.119760 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5764d979d6-l2tqn" event={"ID":"efb24914-9e13-4aed-bfd6-f2f6291a2750","Type":"ContainerDied","Data":"72c7bb2e07d131c44d00530dffe92ee6bceb16f7bfd9b8974fe5b2a10b29c0aa"} Mar 07 14:42:09 crc kubenswrapper[4943]: I0307 14:42:09.125313 4943 generic.go:334] "Generic (PLEG): container finished" podID="14d07994-bd34-4802-8dd5-8fd8afe5e878" containerID="ac2a8956e015025d29002a8478f290cd8b4087387667c0a6c52658f31f2d8176" exitCode=0 Mar 07 14:42:09 crc kubenswrapper[4943]: I0307 14:42:09.126139 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548242-scpk7" event={"ID":"14d07994-bd34-4802-8dd5-8fd8afe5e878","Type":"ContainerDied","Data":"ac2a8956e015025d29002a8478f290cd8b4087387667c0a6c52658f31f2d8176"} Mar 07 14:42:09 crc kubenswrapper[4943]: I0307 14:42:09.152992 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=4.152967298 podStartE2EDuration="4.152967298s" podCreationTimestamp="2026-03-07 14:42:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:42:09.131657614 +0000 UTC m=+171.083794112" watchObservedRunningTime="2026-03-07 14:42:09.152967298 +0000 UTC m=+171.105103796" Mar 07 14:42:09 crc kubenswrapper[4943]: I0307 14:42:09.192136 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xwdpr" Mar 07 14:42:09 crc kubenswrapper[4943]: I0307 14:42:09.486312 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-69ff69cc58-vmx82" Mar 07 14:42:09 crc kubenswrapper[4943]: I0307 14:42:09.515833 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5764d979d6-l2tqn" Mar 07 14:42:09 crc kubenswrapper[4943]: I0307 14:42:09.546794 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-drp64\" (UniqueName: \"kubernetes.io/projected/d41ebe43-4352-4732-bcc2-0ba95c3db967-kube-api-access-drp64\") pod \"d41ebe43-4352-4732-bcc2-0ba95c3db967\" (UID: \"d41ebe43-4352-4732-bcc2-0ba95c3db967\") " Mar 07 14:42:09 crc kubenswrapper[4943]: I0307 14:42:09.546879 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d41ebe43-4352-4732-bcc2-0ba95c3db967-client-ca\") pod \"d41ebe43-4352-4732-bcc2-0ba95c3db967\" (UID: \"d41ebe43-4352-4732-bcc2-0ba95c3db967\") " Mar 07 14:42:09 crc kubenswrapper[4943]: I0307 14:42:09.546906 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d41ebe43-4352-4732-bcc2-0ba95c3db967-config\") pod \"d41ebe43-4352-4732-bcc2-0ba95c3db967\" (UID: \"d41ebe43-4352-4732-bcc2-0ba95c3db967\") " Mar 07 14:42:09 crc kubenswrapper[4943]: I0307 14:42:09.547013 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d41ebe43-4352-4732-bcc2-0ba95c3db967-serving-cert\") pod \"d41ebe43-4352-4732-bcc2-0ba95c3db967\" (UID: \"d41ebe43-4352-4732-bcc2-0ba95c3db967\") " Mar 07 14:42:09 crc kubenswrapper[4943]: I0307 14:42:09.548005 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d41ebe43-4352-4732-bcc2-0ba95c3db967-config" (OuterVolumeSpecName: "config") pod "d41ebe43-4352-4732-bcc2-0ba95c3db967" (UID: "d41ebe43-4352-4732-bcc2-0ba95c3db967"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:42:09 crc kubenswrapper[4943]: I0307 14:42:09.549033 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d41ebe43-4352-4732-bcc2-0ba95c3db967-client-ca" (OuterVolumeSpecName: "client-ca") pod "d41ebe43-4352-4732-bcc2-0ba95c3db967" (UID: "d41ebe43-4352-4732-bcc2-0ba95c3db967"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:42:09 crc kubenswrapper[4943]: I0307 14:42:09.554548 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d41ebe43-4352-4732-bcc2-0ba95c3db967-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d41ebe43-4352-4732-bcc2-0ba95c3db967" (UID: "d41ebe43-4352-4732-bcc2-0ba95c3db967"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:42:09 crc kubenswrapper[4943]: I0307 14:42:09.554558 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d41ebe43-4352-4732-bcc2-0ba95c3db967-kube-api-access-drp64" (OuterVolumeSpecName: "kube-api-access-drp64") pod "d41ebe43-4352-4732-bcc2-0ba95c3db967" (UID: "d41ebe43-4352-4732-bcc2-0ba95c3db967"). InnerVolumeSpecName "kube-api-access-drp64". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:42:09 crc kubenswrapper[4943]: I0307 14:42:09.648799 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nc7f9\" (UniqueName: \"kubernetes.io/projected/efb24914-9e13-4aed-bfd6-f2f6291a2750-kube-api-access-nc7f9\") pod \"efb24914-9e13-4aed-bfd6-f2f6291a2750\" (UID: \"efb24914-9e13-4aed-bfd6-f2f6291a2750\") " Mar 07 14:42:09 crc kubenswrapper[4943]: I0307 14:42:09.648918 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/efb24914-9e13-4aed-bfd6-f2f6291a2750-proxy-ca-bundles\") pod \"efb24914-9e13-4aed-bfd6-f2f6291a2750\" (UID: \"efb24914-9e13-4aed-bfd6-f2f6291a2750\") " Mar 07 14:42:09 crc kubenswrapper[4943]: I0307 14:42:09.648979 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/efb24914-9e13-4aed-bfd6-f2f6291a2750-serving-cert\") pod \"efb24914-9e13-4aed-bfd6-f2f6291a2750\" (UID: \"efb24914-9e13-4aed-bfd6-f2f6291a2750\") " Mar 07 14:42:09 crc kubenswrapper[4943]: I0307 14:42:09.649131 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/efb24914-9e13-4aed-bfd6-f2f6291a2750-client-ca\") pod \"efb24914-9e13-4aed-bfd6-f2f6291a2750\" (UID: \"efb24914-9e13-4aed-bfd6-f2f6291a2750\") " Mar 07 14:42:09 crc kubenswrapper[4943]: I0307 14:42:09.649179 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/efb24914-9e13-4aed-bfd6-f2f6291a2750-config\") pod \"efb24914-9e13-4aed-bfd6-f2f6291a2750\" (UID: \"efb24914-9e13-4aed-bfd6-f2f6291a2750\") " Mar 07 14:42:09 crc kubenswrapper[4943]: I0307 14:42:09.649482 4943 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d41ebe43-4352-4732-bcc2-0ba95c3db967-config\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:09 crc kubenswrapper[4943]: I0307 14:42:09.649502 4943 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d41ebe43-4352-4732-bcc2-0ba95c3db967-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:09 crc kubenswrapper[4943]: I0307 14:42:09.649513 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-drp64\" (UniqueName: \"kubernetes.io/projected/d41ebe43-4352-4732-bcc2-0ba95c3db967-kube-api-access-drp64\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:09 crc kubenswrapper[4943]: I0307 14:42:09.649523 4943 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d41ebe43-4352-4732-bcc2-0ba95c3db967-client-ca\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:09 crc kubenswrapper[4943]: I0307 14:42:09.650177 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/efb24914-9e13-4aed-bfd6-f2f6291a2750-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "efb24914-9e13-4aed-bfd6-f2f6291a2750" (UID: "efb24914-9e13-4aed-bfd6-f2f6291a2750"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:42:09 crc kubenswrapper[4943]: I0307 14:42:09.650221 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/efb24914-9e13-4aed-bfd6-f2f6291a2750-client-ca" (OuterVolumeSpecName: "client-ca") pod "efb24914-9e13-4aed-bfd6-f2f6291a2750" (UID: "efb24914-9e13-4aed-bfd6-f2f6291a2750"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:42:09 crc kubenswrapper[4943]: I0307 14:42:09.650307 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/efb24914-9e13-4aed-bfd6-f2f6291a2750-config" (OuterVolumeSpecName: "config") pod "efb24914-9e13-4aed-bfd6-f2f6291a2750" (UID: "efb24914-9e13-4aed-bfd6-f2f6291a2750"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:42:09 crc kubenswrapper[4943]: I0307 14:42:09.654019 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efb24914-9e13-4aed-bfd6-f2f6291a2750-kube-api-access-nc7f9" (OuterVolumeSpecName: "kube-api-access-nc7f9") pod "efb24914-9e13-4aed-bfd6-f2f6291a2750" (UID: "efb24914-9e13-4aed-bfd6-f2f6291a2750"). InnerVolumeSpecName "kube-api-access-nc7f9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:42:09 crc kubenswrapper[4943]: I0307 14:42:09.654153 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efb24914-9e13-4aed-bfd6-f2f6291a2750-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "efb24914-9e13-4aed-bfd6-f2f6291a2750" (UID: "efb24914-9e13-4aed-bfd6-f2f6291a2750"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:42:09 crc kubenswrapper[4943]: I0307 14:42:09.732483 4943 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2027-01-06 10:26:10.625832759 +0000 UTC Mar 07 14:42:09 crc kubenswrapper[4943]: I0307 14:42:09.732527 4943 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 7315h44m0.89330863s for next certificate rotation Mar 07 14:42:09 crc kubenswrapper[4943]: I0307 14:42:09.750618 4943 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/efb24914-9e13-4aed-bfd6-f2f6291a2750-client-ca\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:09 crc kubenswrapper[4943]: I0307 14:42:09.750672 4943 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/efb24914-9e13-4aed-bfd6-f2f6291a2750-config\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:09 crc kubenswrapper[4943]: I0307 14:42:09.750693 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nc7f9\" (UniqueName: \"kubernetes.io/projected/efb24914-9e13-4aed-bfd6-f2f6291a2750-kube-api-access-nc7f9\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:09 crc kubenswrapper[4943]: I0307 14:42:09.750713 4943 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/efb24914-9e13-4aed-bfd6-f2f6291a2750-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:09 crc kubenswrapper[4943]: I0307 14:42:09.750731 4943 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/efb24914-9e13-4aed-bfd6-f2f6291a2750-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.134397 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5764d979d6-l2tqn" Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.134407 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5764d979d6-l2tqn" event={"ID":"efb24914-9e13-4aed-bfd6-f2f6291a2750","Type":"ContainerDied","Data":"814258638d6cefa4892ddc4035575738a16852793279e7c9a109ea492126cc10"} Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.134996 4943 scope.go:117] "RemoveContainer" containerID="72c7bb2e07d131c44d00530dffe92ee6bceb16f7bfd9b8974fe5b2a10b29c0aa" Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.138234 4943 generic.go:334] "Generic (PLEG): container finished" podID="32ce6d8b-d46d-4d39-ab52-8fb8eed1643a" containerID="949bd4d511b7462a9f8d09f99c6e304851bc5e7dd78fd13568c872c9c4712ab1" exitCode=0 Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.138317 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"32ce6d8b-d46d-4d39-ab52-8fb8eed1643a","Type":"ContainerDied","Data":"949bd4d511b7462a9f8d09f99c6e304851bc5e7dd78fd13568c872c9c4712ab1"} Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.141478 4943 generic.go:334] "Generic (PLEG): container finished" podID="d41ebe43-4352-4732-bcc2-0ba95c3db967" containerID="0c9396ba0ef5b2a92e5c010959bc941492498db680facdb584f938fad62a1cfe" exitCode=0 Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.141669 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-69ff69cc58-vmx82" Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.143710 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-69ff69cc58-vmx82" event={"ID":"d41ebe43-4352-4732-bcc2-0ba95c3db967","Type":"ContainerDied","Data":"0c9396ba0ef5b2a92e5c010959bc941492498db680facdb584f938fad62a1cfe"} Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.143833 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-69ff69cc58-vmx82" event={"ID":"d41ebe43-4352-4732-bcc2-0ba95c3db967","Type":"ContainerDied","Data":"b1b4acd4fda574b57723890cc3532dae99c2529470caa9fadbae1a8ea64968b3"} Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.180734 4943 scope.go:117] "RemoveContainer" containerID="0c9396ba0ef5b2a92e5c010959bc941492498db680facdb584f938fad62a1cfe" Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.182093 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5764d979d6-l2tqn"] Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.187263 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-5764d979d6-l2tqn"] Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.192431 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-69ff69cc58-vmx82"] Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.195213 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-69ff69cc58-vmx82"] Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.204909 4943 scope.go:117] "RemoveContainer" containerID="0c9396ba0ef5b2a92e5c010959bc941492498db680facdb584f938fad62a1cfe" Mar 07 14:42:10 crc kubenswrapper[4943]: E0307 14:42:10.205564 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c9396ba0ef5b2a92e5c010959bc941492498db680facdb584f938fad62a1cfe\": container with ID starting with 0c9396ba0ef5b2a92e5c010959bc941492498db680facdb584f938fad62a1cfe not found: ID does not exist" containerID="0c9396ba0ef5b2a92e5c010959bc941492498db680facdb584f938fad62a1cfe" Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.205594 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c9396ba0ef5b2a92e5c010959bc941492498db680facdb584f938fad62a1cfe"} err="failed to get container status \"0c9396ba0ef5b2a92e5c010959bc941492498db680facdb584f938fad62a1cfe\": rpc error: code = NotFound desc = could not find container \"0c9396ba0ef5b2a92e5c010959bc941492498db680facdb584f938fad62a1cfe\": container with ID starting with 0c9396ba0ef5b2a92e5c010959bc941492498db680facdb584f938fad62a1cfe not found: ID does not exist" Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.426946 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548242-scpk7" Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.563541 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n6c2r\" (UniqueName: \"kubernetes.io/projected/14d07994-bd34-4802-8dd5-8fd8afe5e878-kube-api-access-n6c2r\") pod \"14d07994-bd34-4802-8dd5-8fd8afe5e878\" (UID: \"14d07994-bd34-4802-8dd5-8fd8afe5e878\") " Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.574533 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14d07994-bd34-4802-8dd5-8fd8afe5e878-kube-api-access-n6c2r" (OuterVolumeSpecName: "kube-api-access-n6c2r") pod "14d07994-bd34-4802-8dd5-8fd8afe5e878" (UID: "14d07994-bd34-4802-8dd5-8fd8afe5e878"). InnerVolumeSpecName "kube-api-access-n6c2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.665513 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n6c2r\" (UniqueName: \"kubernetes.io/projected/14d07994-bd34-4802-8dd5-8fd8afe5e878-kube-api-access-n6c2r\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.733512 4943 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-12-27 00:41:59.005558816 +0000 UTC Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.733555 4943 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 7065h59m48.27200927s for next certificate rotation Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.764827 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d41ebe43-4352-4732-bcc2-0ba95c3db967" path="/var/lib/kubelet/pods/d41ebe43-4352-4732-bcc2-0ba95c3db967/volumes" Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.765660 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efb24914-9e13-4aed-bfd6-f2f6291a2750" path="/var/lib/kubelet/pods/efb24914-9e13-4aed-bfd6-f2f6291a2750/volumes" Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.817048 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-f957fc5cb-8r7lr"] Mar 07 14:42:10 crc kubenswrapper[4943]: E0307 14:42:10.817376 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14d07994-bd34-4802-8dd5-8fd8afe5e878" containerName="oc" Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.817392 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="14d07994-bd34-4802-8dd5-8fd8afe5e878" containerName="oc" Mar 07 14:42:10 crc kubenswrapper[4943]: E0307 14:42:10.817400 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efb24914-9e13-4aed-bfd6-f2f6291a2750" containerName="controller-manager" Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.817408 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="efb24914-9e13-4aed-bfd6-f2f6291a2750" containerName="controller-manager" Mar 07 14:42:10 crc kubenswrapper[4943]: E0307 14:42:10.817420 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d41ebe43-4352-4732-bcc2-0ba95c3db967" containerName="route-controller-manager" Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.817427 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="d41ebe43-4352-4732-bcc2-0ba95c3db967" containerName="route-controller-manager" Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.817543 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="d41ebe43-4352-4732-bcc2-0ba95c3db967" containerName="route-controller-manager" Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.817553 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="14d07994-bd34-4802-8dd5-8fd8afe5e878" containerName="oc" Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.817567 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="efb24914-9e13-4aed-bfd6-f2f6291a2750" containerName="controller-manager" Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.818238 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-f957fc5cb-8r7lr" Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.822224 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.823088 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.823228 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.823340 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.824555 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.826470 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.830402 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.842530 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-97747b5c5-wkxmj"] Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.844610 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-97747b5c5-wkxmj" Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.847539 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.847952 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.848275 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.848434 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.848667 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.848859 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.851091 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-97747b5c5-wkxmj"] Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.856474 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-f957fc5cb-8r7lr"] Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.878626 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6758e5ec-0b51-41ef-863d-aa72f41e6bb7-proxy-ca-bundles\") pod \"controller-manager-f957fc5cb-8r7lr\" (UID: \"6758e5ec-0b51-41ef-863d-aa72f41e6bb7\") " pod="openshift-controller-manager/controller-manager-f957fc5cb-8r7lr" Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.878673 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6758e5ec-0b51-41ef-863d-aa72f41e6bb7-client-ca\") pod \"controller-manager-f957fc5cb-8r7lr\" (UID: \"6758e5ec-0b51-41ef-863d-aa72f41e6bb7\") " pod="openshift-controller-manager/controller-manager-f957fc5cb-8r7lr" Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.878716 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6758e5ec-0b51-41ef-863d-aa72f41e6bb7-config\") pod \"controller-manager-f957fc5cb-8r7lr\" (UID: \"6758e5ec-0b51-41ef-863d-aa72f41e6bb7\") " pod="openshift-controller-manager/controller-manager-f957fc5cb-8r7lr" Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.878791 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6758e5ec-0b51-41ef-863d-aa72f41e6bb7-serving-cert\") pod \"controller-manager-f957fc5cb-8r7lr\" (UID: \"6758e5ec-0b51-41ef-863d-aa72f41e6bb7\") " pod="openshift-controller-manager/controller-manager-f957fc5cb-8r7lr" Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.878828 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hj2rp\" (UniqueName: \"kubernetes.io/projected/6758e5ec-0b51-41ef-863d-aa72f41e6bb7-kube-api-access-hj2rp\") pod \"controller-manager-f957fc5cb-8r7lr\" (UID: \"6758e5ec-0b51-41ef-863d-aa72f41e6bb7\") " pod="openshift-controller-manager/controller-manager-f957fc5cb-8r7lr" Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.979850 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4b9e1f64-5c9d-405c-820c-762a9ca4798a-client-ca\") pod \"route-controller-manager-97747b5c5-wkxmj\" (UID: \"4b9e1f64-5c9d-405c-820c-762a9ca4798a\") " pod="openshift-route-controller-manager/route-controller-manager-97747b5c5-wkxmj" Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.979916 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6758e5ec-0b51-41ef-863d-aa72f41e6bb7-serving-cert\") pod \"controller-manager-f957fc5cb-8r7lr\" (UID: \"6758e5ec-0b51-41ef-863d-aa72f41e6bb7\") " pod="openshift-controller-manager/controller-manager-f957fc5cb-8r7lr" Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.979971 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b9e1f64-5c9d-405c-820c-762a9ca4798a-serving-cert\") pod \"route-controller-manager-97747b5c5-wkxmj\" (UID: \"4b9e1f64-5c9d-405c-820c-762a9ca4798a\") " pod="openshift-route-controller-manager/route-controller-manager-97747b5c5-wkxmj" Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.979995 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hj2rp\" (UniqueName: \"kubernetes.io/projected/6758e5ec-0b51-41ef-863d-aa72f41e6bb7-kube-api-access-hj2rp\") pod \"controller-manager-f957fc5cb-8r7lr\" (UID: \"6758e5ec-0b51-41ef-863d-aa72f41e6bb7\") " pod="openshift-controller-manager/controller-manager-f957fc5cb-8r7lr" Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.980107 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvv85\" (UniqueName: \"kubernetes.io/projected/4b9e1f64-5c9d-405c-820c-762a9ca4798a-kube-api-access-kvv85\") pod \"route-controller-manager-97747b5c5-wkxmj\" (UID: \"4b9e1f64-5c9d-405c-820c-762a9ca4798a\") " pod="openshift-route-controller-manager/route-controller-manager-97747b5c5-wkxmj" Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.980208 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b9e1f64-5c9d-405c-820c-762a9ca4798a-config\") pod \"route-controller-manager-97747b5c5-wkxmj\" (UID: \"4b9e1f64-5c9d-405c-820c-762a9ca4798a\") " pod="openshift-route-controller-manager/route-controller-manager-97747b5c5-wkxmj" Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.980262 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6758e5ec-0b51-41ef-863d-aa72f41e6bb7-proxy-ca-bundles\") pod \"controller-manager-f957fc5cb-8r7lr\" (UID: \"6758e5ec-0b51-41ef-863d-aa72f41e6bb7\") " pod="openshift-controller-manager/controller-manager-f957fc5cb-8r7lr" Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.980283 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6758e5ec-0b51-41ef-863d-aa72f41e6bb7-client-ca\") pod \"controller-manager-f957fc5cb-8r7lr\" (UID: \"6758e5ec-0b51-41ef-863d-aa72f41e6bb7\") " pod="openshift-controller-manager/controller-manager-f957fc5cb-8r7lr" Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.980313 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6758e5ec-0b51-41ef-863d-aa72f41e6bb7-config\") pod \"controller-manager-f957fc5cb-8r7lr\" (UID: \"6758e5ec-0b51-41ef-863d-aa72f41e6bb7\") " pod="openshift-controller-manager/controller-manager-f957fc5cb-8r7lr" Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.981124 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6758e5ec-0b51-41ef-863d-aa72f41e6bb7-client-ca\") pod \"controller-manager-f957fc5cb-8r7lr\" (UID: \"6758e5ec-0b51-41ef-863d-aa72f41e6bb7\") " pod="openshift-controller-manager/controller-manager-f957fc5cb-8r7lr" Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.981708 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6758e5ec-0b51-41ef-863d-aa72f41e6bb7-config\") pod \"controller-manager-f957fc5cb-8r7lr\" (UID: \"6758e5ec-0b51-41ef-863d-aa72f41e6bb7\") " pod="openshift-controller-manager/controller-manager-f957fc5cb-8r7lr" Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.981848 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6758e5ec-0b51-41ef-863d-aa72f41e6bb7-proxy-ca-bundles\") pod \"controller-manager-f957fc5cb-8r7lr\" (UID: \"6758e5ec-0b51-41ef-863d-aa72f41e6bb7\") " pod="openshift-controller-manager/controller-manager-f957fc5cb-8r7lr" Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.984015 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6758e5ec-0b51-41ef-863d-aa72f41e6bb7-serving-cert\") pod \"controller-manager-f957fc5cb-8r7lr\" (UID: \"6758e5ec-0b51-41ef-863d-aa72f41e6bb7\") " pod="openshift-controller-manager/controller-manager-f957fc5cb-8r7lr" Mar 07 14:42:10 crc kubenswrapper[4943]: I0307 14:42:10.995365 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hj2rp\" (UniqueName: \"kubernetes.io/projected/6758e5ec-0b51-41ef-863d-aa72f41e6bb7-kube-api-access-hj2rp\") pod \"controller-manager-f957fc5cb-8r7lr\" (UID: \"6758e5ec-0b51-41ef-863d-aa72f41e6bb7\") " pod="openshift-controller-manager/controller-manager-f957fc5cb-8r7lr" Mar 07 14:42:11 crc kubenswrapper[4943]: I0307 14:42:11.081607 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b9e1f64-5c9d-405c-820c-762a9ca4798a-config\") pod \"route-controller-manager-97747b5c5-wkxmj\" (UID: \"4b9e1f64-5c9d-405c-820c-762a9ca4798a\") " pod="openshift-route-controller-manager/route-controller-manager-97747b5c5-wkxmj" Mar 07 14:42:11 crc kubenswrapper[4943]: I0307 14:42:11.081709 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4b9e1f64-5c9d-405c-820c-762a9ca4798a-client-ca\") pod \"route-controller-manager-97747b5c5-wkxmj\" (UID: \"4b9e1f64-5c9d-405c-820c-762a9ca4798a\") " pod="openshift-route-controller-manager/route-controller-manager-97747b5c5-wkxmj" Mar 07 14:42:11 crc kubenswrapper[4943]: I0307 14:42:11.081771 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b9e1f64-5c9d-405c-820c-762a9ca4798a-serving-cert\") pod \"route-controller-manager-97747b5c5-wkxmj\" (UID: \"4b9e1f64-5c9d-405c-820c-762a9ca4798a\") " pod="openshift-route-controller-manager/route-controller-manager-97747b5c5-wkxmj" Mar 07 14:42:11 crc kubenswrapper[4943]: I0307 14:42:11.081804 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvv85\" (UniqueName: \"kubernetes.io/projected/4b9e1f64-5c9d-405c-820c-762a9ca4798a-kube-api-access-kvv85\") pod \"route-controller-manager-97747b5c5-wkxmj\" (UID: \"4b9e1f64-5c9d-405c-820c-762a9ca4798a\") " pod="openshift-route-controller-manager/route-controller-manager-97747b5c5-wkxmj" Mar 07 14:42:11 crc kubenswrapper[4943]: I0307 14:42:11.083349 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4b9e1f64-5c9d-405c-820c-762a9ca4798a-client-ca\") pod \"route-controller-manager-97747b5c5-wkxmj\" (UID: \"4b9e1f64-5c9d-405c-820c-762a9ca4798a\") " pod="openshift-route-controller-manager/route-controller-manager-97747b5c5-wkxmj" Mar 07 14:42:11 crc kubenswrapper[4943]: I0307 14:42:11.083439 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b9e1f64-5c9d-405c-820c-762a9ca4798a-config\") pod \"route-controller-manager-97747b5c5-wkxmj\" (UID: \"4b9e1f64-5c9d-405c-820c-762a9ca4798a\") " pod="openshift-route-controller-manager/route-controller-manager-97747b5c5-wkxmj" Mar 07 14:42:11 crc kubenswrapper[4943]: I0307 14:42:11.087225 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b9e1f64-5c9d-405c-820c-762a9ca4798a-serving-cert\") pod \"route-controller-manager-97747b5c5-wkxmj\" (UID: \"4b9e1f64-5c9d-405c-820c-762a9ca4798a\") " pod="openshift-route-controller-manager/route-controller-manager-97747b5c5-wkxmj" Mar 07 14:42:11 crc kubenswrapper[4943]: I0307 14:42:11.107705 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvv85\" (UniqueName: \"kubernetes.io/projected/4b9e1f64-5c9d-405c-820c-762a9ca4798a-kube-api-access-kvv85\") pod \"route-controller-manager-97747b5c5-wkxmj\" (UID: \"4b9e1f64-5c9d-405c-820c-762a9ca4798a\") " pod="openshift-route-controller-manager/route-controller-manager-97747b5c5-wkxmj" Mar 07 14:42:11 crc kubenswrapper[4943]: I0307 14:42:11.151852 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548242-scpk7" event={"ID":"14d07994-bd34-4802-8dd5-8fd8afe5e878","Type":"ContainerDied","Data":"2b619c7fa4f294863fb14b75a2ddfe9cb31a5bb989803badfb17c0ea218fed2d"} Mar 07 14:42:11 crc kubenswrapper[4943]: I0307 14:42:11.151939 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2b619c7fa4f294863fb14b75a2ddfe9cb31a5bb989803badfb17c0ea218fed2d" Mar 07 14:42:11 crc kubenswrapper[4943]: I0307 14:42:11.151958 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548242-scpk7" Mar 07 14:42:11 crc kubenswrapper[4943]: I0307 14:42:11.154491 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jrc6h" event={"ID":"f10fa8f5-f504-40c9-81a5-d1658c1ef268","Type":"ContainerDied","Data":"20bdfe5cb0c8c7149d03c99da6b8dec4af6ec25d0b1fa3174e4cced7f6a7790b"} Mar 07 14:42:11 crc kubenswrapper[4943]: I0307 14:42:11.154497 4943 generic.go:334] "Generic (PLEG): container finished" podID="f10fa8f5-f504-40c9-81a5-d1658c1ef268" containerID="20bdfe5cb0c8c7149d03c99da6b8dec4af6ec25d0b1fa3174e4cced7f6a7790b" exitCode=0 Mar 07 14:42:11 crc kubenswrapper[4943]: I0307 14:42:11.192723 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-f957fc5cb-8r7lr" Mar 07 14:42:11 crc kubenswrapper[4943]: I0307 14:42:11.203608 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-97747b5c5-wkxmj" Mar 07 14:42:11 crc kubenswrapper[4943]: I0307 14:42:11.497795 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 07 14:42:11 crc kubenswrapper[4943]: I0307 14:42:11.590387 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/32ce6d8b-d46d-4d39-ab52-8fb8eed1643a-kube-api-access\") pod \"32ce6d8b-d46d-4d39-ab52-8fb8eed1643a\" (UID: \"32ce6d8b-d46d-4d39-ab52-8fb8eed1643a\") " Mar 07 14:42:11 crc kubenswrapper[4943]: I0307 14:42:11.590465 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/32ce6d8b-d46d-4d39-ab52-8fb8eed1643a-kubelet-dir\") pod \"32ce6d8b-d46d-4d39-ab52-8fb8eed1643a\" (UID: \"32ce6d8b-d46d-4d39-ab52-8fb8eed1643a\") " Mar 07 14:42:11 crc kubenswrapper[4943]: I0307 14:42:11.590744 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/32ce6d8b-d46d-4d39-ab52-8fb8eed1643a-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "32ce6d8b-d46d-4d39-ab52-8fb8eed1643a" (UID: "32ce6d8b-d46d-4d39-ab52-8fb8eed1643a"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 07 14:42:11 crc kubenswrapper[4943]: I0307 14:42:11.600067 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32ce6d8b-d46d-4d39-ab52-8fb8eed1643a-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "32ce6d8b-d46d-4d39-ab52-8fb8eed1643a" (UID: "32ce6d8b-d46d-4d39-ab52-8fb8eed1643a"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:42:11 crc kubenswrapper[4943]: I0307 14:42:11.673801 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-f957fc5cb-8r7lr"] Mar 07 14:42:11 crc kubenswrapper[4943]: I0307 14:42:11.682533 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-97747b5c5-wkxmj"] Mar 07 14:42:11 crc kubenswrapper[4943]: W0307 14:42:11.687461 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6758e5ec_0b51_41ef_863d_aa72f41e6bb7.slice/crio-ad473c1eafdda09c6b7a142f4bdf574574b80c7fbf9fab92a4433393e9d7971d WatchSource:0}: Error finding container ad473c1eafdda09c6b7a142f4bdf574574b80c7fbf9fab92a4433393e9d7971d: Status 404 returned error can't find the container with id ad473c1eafdda09c6b7a142f4bdf574574b80c7fbf9fab92a4433393e9d7971d Mar 07 14:42:11 crc kubenswrapper[4943]: W0307 14:42:11.689674 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4b9e1f64_5c9d_405c_820c_762a9ca4798a.slice/crio-34940292bdcaa5d49d8fccfd644e02091897a91cb218f557eae574d10251ad55 WatchSource:0}: Error finding container 34940292bdcaa5d49d8fccfd644e02091897a91cb218f557eae574d10251ad55: Status 404 returned error can't find the container with id 34940292bdcaa5d49d8fccfd644e02091897a91cb218f557eae574d10251ad55 Mar 07 14:42:11 crc kubenswrapper[4943]: I0307 14:42:11.691638 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/32ce6d8b-d46d-4d39-ab52-8fb8eed1643a-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:11 crc kubenswrapper[4943]: I0307 14:42:11.691748 4943 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/32ce6d8b-d46d-4d39-ab52-8fb8eed1643a-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:12 crc kubenswrapper[4943]: I0307 14:42:12.161680 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 07 14:42:12 crc kubenswrapper[4943]: I0307 14:42:12.161667 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"32ce6d8b-d46d-4d39-ab52-8fb8eed1643a","Type":"ContainerDied","Data":"6d31df4647101e318d6b67d073a8e0ab7f8f7ee50be9d425475035138ebf5d68"} Mar 07 14:42:12 crc kubenswrapper[4943]: I0307 14:42:12.161794 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6d31df4647101e318d6b67d073a8e0ab7f8f7ee50be9d425475035138ebf5d68" Mar 07 14:42:12 crc kubenswrapper[4943]: I0307 14:42:12.163081 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-f957fc5cb-8r7lr" event={"ID":"6758e5ec-0b51-41ef-863d-aa72f41e6bb7","Type":"ContainerStarted","Data":"72b50585ad9e852d9b58596507515bc87cac3e6c9e04731e4dc63a89ae20103c"} Mar 07 14:42:12 crc kubenswrapper[4943]: I0307 14:42:12.163129 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-f957fc5cb-8r7lr" event={"ID":"6758e5ec-0b51-41ef-863d-aa72f41e6bb7","Type":"ContainerStarted","Data":"ad473c1eafdda09c6b7a142f4bdf574574b80c7fbf9fab92a4433393e9d7971d"} Mar 07 14:42:12 crc kubenswrapper[4943]: I0307 14:42:12.163392 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-f957fc5cb-8r7lr" Mar 07 14:42:12 crc kubenswrapper[4943]: I0307 14:42:12.166517 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-97747b5c5-wkxmj" event={"ID":"4b9e1f64-5c9d-405c-820c-762a9ca4798a","Type":"ContainerStarted","Data":"55524a95d9cfc406b363ac3a85ba9c341ea971eb1b3979f5c12d750721660263"} Mar 07 14:42:12 crc kubenswrapper[4943]: I0307 14:42:12.166547 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-97747b5c5-wkxmj" event={"ID":"4b9e1f64-5c9d-405c-820c-762a9ca4798a","Type":"ContainerStarted","Data":"34940292bdcaa5d49d8fccfd644e02091897a91cb218f557eae574d10251ad55"} Mar 07 14:42:12 crc kubenswrapper[4943]: I0307 14:42:12.166715 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-97747b5c5-wkxmj" Mar 07 14:42:12 crc kubenswrapper[4943]: I0307 14:42:12.169629 4943 generic.go:334] "Generic (PLEG): container finished" podID="74a7b99c-3412-438e-a835-1e153a56047b" containerID="ad78b2587ac7c0ce5bc3a0ea769519e07e704a8ecaaf863e3c02fc49d954d63a" exitCode=0 Mar 07 14:42:12 crc kubenswrapper[4943]: I0307 14:42:12.169697 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vkcxl" event={"ID":"74a7b99c-3412-438e-a835-1e153a56047b","Type":"ContainerDied","Data":"ad78b2587ac7c0ce5bc3a0ea769519e07e704a8ecaaf863e3c02fc49d954d63a"} Mar 07 14:42:12 crc kubenswrapper[4943]: I0307 14:42:12.177154 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jrc6h" event={"ID":"f10fa8f5-f504-40c9-81a5-d1658c1ef268","Type":"ContainerStarted","Data":"2589f77538011f2ab019b72637e957313f033e18e37f87b0825a4312ec83bbd5"} Mar 07 14:42:12 crc kubenswrapper[4943]: I0307 14:42:12.179912 4943 generic.go:334] "Generic (PLEG): container finished" podID="23fa83e7-df77-47e2-bb98-3344e48ab6bd" containerID="0b7ee43e866c549159478b2d441e21f9c48f78bb88ed5e53a91b6512934df95c" exitCode=0 Mar 07 14:42:12 crc kubenswrapper[4943]: I0307 14:42:12.180006 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9sf6k" event={"ID":"23fa83e7-df77-47e2-bb98-3344e48ab6bd","Type":"ContainerDied","Data":"0b7ee43e866c549159478b2d441e21f9c48f78bb88ed5e53a91b6512934df95c"} Mar 07 14:42:12 crc kubenswrapper[4943]: I0307 14:42:12.203362 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-f957fc5cb-8r7lr" podStartSLOduration=4.203337576 podStartE2EDuration="4.203337576s" podCreationTimestamp="2026-03-07 14:42:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:42:12.201711467 +0000 UTC m=+174.153847965" watchObservedRunningTime="2026-03-07 14:42:12.203337576 +0000 UTC m=+174.155474074" Mar 07 14:42:12 crc kubenswrapper[4943]: I0307 14:42:12.206394 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-f957fc5cb-8r7lr" Mar 07 14:42:12 crc kubenswrapper[4943]: I0307 14:42:12.263481 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-97747b5c5-wkxmj" podStartSLOduration=3.263462915 podStartE2EDuration="3.263462915s" podCreationTimestamp="2026-03-07 14:42:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:42:12.246561268 +0000 UTC m=+174.198697766" watchObservedRunningTime="2026-03-07 14:42:12.263462915 +0000 UTC m=+174.215599403" Mar 07 14:42:12 crc kubenswrapper[4943]: I0307 14:42:12.266021 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jrc6h" podStartSLOduration=3.720348693 podStartE2EDuration="47.266014346s" podCreationTimestamp="2026-03-07 14:41:25 +0000 UTC" firstStartedPulling="2026-03-07 14:41:28.044435919 +0000 UTC m=+129.996572417" lastFinishedPulling="2026-03-07 14:42:11.590101582 +0000 UTC m=+173.542238070" observedRunningTime="2026-03-07 14:42:12.261661981 +0000 UTC m=+174.213798479" watchObservedRunningTime="2026-03-07 14:42:12.266014346 +0000 UTC m=+174.218150844" Mar 07 14:42:12 crc kubenswrapper[4943]: I0307 14:42:12.405636 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-97747b5c5-wkxmj" Mar 07 14:42:12 crc kubenswrapper[4943]: I0307 14:42:12.618352 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 07 14:42:12 crc kubenswrapper[4943]: E0307 14:42:12.618568 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32ce6d8b-d46d-4d39-ab52-8fb8eed1643a" containerName="pruner" Mar 07 14:42:12 crc kubenswrapper[4943]: I0307 14:42:12.618580 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="32ce6d8b-d46d-4d39-ab52-8fb8eed1643a" containerName="pruner" Mar 07 14:42:12 crc kubenswrapper[4943]: I0307 14:42:12.618679 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="32ce6d8b-d46d-4d39-ab52-8fb8eed1643a" containerName="pruner" Mar 07 14:42:12 crc kubenswrapper[4943]: I0307 14:42:12.619031 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 07 14:42:12 crc kubenswrapper[4943]: I0307 14:42:12.623262 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 07 14:42:12 crc kubenswrapper[4943]: I0307 14:42:12.623277 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 07 14:42:12 crc kubenswrapper[4943]: I0307 14:42:12.688541 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 07 14:42:12 crc kubenswrapper[4943]: I0307 14:42:12.708188 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/979a9c5a-4f69-4b26-8127-f760a697ae91-kubelet-dir\") pod \"installer-9-crc\" (UID: \"979a9c5a-4f69-4b26-8127-f760a697ae91\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 07 14:42:12 crc kubenswrapper[4943]: I0307 14:42:12.708298 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/979a9c5a-4f69-4b26-8127-f760a697ae91-var-lock\") pod \"installer-9-crc\" (UID: \"979a9c5a-4f69-4b26-8127-f760a697ae91\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 07 14:42:12 crc kubenswrapper[4943]: I0307 14:42:12.708347 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/979a9c5a-4f69-4b26-8127-f760a697ae91-kube-api-access\") pod \"installer-9-crc\" (UID: \"979a9c5a-4f69-4b26-8127-f760a697ae91\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 07 14:42:12 crc kubenswrapper[4943]: I0307 14:42:12.809653 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/979a9c5a-4f69-4b26-8127-f760a697ae91-var-lock\") pod \"installer-9-crc\" (UID: \"979a9c5a-4f69-4b26-8127-f760a697ae91\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 07 14:42:12 crc kubenswrapper[4943]: I0307 14:42:12.809887 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/979a9c5a-4f69-4b26-8127-f760a697ae91-var-lock\") pod \"installer-9-crc\" (UID: \"979a9c5a-4f69-4b26-8127-f760a697ae91\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 07 14:42:12 crc kubenswrapper[4943]: I0307 14:42:12.810020 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/979a9c5a-4f69-4b26-8127-f760a697ae91-kube-api-access\") pod \"installer-9-crc\" (UID: \"979a9c5a-4f69-4b26-8127-f760a697ae91\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 07 14:42:12 crc kubenswrapper[4943]: I0307 14:42:12.810107 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/979a9c5a-4f69-4b26-8127-f760a697ae91-kubelet-dir\") pod \"installer-9-crc\" (UID: \"979a9c5a-4f69-4b26-8127-f760a697ae91\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 07 14:42:12 crc kubenswrapper[4943]: I0307 14:42:12.810316 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/979a9c5a-4f69-4b26-8127-f760a697ae91-kubelet-dir\") pod \"installer-9-crc\" (UID: \"979a9c5a-4f69-4b26-8127-f760a697ae91\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 07 14:42:12 crc kubenswrapper[4943]: I0307 14:42:12.833293 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/979a9c5a-4f69-4b26-8127-f760a697ae91-kube-api-access\") pod \"installer-9-crc\" (UID: \"979a9c5a-4f69-4b26-8127-f760a697ae91\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 07 14:42:12 crc kubenswrapper[4943]: I0307 14:42:12.998073 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 07 14:42:13 crc kubenswrapper[4943]: I0307 14:42:13.205973 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vkcxl" event={"ID":"74a7b99c-3412-438e-a835-1e153a56047b","Type":"ContainerStarted","Data":"d5b0b520a7e836b5a5d4d0c817a049499e6e7943b8fbc86b71ccfd77656df4bf"} Mar 07 14:42:13 crc kubenswrapper[4943]: I0307 14:42:13.215002 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9sf6k" event={"ID":"23fa83e7-df77-47e2-bb98-3344e48ab6bd","Type":"ContainerStarted","Data":"f2e01ac1b3b98ea5ad9485135e103568e4998789d36b202b8123967b665c26ca"} Mar 07 14:42:13 crc kubenswrapper[4943]: I0307 14:42:13.217907 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f95bw" event={"ID":"a41f89c1-5f6d-46bc-801b-a29ebebf4468","Type":"ContainerStarted","Data":"8be7d1144cb9e9546f622f3ad7227b4339e050c7ecf5514fcf23425048fc8ff5"} Mar 07 14:42:13 crc kubenswrapper[4943]: I0307 14:42:13.244021 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-vkcxl" podStartSLOduration=3.626978148 podStartE2EDuration="48.244001997s" podCreationTimestamp="2026-03-07 14:41:25 +0000 UTC" firstStartedPulling="2026-03-07 14:41:27.977125323 +0000 UTC m=+129.929261821" lastFinishedPulling="2026-03-07 14:42:12.594149172 +0000 UTC m=+174.546285670" observedRunningTime="2026-03-07 14:42:13.227859078 +0000 UTC m=+175.179995576" watchObservedRunningTime="2026-03-07 14:42:13.244001997 +0000 UTC m=+175.196138495" Mar 07 14:42:13 crc kubenswrapper[4943]: I0307 14:42:13.465100 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9sf6k" podStartSLOduration=2.97517953 podStartE2EDuration="47.465073283s" podCreationTimestamp="2026-03-07 14:41:26 +0000 UTC" firstStartedPulling="2026-03-07 14:41:28.146271191 +0000 UTC m=+130.098407699" lastFinishedPulling="2026-03-07 14:42:12.636164954 +0000 UTC m=+174.588301452" observedRunningTime="2026-03-07 14:42:13.259154682 +0000 UTC m=+175.211291190" watchObservedRunningTime="2026-03-07 14:42:13.465073283 +0000 UTC m=+175.417209781" Mar 07 14:42:13 crc kubenswrapper[4943]: I0307 14:42:13.470459 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 07 14:42:14 crc kubenswrapper[4943]: I0307 14:42:14.225642 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rvkgz" event={"ID":"4ae1cf01-1400-4500-9b84-8435a6541612","Type":"ContainerStarted","Data":"a219180b9098cbaad437ff15fcfcb8445f7a4b73c0fe03c40cf487762dee6dad"} Mar 07 14:42:14 crc kubenswrapper[4943]: I0307 14:42:14.228430 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"979a9c5a-4f69-4b26-8127-f760a697ae91","Type":"ContainerStarted","Data":"ed8efad2a01435fb5db9bfd68f4303f464293e51ca1a7be35011ad53a6dfb3bf"} Mar 07 14:42:14 crc kubenswrapper[4943]: I0307 14:42:14.228462 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"979a9c5a-4f69-4b26-8127-f760a697ae91","Type":"ContainerStarted","Data":"b5d195dd1186d19e7481f63b38ef59140069d31599dbe38908f4898eae4c0701"} Mar 07 14:42:14 crc kubenswrapper[4943]: I0307 14:42:14.231242 4943 generic.go:334] "Generic (PLEG): container finished" podID="a41f89c1-5f6d-46bc-801b-a29ebebf4468" containerID="8be7d1144cb9e9546f622f3ad7227b4339e050c7ecf5514fcf23425048fc8ff5" exitCode=0 Mar 07 14:42:14 crc kubenswrapper[4943]: I0307 14:42:14.231617 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f95bw" event={"ID":"a41f89c1-5f6d-46bc-801b-a29ebebf4468","Type":"ContainerDied","Data":"8be7d1144cb9e9546f622f3ad7227b4339e050c7ecf5514fcf23425048fc8ff5"} Mar 07 14:42:14 crc kubenswrapper[4943]: I0307 14:42:14.271858 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=2.27183318 podStartE2EDuration="2.27183318s" podCreationTimestamp="2026-03-07 14:42:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:42:14.26854335 +0000 UTC m=+176.220679848" watchObservedRunningTime="2026-03-07 14:42:14.27183318 +0000 UTC m=+176.223969688" Mar 07 14:42:15 crc kubenswrapper[4943]: I0307 14:42:15.239570 4943 generic.go:334] "Generic (PLEG): container finished" podID="4ae1cf01-1400-4500-9b84-8435a6541612" containerID="a219180b9098cbaad437ff15fcfcb8445f7a4b73c0fe03c40cf487762dee6dad" exitCode=0 Mar 07 14:42:15 crc kubenswrapper[4943]: I0307 14:42:15.239651 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rvkgz" event={"ID":"4ae1cf01-1400-4500-9b84-8435a6541612","Type":"ContainerDied","Data":"a219180b9098cbaad437ff15fcfcb8445f7a4b73c0fe03c40cf487762dee6dad"} Mar 07 14:42:16 crc kubenswrapper[4943]: I0307 14:42:16.174700 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jrc6h" Mar 07 14:42:16 crc kubenswrapper[4943]: I0307 14:42:16.175401 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jrc6h" Mar 07 14:42:16 crc kubenswrapper[4943]: I0307 14:42:16.239591 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jrc6h" Mar 07 14:42:16 crc kubenswrapper[4943]: I0307 14:42:16.249376 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f95bw" event={"ID":"a41f89c1-5f6d-46bc-801b-a29ebebf4468","Type":"ContainerStarted","Data":"1b1573f313b949b3ea11b97a155d031e94aeebb27dc909c06ba03f22ba4e6cfa"} Mar 07 14:42:16 crc kubenswrapper[4943]: I0307 14:42:16.296840 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jrc6h" Mar 07 14:42:16 crc kubenswrapper[4943]: I0307 14:42:16.300469 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-f95bw" podStartSLOduration=4.679027464 podStartE2EDuration="48.300436953s" podCreationTimestamp="2026-03-07 14:41:28 +0000 UTC" firstStartedPulling="2026-03-07 14:41:31.535674128 +0000 UTC m=+133.487810626" lastFinishedPulling="2026-03-07 14:42:15.157083577 +0000 UTC m=+177.109220115" observedRunningTime="2026-03-07 14:42:16.295690669 +0000 UTC m=+178.247827177" watchObservedRunningTime="2026-03-07 14:42:16.300436953 +0000 UTC m=+178.252573451" Mar 07 14:42:16 crc kubenswrapper[4943]: I0307 14:42:16.370355 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-vkcxl" Mar 07 14:42:16 crc kubenswrapper[4943]: I0307 14:42:16.370436 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-vkcxl" Mar 07 14:42:16 crc kubenswrapper[4943]: I0307 14:42:16.418327 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-vkcxl" Mar 07 14:42:16 crc kubenswrapper[4943]: I0307 14:42:16.527904 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9sf6k" Mar 07 14:42:16 crc kubenswrapper[4943]: I0307 14:42:16.528030 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9sf6k" Mar 07 14:42:16 crc kubenswrapper[4943]: I0307 14:42:16.567533 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9sf6k" Mar 07 14:42:17 crc kubenswrapper[4943]: I0307 14:42:17.292469 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9sf6k" Mar 07 14:42:17 crc kubenswrapper[4943]: I0307 14:42:17.303502 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-vkcxl" Mar 07 14:42:18 crc kubenswrapper[4943]: I0307 14:42:18.142017 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vkcxl"] Mar 07 14:42:19 crc kubenswrapper[4943]: I0307 14:42:19.153168 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-f95bw" Mar 07 14:42:19 crc kubenswrapper[4943]: I0307 14:42:19.153627 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-f95bw" Mar 07 14:42:19 crc kubenswrapper[4943]: I0307 14:42:19.266435 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-vkcxl" podUID="74a7b99c-3412-438e-a835-1e153a56047b" containerName="registry-server" containerID="cri-o://d5b0b520a7e836b5a5d4d0c817a049499e6e7943b8fbc86b71ccfd77656df4bf" gracePeriod=2 Mar 07 14:42:20 crc kubenswrapper[4943]: E0307 14:42:20.183594 4943 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod74a7b99c_3412_438e_a835_1e153a56047b.slice/crio-conmon-d5b0b520a7e836b5a5d4d0c817a049499e6e7943b8fbc86b71ccfd77656df4bf.scope\": RecentStats: unable to find data in memory cache]" Mar 07 14:42:20 crc kubenswrapper[4943]: I0307 14:42:20.196854 4943 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-f95bw" podUID="a41f89c1-5f6d-46bc-801b-a29ebebf4468" containerName="registry-server" probeResult="failure" output=< Mar 07 14:42:20 crc kubenswrapper[4943]: timeout: failed to connect service ":50051" within 1s Mar 07 14:42:20 crc kubenswrapper[4943]: > Mar 07 14:42:20 crc kubenswrapper[4943]: I0307 14:42:20.273639 4943 generic.go:334] "Generic (PLEG): container finished" podID="6a05c7ae-f2a1-4116-9038-ac838fa88af7" containerID="dacb590c8ccf77514c2fe8c0e0bc44da9380e7b5bbfd99d17df8072d8e55f4fe" exitCode=0 Mar 07 14:42:20 crc kubenswrapper[4943]: I0307 14:42:20.273744 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p767n" event={"ID":"6a05c7ae-f2a1-4116-9038-ac838fa88af7","Type":"ContainerDied","Data":"dacb590c8ccf77514c2fe8c0e0bc44da9380e7b5bbfd99d17df8072d8e55f4fe"} Mar 07 14:42:20 crc kubenswrapper[4943]: I0307 14:42:20.277150 4943 generic.go:334] "Generic (PLEG): container finished" podID="74a7b99c-3412-438e-a835-1e153a56047b" containerID="d5b0b520a7e836b5a5d4d0c817a049499e6e7943b8fbc86b71ccfd77656df4bf" exitCode=0 Mar 07 14:42:20 crc kubenswrapper[4943]: I0307 14:42:20.277206 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vkcxl" event={"ID":"74a7b99c-3412-438e-a835-1e153a56047b","Type":"ContainerDied","Data":"d5b0b520a7e836b5a5d4d0c817a049499e6e7943b8fbc86b71ccfd77656df4bf"} Mar 07 14:42:20 crc kubenswrapper[4943]: I0307 14:42:20.280324 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rvkgz" event={"ID":"4ae1cf01-1400-4500-9b84-8435a6541612","Type":"ContainerStarted","Data":"a1c0898bb09c451613fd5085b4e13b127657ce13de6cdaff591a3d96bbc8c225"} Mar 07 14:42:20 crc kubenswrapper[4943]: I0307 14:42:20.283101 4943 generic.go:334] "Generic (PLEG): container finished" podID="0d1657d7-a787-4648-91f6-e0ff987cede2" containerID="fac900a8e0187f803bf8032d0085fb713054b6666413f52381eaf48989c9cb80" exitCode=0 Mar 07 14:42:20 crc kubenswrapper[4943]: I0307 14:42:20.283127 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dfqzr" event={"ID":"0d1657d7-a787-4648-91f6-e0ff987cede2","Type":"ContainerDied","Data":"fac900a8e0187f803bf8032d0085fb713054b6666413f52381eaf48989c9cb80"} Mar 07 14:42:20 crc kubenswrapper[4943]: I0307 14:42:20.551448 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rvkgz" podStartSLOduration=4.236732365 podStartE2EDuration="51.551390155s" podCreationTimestamp="2026-03-07 14:41:29 +0000 UTC" firstStartedPulling="2026-03-07 14:41:31.488428996 +0000 UTC m=+133.440565494" lastFinishedPulling="2026-03-07 14:42:18.803086776 +0000 UTC m=+180.755223284" observedRunningTime="2026-03-07 14:42:20.343655811 +0000 UTC m=+182.295792339" watchObservedRunningTime="2026-03-07 14:42:20.551390155 +0000 UTC m=+182.503526693" Mar 07 14:42:20 crc kubenswrapper[4943]: I0307 14:42:20.554013 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9sf6k"] Mar 07 14:42:20 crc kubenswrapper[4943]: I0307 14:42:20.554380 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9sf6k" podUID="23fa83e7-df77-47e2-bb98-3344e48ab6bd" containerName="registry-server" containerID="cri-o://f2e01ac1b3b98ea5ad9485135e103568e4998789d36b202b8123967b665c26ca" gracePeriod=2 Mar 07 14:42:20 crc kubenswrapper[4943]: I0307 14:42:20.975245 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vkcxl" Mar 07 14:42:21 crc kubenswrapper[4943]: I0307 14:42:21.147680 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74a7b99c-3412-438e-a835-1e153a56047b-catalog-content\") pod \"74a7b99c-3412-438e-a835-1e153a56047b\" (UID: \"74a7b99c-3412-438e-a835-1e153a56047b\") " Mar 07 14:42:21 crc kubenswrapper[4943]: I0307 14:42:21.147960 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mq6g7\" (UniqueName: \"kubernetes.io/projected/74a7b99c-3412-438e-a835-1e153a56047b-kube-api-access-mq6g7\") pod \"74a7b99c-3412-438e-a835-1e153a56047b\" (UID: \"74a7b99c-3412-438e-a835-1e153a56047b\") " Mar 07 14:42:21 crc kubenswrapper[4943]: I0307 14:42:21.148030 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74a7b99c-3412-438e-a835-1e153a56047b-utilities\") pod \"74a7b99c-3412-438e-a835-1e153a56047b\" (UID: \"74a7b99c-3412-438e-a835-1e153a56047b\") " Mar 07 14:42:21 crc kubenswrapper[4943]: I0307 14:42:21.149220 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74a7b99c-3412-438e-a835-1e153a56047b-utilities" (OuterVolumeSpecName: "utilities") pod "74a7b99c-3412-438e-a835-1e153a56047b" (UID: "74a7b99c-3412-438e-a835-1e153a56047b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:42:21 crc kubenswrapper[4943]: I0307 14:42:21.159639 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74a7b99c-3412-438e-a835-1e153a56047b-kube-api-access-mq6g7" (OuterVolumeSpecName: "kube-api-access-mq6g7") pod "74a7b99c-3412-438e-a835-1e153a56047b" (UID: "74a7b99c-3412-438e-a835-1e153a56047b"). InnerVolumeSpecName "kube-api-access-mq6g7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:42:21 crc kubenswrapper[4943]: I0307 14:42:21.230432 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74a7b99c-3412-438e-a835-1e153a56047b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "74a7b99c-3412-438e-a835-1e153a56047b" (UID: "74a7b99c-3412-438e-a835-1e153a56047b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:42:21 crc kubenswrapper[4943]: I0307 14:42:21.250170 4943 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74a7b99c-3412-438e-a835-1e153a56047b-utilities\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:21 crc kubenswrapper[4943]: I0307 14:42:21.250206 4943 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74a7b99c-3412-438e-a835-1e153a56047b-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:21 crc kubenswrapper[4943]: I0307 14:42:21.250219 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mq6g7\" (UniqueName: \"kubernetes.io/projected/74a7b99c-3412-438e-a835-1e153a56047b-kube-api-access-mq6g7\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:21 crc kubenswrapper[4943]: I0307 14:42:21.290772 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vkcxl" event={"ID":"74a7b99c-3412-438e-a835-1e153a56047b","Type":"ContainerDied","Data":"3deec9910ce9875b285512643f55abea6acc1ca773058fdcc8e2d214d5bdbc21"} Mar 07 14:42:21 crc kubenswrapper[4943]: I0307 14:42:21.290828 4943 scope.go:117] "RemoveContainer" containerID="d5b0b520a7e836b5a5d4d0c817a049499e6e7943b8fbc86b71ccfd77656df4bf" Mar 07 14:42:21 crc kubenswrapper[4943]: I0307 14:42:21.290947 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vkcxl" Mar 07 14:42:21 crc kubenswrapper[4943]: I0307 14:42:21.297128 4943 generic.go:334] "Generic (PLEG): container finished" podID="23fa83e7-df77-47e2-bb98-3344e48ab6bd" containerID="f2e01ac1b3b98ea5ad9485135e103568e4998789d36b202b8123967b665c26ca" exitCode=0 Mar 07 14:42:21 crc kubenswrapper[4943]: I0307 14:42:21.297159 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9sf6k" event={"ID":"23fa83e7-df77-47e2-bb98-3344e48ab6bd","Type":"ContainerDied","Data":"f2e01ac1b3b98ea5ad9485135e103568e4998789d36b202b8123967b665c26ca"} Mar 07 14:42:21 crc kubenswrapper[4943]: I0307 14:42:21.324064 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vkcxl"] Mar 07 14:42:21 crc kubenswrapper[4943]: I0307 14:42:21.330006 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-vkcxl"] Mar 07 14:42:21 crc kubenswrapper[4943]: I0307 14:42:21.893710 4943 scope.go:117] "RemoveContainer" containerID="ad78b2587ac7c0ce5bc3a0ea769519e07e704a8ecaaf863e3c02fc49d954d63a" Mar 07 14:42:22 crc kubenswrapper[4943]: I0307 14:42:22.248385 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9sf6k" Mar 07 14:42:22 crc kubenswrapper[4943]: I0307 14:42:22.311427 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9sf6k" event={"ID":"23fa83e7-df77-47e2-bb98-3344e48ab6bd","Type":"ContainerDied","Data":"9139f60cef576a3909eb49dc2cb92a68a6bb168e3dc55e778d8644ab62cd64f2"} Mar 07 14:42:22 crc kubenswrapper[4943]: I0307 14:42:22.311488 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9sf6k" Mar 07 14:42:22 crc kubenswrapper[4943]: I0307 14:42:22.364440 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23fa83e7-df77-47e2-bb98-3344e48ab6bd-catalog-content\") pod \"23fa83e7-df77-47e2-bb98-3344e48ab6bd\" (UID: \"23fa83e7-df77-47e2-bb98-3344e48ab6bd\") " Mar 07 14:42:22 crc kubenswrapper[4943]: I0307 14:42:22.364490 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23fa83e7-df77-47e2-bb98-3344e48ab6bd-utilities\") pod \"23fa83e7-df77-47e2-bb98-3344e48ab6bd\" (UID: \"23fa83e7-df77-47e2-bb98-3344e48ab6bd\") " Mar 07 14:42:22 crc kubenswrapper[4943]: I0307 14:42:22.364573 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f8qh5\" (UniqueName: \"kubernetes.io/projected/23fa83e7-df77-47e2-bb98-3344e48ab6bd-kube-api-access-f8qh5\") pod \"23fa83e7-df77-47e2-bb98-3344e48ab6bd\" (UID: \"23fa83e7-df77-47e2-bb98-3344e48ab6bd\") " Mar 07 14:42:22 crc kubenswrapper[4943]: I0307 14:42:22.366091 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23fa83e7-df77-47e2-bb98-3344e48ab6bd-utilities" (OuterVolumeSpecName: "utilities") pod "23fa83e7-df77-47e2-bb98-3344e48ab6bd" (UID: "23fa83e7-df77-47e2-bb98-3344e48ab6bd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:42:22 crc kubenswrapper[4943]: I0307 14:42:22.370678 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23fa83e7-df77-47e2-bb98-3344e48ab6bd-kube-api-access-f8qh5" (OuterVolumeSpecName: "kube-api-access-f8qh5") pod "23fa83e7-df77-47e2-bb98-3344e48ab6bd" (UID: "23fa83e7-df77-47e2-bb98-3344e48ab6bd"). InnerVolumeSpecName "kube-api-access-f8qh5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:42:22 crc kubenswrapper[4943]: I0307 14:42:22.450140 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23fa83e7-df77-47e2-bb98-3344e48ab6bd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "23fa83e7-df77-47e2-bb98-3344e48ab6bd" (UID: "23fa83e7-df77-47e2-bb98-3344e48ab6bd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:42:22 crc kubenswrapper[4943]: I0307 14:42:22.466052 4943 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23fa83e7-df77-47e2-bb98-3344e48ab6bd-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:22 crc kubenswrapper[4943]: I0307 14:42:22.466074 4943 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23fa83e7-df77-47e2-bb98-3344e48ab6bd-utilities\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:22 crc kubenswrapper[4943]: I0307 14:42:22.466111 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f8qh5\" (UniqueName: \"kubernetes.io/projected/23fa83e7-df77-47e2-bb98-3344e48ab6bd-kube-api-access-f8qh5\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:22 crc kubenswrapper[4943]: I0307 14:42:22.499764 4943 scope.go:117] "RemoveContainer" containerID="b693f2dfa58f9dd4c4a864a088610221ba9a30f644ee4c4624a68a73422f8056" Mar 07 14:42:22 crc kubenswrapper[4943]: I0307 14:42:22.514214 4943 scope.go:117] "RemoveContainer" containerID="f2e01ac1b3b98ea5ad9485135e103568e4998789d36b202b8123967b665c26ca" Mar 07 14:42:22 crc kubenswrapper[4943]: I0307 14:42:22.537941 4943 scope.go:117] "RemoveContainer" containerID="0b7ee43e866c549159478b2d441e21f9c48f78bb88ed5e53a91b6512934df95c" Mar 07 14:42:22 crc kubenswrapper[4943]: I0307 14:42:22.564896 4943 scope.go:117] "RemoveContainer" containerID="7d1b2a44c52125e4bda1275594a9ef1849cbe7f97f88ce6b456365004f8a327d" Mar 07 14:42:22 crc kubenswrapper[4943]: I0307 14:42:22.658170 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9sf6k"] Mar 07 14:42:22 crc kubenswrapper[4943]: I0307 14:42:22.663662 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9sf6k"] Mar 07 14:42:22 crc kubenswrapper[4943]: I0307 14:42:22.767757 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23fa83e7-df77-47e2-bb98-3344e48ab6bd" path="/var/lib/kubelet/pods/23fa83e7-df77-47e2-bb98-3344e48ab6bd/volumes" Mar 07 14:42:22 crc kubenswrapper[4943]: I0307 14:42:22.769027 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74a7b99c-3412-438e-a835-1e153a56047b" path="/var/lib/kubelet/pods/74a7b99c-3412-438e-a835-1e153a56047b/volumes" Mar 07 14:42:23 crc kubenswrapper[4943]: I0307 14:42:23.322828 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p767n" event={"ID":"6a05c7ae-f2a1-4116-9038-ac838fa88af7","Type":"ContainerStarted","Data":"195c1f3f178736421b8271566e1f002a2c5619c4109019c84bfd76a1c3e498a6"} Mar 07 14:42:23 crc kubenswrapper[4943]: I0307 14:42:23.329269 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dfqzr" event={"ID":"0d1657d7-a787-4648-91f6-e0ff987cede2","Type":"ContainerStarted","Data":"2a6e8e90c36b929a9f6b254045d721389004a538bf9b2b26618515396b8cecda"} Mar 07 14:42:23 crc kubenswrapper[4943]: I0307 14:42:23.352553 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-p767n" podStartSLOduration=5.395954942 podStartE2EDuration="56.352537711s" podCreationTimestamp="2026-03-07 14:41:27 +0000 UTC" firstStartedPulling="2026-03-07 14:41:31.543356751 +0000 UTC m=+133.495493239" lastFinishedPulling="2026-03-07 14:42:22.4999395 +0000 UTC m=+184.452076008" observedRunningTime="2026-03-07 14:42:23.350980503 +0000 UTC m=+185.303116991" watchObservedRunningTime="2026-03-07 14:42:23.352537711 +0000 UTC m=+185.304674209" Mar 07 14:42:23 crc kubenswrapper[4943]: I0307 14:42:23.382901 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dfqzr" podStartSLOduration=3.615123972 podStartE2EDuration="58.382881442s" podCreationTimestamp="2026-03-07 14:41:25 +0000 UTC" firstStartedPulling="2026-03-07 14:41:27.986709352 +0000 UTC m=+129.938845850" lastFinishedPulling="2026-03-07 14:42:22.754466792 +0000 UTC m=+184.706603320" observedRunningTime="2026-03-07 14:42:23.381283274 +0000 UTC m=+185.333419782" watchObservedRunningTime="2026-03-07 14:42:23.382881442 +0000 UTC m=+185.335017940" Mar 07 14:42:26 crc kubenswrapper[4943]: I0307 14:42:26.035002 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dfqzr" Mar 07 14:42:26 crc kubenswrapper[4943]: I0307 14:42:26.037617 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dfqzr" Mar 07 14:42:26 crc kubenswrapper[4943]: I0307 14:42:26.123835 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dfqzr" Mar 07 14:42:28 crc kubenswrapper[4943]: I0307 14:42:28.570872 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-p767n" Mar 07 14:42:28 crc kubenswrapper[4943]: I0307 14:42:28.571260 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-p767n" Mar 07 14:42:28 crc kubenswrapper[4943]: I0307 14:42:28.606720 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-p767n" Mar 07 14:42:29 crc kubenswrapper[4943]: I0307 14:42:29.039120 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-f957fc5cb-8r7lr"] Mar 07 14:42:29 crc kubenswrapper[4943]: I0307 14:42:29.039444 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-f957fc5cb-8r7lr" podUID="6758e5ec-0b51-41ef-863d-aa72f41e6bb7" containerName="controller-manager" containerID="cri-o://72b50585ad9e852d9b58596507515bc87cac3e6c9e04731e4dc63a89ae20103c" gracePeriod=30 Mar 07 14:42:29 crc kubenswrapper[4943]: I0307 14:42:29.064019 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-97747b5c5-wkxmj"] Mar 07 14:42:29 crc kubenswrapper[4943]: I0307 14:42:29.064584 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-97747b5c5-wkxmj" podUID="4b9e1f64-5c9d-405c-820c-762a9ca4798a" containerName="route-controller-manager" containerID="cri-o://55524a95d9cfc406b363ac3a85ba9c341ea971eb1b3979f5c12d750721660263" gracePeriod=30 Mar 07 14:42:29 crc kubenswrapper[4943]: I0307 14:42:29.196321 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-f95bw" Mar 07 14:42:29 crc kubenswrapper[4943]: I0307 14:42:29.257613 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-f95bw" Mar 07 14:42:29 crc kubenswrapper[4943]: I0307 14:42:29.398288 4943 generic.go:334] "Generic (PLEG): container finished" podID="6758e5ec-0b51-41ef-863d-aa72f41e6bb7" containerID="72b50585ad9e852d9b58596507515bc87cac3e6c9e04731e4dc63a89ae20103c" exitCode=0 Mar 07 14:42:29 crc kubenswrapper[4943]: I0307 14:42:29.398725 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-f957fc5cb-8r7lr" event={"ID":"6758e5ec-0b51-41ef-863d-aa72f41e6bb7","Type":"ContainerDied","Data":"72b50585ad9e852d9b58596507515bc87cac3e6c9e04731e4dc63a89ae20103c"} Mar 07 14:42:29 crc kubenswrapper[4943]: I0307 14:42:29.401450 4943 generic.go:334] "Generic (PLEG): container finished" podID="4b9e1f64-5c9d-405c-820c-762a9ca4798a" containerID="55524a95d9cfc406b363ac3a85ba9c341ea971eb1b3979f5c12d750721660263" exitCode=0 Mar 07 14:42:29 crc kubenswrapper[4943]: I0307 14:42:29.402317 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-97747b5c5-wkxmj" event={"ID":"4b9e1f64-5c9d-405c-820c-762a9ca4798a","Type":"ContainerDied","Data":"55524a95d9cfc406b363ac3a85ba9c341ea971eb1b3979f5c12d750721660263"} Mar 07 14:42:29 crc kubenswrapper[4943]: I0307 14:42:29.460149 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-p767n" Mar 07 14:42:29 crc kubenswrapper[4943]: I0307 14:42:29.579499 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-97747b5c5-wkxmj" Mar 07 14:42:29 crc kubenswrapper[4943]: I0307 14:42:29.664575 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4b9e1f64-5c9d-405c-820c-762a9ca4798a-client-ca\") pod \"4b9e1f64-5c9d-405c-820c-762a9ca4798a\" (UID: \"4b9e1f64-5c9d-405c-820c-762a9ca4798a\") " Mar 07 14:42:29 crc kubenswrapper[4943]: I0307 14:42:29.664628 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b9e1f64-5c9d-405c-820c-762a9ca4798a-config\") pod \"4b9e1f64-5c9d-405c-820c-762a9ca4798a\" (UID: \"4b9e1f64-5c9d-405c-820c-762a9ca4798a\") " Mar 07 14:42:29 crc kubenswrapper[4943]: I0307 14:42:29.664708 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b9e1f64-5c9d-405c-820c-762a9ca4798a-serving-cert\") pod \"4b9e1f64-5c9d-405c-820c-762a9ca4798a\" (UID: \"4b9e1f64-5c9d-405c-820c-762a9ca4798a\") " Mar 07 14:42:29 crc kubenswrapper[4943]: I0307 14:42:29.664733 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kvv85\" (UniqueName: \"kubernetes.io/projected/4b9e1f64-5c9d-405c-820c-762a9ca4798a-kube-api-access-kvv85\") pod \"4b9e1f64-5c9d-405c-820c-762a9ca4798a\" (UID: \"4b9e1f64-5c9d-405c-820c-762a9ca4798a\") " Mar 07 14:42:29 crc kubenswrapper[4943]: I0307 14:42:29.665606 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b9e1f64-5c9d-405c-820c-762a9ca4798a-client-ca" (OuterVolumeSpecName: "client-ca") pod "4b9e1f64-5c9d-405c-820c-762a9ca4798a" (UID: "4b9e1f64-5c9d-405c-820c-762a9ca4798a"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:42:29 crc kubenswrapper[4943]: I0307 14:42:29.665883 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b9e1f64-5c9d-405c-820c-762a9ca4798a-config" (OuterVolumeSpecName: "config") pod "4b9e1f64-5c9d-405c-820c-762a9ca4798a" (UID: "4b9e1f64-5c9d-405c-820c-762a9ca4798a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:42:29 crc kubenswrapper[4943]: I0307 14:42:29.674272 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b9e1f64-5c9d-405c-820c-762a9ca4798a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "4b9e1f64-5c9d-405c-820c-762a9ca4798a" (UID: "4b9e1f64-5c9d-405c-820c-762a9ca4798a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:42:29 crc kubenswrapper[4943]: I0307 14:42:29.676628 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b9e1f64-5c9d-405c-820c-762a9ca4798a-kube-api-access-kvv85" (OuterVolumeSpecName: "kube-api-access-kvv85") pod "4b9e1f64-5c9d-405c-820c-762a9ca4798a" (UID: "4b9e1f64-5c9d-405c-820c-762a9ca4798a"). InnerVolumeSpecName "kube-api-access-kvv85". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:42:29 crc kubenswrapper[4943]: I0307 14:42:29.724749 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-f957fc5cb-8r7lr" Mar 07 14:42:29 crc kubenswrapper[4943]: I0307 14:42:29.766669 4943 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b9e1f64-5c9d-405c-820c-762a9ca4798a-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:29 crc kubenswrapper[4943]: I0307 14:42:29.766715 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kvv85\" (UniqueName: \"kubernetes.io/projected/4b9e1f64-5c9d-405c-820c-762a9ca4798a-kube-api-access-kvv85\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:29 crc kubenswrapper[4943]: I0307 14:42:29.766730 4943 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4b9e1f64-5c9d-405c-820c-762a9ca4798a-client-ca\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:29 crc kubenswrapper[4943]: I0307 14:42:29.766744 4943 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b9e1f64-5c9d-405c-820c-762a9ca4798a-config\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:29 crc kubenswrapper[4943]: I0307 14:42:29.817709 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rvkgz" Mar 07 14:42:29 crc kubenswrapper[4943]: I0307 14:42:29.817782 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rvkgz" Mar 07 14:42:29 crc kubenswrapper[4943]: I0307 14:42:29.867971 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6758e5ec-0b51-41ef-863d-aa72f41e6bb7-proxy-ca-bundles\") pod \"6758e5ec-0b51-41ef-863d-aa72f41e6bb7\" (UID: \"6758e5ec-0b51-41ef-863d-aa72f41e6bb7\") " Mar 07 14:42:29 crc kubenswrapper[4943]: I0307 14:42:29.868025 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6758e5ec-0b51-41ef-863d-aa72f41e6bb7-serving-cert\") pod \"6758e5ec-0b51-41ef-863d-aa72f41e6bb7\" (UID: \"6758e5ec-0b51-41ef-863d-aa72f41e6bb7\") " Mar 07 14:42:29 crc kubenswrapper[4943]: I0307 14:42:29.868086 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6758e5ec-0b51-41ef-863d-aa72f41e6bb7-client-ca\") pod \"6758e5ec-0b51-41ef-863d-aa72f41e6bb7\" (UID: \"6758e5ec-0b51-41ef-863d-aa72f41e6bb7\") " Mar 07 14:42:29 crc kubenswrapper[4943]: I0307 14:42:29.868105 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6758e5ec-0b51-41ef-863d-aa72f41e6bb7-config\") pod \"6758e5ec-0b51-41ef-863d-aa72f41e6bb7\" (UID: \"6758e5ec-0b51-41ef-863d-aa72f41e6bb7\") " Mar 07 14:42:29 crc kubenswrapper[4943]: I0307 14:42:29.868176 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hj2rp\" (UniqueName: \"kubernetes.io/projected/6758e5ec-0b51-41ef-863d-aa72f41e6bb7-kube-api-access-hj2rp\") pod \"6758e5ec-0b51-41ef-863d-aa72f41e6bb7\" (UID: \"6758e5ec-0b51-41ef-863d-aa72f41e6bb7\") " Mar 07 14:42:29 crc kubenswrapper[4943]: I0307 14:42:29.868784 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6758e5ec-0b51-41ef-863d-aa72f41e6bb7-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "6758e5ec-0b51-41ef-863d-aa72f41e6bb7" (UID: "6758e5ec-0b51-41ef-863d-aa72f41e6bb7"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:42:29 crc kubenswrapper[4943]: I0307 14:42:29.868955 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6758e5ec-0b51-41ef-863d-aa72f41e6bb7-config" (OuterVolumeSpecName: "config") pod "6758e5ec-0b51-41ef-863d-aa72f41e6bb7" (UID: "6758e5ec-0b51-41ef-863d-aa72f41e6bb7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:42:29 crc kubenswrapper[4943]: I0307 14:42:29.868956 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6758e5ec-0b51-41ef-863d-aa72f41e6bb7-client-ca" (OuterVolumeSpecName: "client-ca") pod "6758e5ec-0b51-41ef-863d-aa72f41e6bb7" (UID: "6758e5ec-0b51-41ef-863d-aa72f41e6bb7"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:42:29 crc kubenswrapper[4943]: I0307 14:42:29.870874 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6758e5ec-0b51-41ef-863d-aa72f41e6bb7-kube-api-access-hj2rp" (OuterVolumeSpecName: "kube-api-access-hj2rp") pod "6758e5ec-0b51-41ef-863d-aa72f41e6bb7" (UID: "6758e5ec-0b51-41ef-863d-aa72f41e6bb7"). InnerVolumeSpecName "kube-api-access-hj2rp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:42:29 crc kubenswrapper[4943]: I0307 14:42:29.870978 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6758e5ec-0b51-41ef-863d-aa72f41e6bb7-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6758e5ec-0b51-41ef-863d-aa72f41e6bb7" (UID: "6758e5ec-0b51-41ef-863d-aa72f41e6bb7"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:42:29 crc kubenswrapper[4943]: I0307 14:42:29.871094 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rvkgz" Mar 07 14:42:29 crc kubenswrapper[4943]: I0307 14:42:29.969602 4943 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6758e5ec-0b51-41ef-863d-aa72f41e6bb7-client-ca\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:29 crc kubenswrapper[4943]: I0307 14:42:29.969638 4943 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6758e5ec-0b51-41ef-863d-aa72f41e6bb7-config\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:29 crc kubenswrapper[4943]: I0307 14:42:29.969649 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hj2rp\" (UniqueName: \"kubernetes.io/projected/6758e5ec-0b51-41ef-863d-aa72f41e6bb7-kube-api-access-hj2rp\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:29 crc kubenswrapper[4943]: I0307 14:42:29.969661 4943 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6758e5ec-0b51-41ef-863d-aa72f41e6bb7-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:29 crc kubenswrapper[4943]: I0307 14:42:29.969671 4943 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6758e5ec-0b51-41ef-863d-aa72f41e6bb7-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:30 crc kubenswrapper[4943]: I0307 14:42:30.426028 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-f957fc5cb-8r7lr" event={"ID":"6758e5ec-0b51-41ef-863d-aa72f41e6bb7","Type":"ContainerDied","Data":"ad473c1eafdda09c6b7a142f4bdf574574b80c7fbf9fab92a4433393e9d7971d"} Mar 07 14:42:30 crc kubenswrapper[4943]: I0307 14:42:30.426169 4943 scope.go:117] "RemoveContainer" containerID="72b50585ad9e852d9b58596507515bc87cac3e6c9e04731e4dc63a89ae20103c" Mar 07 14:42:30 crc kubenswrapper[4943]: I0307 14:42:30.426198 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-f957fc5cb-8r7lr" Mar 07 14:42:30 crc kubenswrapper[4943]: I0307 14:42:30.430383 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-97747b5c5-wkxmj" event={"ID":"4b9e1f64-5c9d-405c-820c-762a9ca4798a","Type":"ContainerDied","Data":"34940292bdcaa5d49d8fccfd644e02091897a91cb218f557eae574d10251ad55"} Mar 07 14:42:30 crc kubenswrapper[4943]: I0307 14:42:30.430477 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-97747b5c5-wkxmj" Mar 07 14:42:30 crc kubenswrapper[4943]: I0307 14:42:30.456197 4943 scope.go:117] "RemoveContainer" containerID="55524a95d9cfc406b363ac3a85ba9c341ea971eb1b3979f5c12d750721660263" Mar 07 14:42:30 crc kubenswrapper[4943]: I0307 14:42:30.490204 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-97747b5c5-wkxmj"] Mar 07 14:42:30 crc kubenswrapper[4943]: I0307 14:42:30.495482 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-97747b5c5-wkxmj"] Mar 07 14:42:30 crc kubenswrapper[4943]: I0307 14:42:30.508692 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-f957fc5cb-8r7lr"] Mar 07 14:42:30 crc kubenswrapper[4943]: I0307 14:42:30.514124 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-f957fc5cb-8r7lr"] Mar 07 14:42:30 crc kubenswrapper[4943]: I0307 14:42:30.516827 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rvkgz" Mar 07 14:42:30 crc kubenswrapper[4943]: I0307 14:42:30.546087 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-p767n"] Mar 07 14:42:30 crc kubenswrapper[4943]: I0307 14:42:30.767893 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b9e1f64-5c9d-405c-820c-762a9ca4798a" path="/var/lib/kubelet/pods/4b9e1f64-5c9d-405c-820c-762a9ca4798a/volumes" Mar 07 14:42:30 crc kubenswrapper[4943]: I0307 14:42:30.768984 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6758e5ec-0b51-41ef-863d-aa72f41e6bb7" path="/var/lib/kubelet/pods/6758e5ec-0b51-41ef-863d-aa72f41e6bb7/volumes" Mar 07 14:42:30 crc kubenswrapper[4943]: I0307 14:42:30.815172 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-65d6c4859f-tqqv5"] Mar 07 14:42:30 crc kubenswrapper[4943]: E0307 14:42:30.815640 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23fa83e7-df77-47e2-bb98-3344e48ab6bd" containerName="extract-content" Mar 07 14:42:30 crc kubenswrapper[4943]: I0307 14:42:30.815675 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="23fa83e7-df77-47e2-bb98-3344e48ab6bd" containerName="extract-content" Mar 07 14:42:30 crc kubenswrapper[4943]: E0307 14:42:30.815701 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74a7b99c-3412-438e-a835-1e153a56047b" containerName="registry-server" Mar 07 14:42:30 crc kubenswrapper[4943]: I0307 14:42:30.815718 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="74a7b99c-3412-438e-a835-1e153a56047b" containerName="registry-server" Mar 07 14:42:30 crc kubenswrapper[4943]: E0307 14:42:30.815776 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6758e5ec-0b51-41ef-863d-aa72f41e6bb7" containerName="controller-manager" Mar 07 14:42:30 crc kubenswrapper[4943]: I0307 14:42:30.815791 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="6758e5ec-0b51-41ef-863d-aa72f41e6bb7" containerName="controller-manager" Mar 07 14:42:30 crc kubenswrapper[4943]: E0307 14:42:30.815878 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b9e1f64-5c9d-405c-820c-762a9ca4798a" containerName="route-controller-manager" Mar 07 14:42:30 crc kubenswrapper[4943]: I0307 14:42:30.815893 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b9e1f64-5c9d-405c-820c-762a9ca4798a" containerName="route-controller-manager" Mar 07 14:42:30 crc kubenswrapper[4943]: E0307 14:42:30.815961 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23fa83e7-df77-47e2-bb98-3344e48ab6bd" containerName="registry-server" Mar 07 14:42:30 crc kubenswrapper[4943]: I0307 14:42:30.815975 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="23fa83e7-df77-47e2-bb98-3344e48ab6bd" containerName="registry-server" Mar 07 14:42:30 crc kubenswrapper[4943]: E0307 14:42:30.816059 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23fa83e7-df77-47e2-bb98-3344e48ab6bd" containerName="extract-utilities" Mar 07 14:42:30 crc kubenswrapper[4943]: I0307 14:42:30.816073 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="23fa83e7-df77-47e2-bb98-3344e48ab6bd" containerName="extract-utilities" Mar 07 14:42:30 crc kubenswrapper[4943]: E0307 14:42:30.816088 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74a7b99c-3412-438e-a835-1e153a56047b" containerName="extract-content" Mar 07 14:42:30 crc kubenswrapper[4943]: I0307 14:42:30.816955 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="74a7b99c-3412-438e-a835-1e153a56047b" containerName="extract-content" Mar 07 14:42:30 crc kubenswrapper[4943]: E0307 14:42:30.816980 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74a7b99c-3412-438e-a835-1e153a56047b" containerName="extract-utilities" Mar 07 14:42:30 crc kubenswrapper[4943]: I0307 14:42:30.816998 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="74a7b99c-3412-438e-a835-1e153a56047b" containerName="extract-utilities" Mar 07 14:42:30 crc kubenswrapper[4943]: I0307 14:42:30.817714 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b9e1f64-5c9d-405c-820c-762a9ca4798a" containerName="route-controller-manager" Mar 07 14:42:30 crc kubenswrapper[4943]: I0307 14:42:30.817858 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="74a7b99c-3412-438e-a835-1e153a56047b" containerName="registry-server" Mar 07 14:42:30 crc kubenswrapper[4943]: I0307 14:42:30.817896 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="6758e5ec-0b51-41ef-863d-aa72f41e6bb7" containerName="controller-manager" Mar 07 14:42:30 crc kubenswrapper[4943]: I0307 14:42:30.817913 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="23fa83e7-df77-47e2-bb98-3344e48ab6bd" containerName="registry-server" Mar 07 14:42:30 crc kubenswrapper[4943]: I0307 14:42:30.821361 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-65d6c4859f-tqqv5" Mar 07 14:42:30 crc kubenswrapper[4943]: I0307 14:42:30.824206 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-574f768f98-q28jc"] Mar 07 14:42:30 crc kubenswrapper[4943]: I0307 14:42:30.826031 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-574f768f98-q28jc" Mar 07 14:42:30 crc kubenswrapper[4943]: I0307 14:42:30.831242 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 07 14:42:30 crc kubenswrapper[4943]: I0307 14:42:30.831284 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 07 14:42:30 crc kubenswrapper[4943]: I0307 14:42:30.831741 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 07 14:42:30 crc kubenswrapper[4943]: I0307 14:42:30.831852 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 07 14:42:30 crc kubenswrapper[4943]: I0307 14:42:30.831879 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 07 14:42:30 crc kubenswrapper[4943]: I0307 14:42:30.832019 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 07 14:42:30 crc kubenswrapper[4943]: I0307 14:42:30.832898 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 07 14:42:30 crc kubenswrapper[4943]: I0307 14:42:30.833009 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 07 14:42:30 crc kubenswrapper[4943]: I0307 14:42:30.833348 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 07 14:42:30 crc kubenswrapper[4943]: I0307 14:42:30.833379 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 07 14:42:30 crc kubenswrapper[4943]: I0307 14:42:30.833536 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 07 14:42:30 crc kubenswrapper[4943]: I0307 14:42:30.834258 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 07 14:42:30 crc kubenswrapper[4943]: I0307 14:42:30.838088 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 07 14:42:30 crc kubenswrapper[4943]: I0307 14:42:30.840695 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-65d6c4859f-tqqv5"] Mar 07 14:42:30 crc kubenswrapper[4943]: I0307 14:42:30.847469 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-574f768f98-q28jc"] Mar 07 14:42:30 crc kubenswrapper[4943]: I0307 14:42:30.986735 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfpc8\" (UniqueName: \"kubernetes.io/projected/2225553d-99ab-4be0-8515-190337057bac-kube-api-access-cfpc8\") pod \"controller-manager-65d6c4859f-tqqv5\" (UID: \"2225553d-99ab-4be0-8515-190337057bac\") " pod="openshift-controller-manager/controller-manager-65d6c4859f-tqqv5" Mar 07 14:42:30 crc kubenswrapper[4943]: I0307 14:42:30.986807 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2225553d-99ab-4be0-8515-190337057bac-serving-cert\") pod \"controller-manager-65d6c4859f-tqqv5\" (UID: \"2225553d-99ab-4be0-8515-190337057bac\") " pod="openshift-controller-manager/controller-manager-65d6c4859f-tqqv5" Mar 07 14:42:30 crc kubenswrapper[4943]: I0307 14:42:30.986867 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2225553d-99ab-4be0-8515-190337057bac-config\") pod \"controller-manager-65d6c4859f-tqqv5\" (UID: \"2225553d-99ab-4be0-8515-190337057bac\") " pod="openshift-controller-manager/controller-manager-65d6c4859f-tqqv5" Mar 07 14:42:30 crc kubenswrapper[4943]: I0307 14:42:30.987064 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a-serving-cert\") pod \"route-controller-manager-574f768f98-q28jc\" (UID: \"b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a\") " pod="openshift-route-controller-manager/route-controller-manager-574f768f98-q28jc" Mar 07 14:42:30 crc kubenswrapper[4943]: I0307 14:42:30.987175 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2225553d-99ab-4be0-8515-190337057bac-proxy-ca-bundles\") pod \"controller-manager-65d6c4859f-tqqv5\" (UID: \"2225553d-99ab-4be0-8515-190337057bac\") " pod="openshift-controller-manager/controller-manager-65d6c4859f-tqqv5" Mar 07 14:42:30 crc kubenswrapper[4943]: I0307 14:42:30.987260 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a-client-ca\") pod \"route-controller-manager-574f768f98-q28jc\" (UID: \"b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a\") " pod="openshift-route-controller-manager/route-controller-manager-574f768f98-q28jc" Mar 07 14:42:30 crc kubenswrapper[4943]: I0307 14:42:30.987316 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxkrz\" (UniqueName: \"kubernetes.io/projected/b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a-kube-api-access-mxkrz\") pod \"route-controller-manager-574f768f98-q28jc\" (UID: \"b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a\") " pod="openshift-route-controller-manager/route-controller-manager-574f768f98-q28jc" Mar 07 14:42:30 crc kubenswrapper[4943]: I0307 14:42:30.987380 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a-config\") pod \"route-controller-manager-574f768f98-q28jc\" (UID: \"b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a\") " pod="openshift-route-controller-manager/route-controller-manager-574f768f98-q28jc" Mar 07 14:42:30 crc kubenswrapper[4943]: I0307 14:42:30.987465 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2225553d-99ab-4be0-8515-190337057bac-client-ca\") pod \"controller-manager-65d6c4859f-tqqv5\" (UID: \"2225553d-99ab-4be0-8515-190337057bac\") " pod="openshift-controller-manager/controller-manager-65d6c4859f-tqqv5" Mar 07 14:42:31 crc kubenswrapper[4943]: I0307 14:42:31.088350 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2225553d-99ab-4be0-8515-190337057bac-client-ca\") pod \"controller-manager-65d6c4859f-tqqv5\" (UID: \"2225553d-99ab-4be0-8515-190337057bac\") " pod="openshift-controller-manager/controller-manager-65d6c4859f-tqqv5" Mar 07 14:42:31 crc kubenswrapper[4943]: I0307 14:42:31.088467 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfpc8\" (UniqueName: \"kubernetes.io/projected/2225553d-99ab-4be0-8515-190337057bac-kube-api-access-cfpc8\") pod \"controller-manager-65d6c4859f-tqqv5\" (UID: \"2225553d-99ab-4be0-8515-190337057bac\") " pod="openshift-controller-manager/controller-manager-65d6c4859f-tqqv5" Mar 07 14:42:31 crc kubenswrapper[4943]: I0307 14:42:31.088507 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2225553d-99ab-4be0-8515-190337057bac-serving-cert\") pod \"controller-manager-65d6c4859f-tqqv5\" (UID: \"2225553d-99ab-4be0-8515-190337057bac\") " pod="openshift-controller-manager/controller-manager-65d6c4859f-tqqv5" Mar 07 14:42:31 crc kubenswrapper[4943]: I0307 14:42:31.088561 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2225553d-99ab-4be0-8515-190337057bac-config\") pod \"controller-manager-65d6c4859f-tqqv5\" (UID: \"2225553d-99ab-4be0-8515-190337057bac\") " pod="openshift-controller-manager/controller-manager-65d6c4859f-tqqv5" Mar 07 14:42:31 crc kubenswrapper[4943]: I0307 14:42:31.088630 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a-serving-cert\") pod \"route-controller-manager-574f768f98-q28jc\" (UID: \"b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a\") " pod="openshift-route-controller-manager/route-controller-manager-574f768f98-q28jc" Mar 07 14:42:31 crc kubenswrapper[4943]: I0307 14:42:31.088665 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2225553d-99ab-4be0-8515-190337057bac-proxy-ca-bundles\") pod \"controller-manager-65d6c4859f-tqqv5\" (UID: \"2225553d-99ab-4be0-8515-190337057bac\") " pod="openshift-controller-manager/controller-manager-65d6c4859f-tqqv5" Mar 07 14:42:31 crc kubenswrapper[4943]: I0307 14:42:31.088710 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a-client-ca\") pod \"route-controller-manager-574f768f98-q28jc\" (UID: \"b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a\") " pod="openshift-route-controller-manager/route-controller-manager-574f768f98-q28jc" Mar 07 14:42:31 crc kubenswrapper[4943]: I0307 14:42:31.088747 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxkrz\" (UniqueName: \"kubernetes.io/projected/b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a-kube-api-access-mxkrz\") pod \"route-controller-manager-574f768f98-q28jc\" (UID: \"b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a\") " pod="openshift-route-controller-manager/route-controller-manager-574f768f98-q28jc" Mar 07 14:42:31 crc kubenswrapper[4943]: I0307 14:42:31.088792 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a-config\") pod \"route-controller-manager-574f768f98-q28jc\" (UID: \"b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a\") " pod="openshift-route-controller-manager/route-controller-manager-574f768f98-q28jc" Mar 07 14:42:31 crc kubenswrapper[4943]: I0307 14:42:31.090355 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a-client-ca\") pod \"route-controller-manager-574f768f98-q28jc\" (UID: \"b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a\") " pod="openshift-route-controller-manager/route-controller-manager-574f768f98-q28jc" Mar 07 14:42:31 crc kubenswrapper[4943]: I0307 14:42:31.090473 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2225553d-99ab-4be0-8515-190337057bac-client-ca\") pod \"controller-manager-65d6c4859f-tqqv5\" (UID: \"2225553d-99ab-4be0-8515-190337057bac\") " pod="openshift-controller-manager/controller-manager-65d6c4859f-tqqv5" Mar 07 14:42:31 crc kubenswrapper[4943]: I0307 14:42:31.091043 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a-config\") pod \"route-controller-manager-574f768f98-q28jc\" (UID: \"b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a\") " pod="openshift-route-controller-manager/route-controller-manager-574f768f98-q28jc" Mar 07 14:42:31 crc kubenswrapper[4943]: I0307 14:42:31.091449 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2225553d-99ab-4be0-8515-190337057bac-config\") pod \"controller-manager-65d6c4859f-tqqv5\" (UID: \"2225553d-99ab-4be0-8515-190337057bac\") " pod="openshift-controller-manager/controller-manager-65d6c4859f-tqqv5" Mar 07 14:42:31 crc kubenswrapper[4943]: I0307 14:42:31.092221 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2225553d-99ab-4be0-8515-190337057bac-proxy-ca-bundles\") pod \"controller-manager-65d6c4859f-tqqv5\" (UID: \"2225553d-99ab-4be0-8515-190337057bac\") " pod="openshift-controller-manager/controller-manager-65d6c4859f-tqqv5" Mar 07 14:42:31 crc kubenswrapper[4943]: I0307 14:42:31.095128 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a-serving-cert\") pod \"route-controller-manager-574f768f98-q28jc\" (UID: \"b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a\") " pod="openshift-route-controller-manager/route-controller-manager-574f768f98-q28jc" Mar 07 14:42:31 crc kubenswrapper[4943]: I0307 14:42:31.098089 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2225553d-99ab-4be0-8515-190337057bac-serving-cert\") pod \"controller-manager-65d6c4859f-tqqv5\" (UID: \"2225553d-99ab-4be0-8515-190337057bac\") " pod="openshift-controller-manager/controller-manager-65d6c4859f-tqqv5" Mar 07 14:42:31 crc kubenswrapper[4943]: I0307 14:42:31.119319 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfpc8\" (UniqueName: \"kubernetes.io/projected/2225553d-99ab-4be0-8515-190337057bac-kube-api-access-cfpc8\") pod \"controller-manager-65d6c4859f-tqqv5\" (UID: \"2225553d-99ab-4be0-8515-190337057bac\") " pod="openshift-controller-manager/controller-manager-65d6c4859f-tqqv5" Mar 07 14:42:31 crc kubenswrapper[4943]: I0307 14:42:31.119863 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxkrz\" (UniqueName: \"kubernetes.io/projected/b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a-kube-api-access-mxkrz\") pod \"route-controller-manager-574f768f98-q28jc\" (UID: \"b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a\") " pod="openshift-route-controller-manager/route-controller-manager-574f768f98-q28jc" Mar 07 14:42:31 crc kubenswrapper[4943]: I0307 14:42:31.162264 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-65d6c4859f-tqqv5" Mar 07 14:42:31 crc kubenswrapper[4943]: I0307 14:42:31.182052 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-574f768f98-q28jc" Mar 07 14:42:31 crc kubenswrapper[4943]: I0307 14:42:31.444567 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-p767n" podUID="6a05c7ae-f2a1-4116-9038-ac838fa88af7" containerName="registry-server" containerID="cri-o://195c1f3f178736421b8271566e1f002a2c5619c4109019c84bfd76a1c3e498a6" gracePeriod=2 Mar 07 14:42:31 crc kubenswrapper[4943]: I0307 14:42:31.541723 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rvkgz"] Mar 07 14:42:31 crc kubenswrapper[4943]: I0307 14:42:31.655454 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-65d6c4859f-tqqv5"] Mar 07 14:42:31 crc kubenswrapper[4943]: W0307 14:42:31.672186 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2225553d_99ab_4be0_8515_190337057bac.slice/crio-05b5f181f70bc4b4f9181b0884a852d540a7e9b8be49a9811a0f476c20c617c3 WatchSource:0}: Error finding container 05b5f181f70bc4b4f9181b0884a852d540a7e9b8be49a9811a0f476c20c617c3: Status 404 returned error can't find the container with id 05b5f181f70bc4b4f9181b0884a852d540a7e9b8be49a9811a0f476c20c617c3 Mar 07 14:42:31 crc kubenswrapper[4943]: I0307 14:42:31.700843 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-574f768f98-q28jc"] Mar 07 14:42:31 crc kubenswrapper[4943]: W0307 14:42:31.716509 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb5e747b4_9fec_42f5_8ffe_dcd7f2bcf28a.slice/crio-cc04508d21d1e615d72367ea89dac7d015de1f508e987eb4abe91624f3bfe8e4 WatchSource:0}: Error finding container cc04508d21d1e615d72367ea89dac7d015de1f508e987eb4abe91624f3bfe8e4: Status 404 returned error can't find the container with id cc04508d21d1e615d72367ea89dac7d015de1f508e987eb4abe91624f3bfe8e4 Mar 07 14:42:31 crc kubenswrapper[4943]: I0307 14:42:31.816911 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p767n" Mar 07 14:42:32 crc kubenswrapper[4943]: I0307 14:42:32.001598 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a05c7ae-f2a1-4116-9038-ac838fa88af7-catalog-content\") pod \"6a05c7ae-f2a1-4116-9038-ac838fa88af7\" (UID: \"6a05c7ae-f2a1-4116-9038-ac838fa88af7\") " Mar 07 14:42:32 crc kubenswrapper[4943]: I0307 14:42:32.001922 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rf8ld\" (UniqueName: \"kubernetes.io/projected/6a05c7ae-f2a1-4116-9038-ac838fa88af7-kube-api-access-rf8ld\") pod \"6a05c7ae-f2a1-4116-9038-ac838fa88af7\" (UID: \"6a05c7ae-f2a1-4116-9038-ac838fa88af7\") " Mar 07 14:42:32 crc kubenswrapper[4943]: I0307 14:42:32.001991 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a05c7ae-f2a1-4116-9038-ac838fa88af7-utilities\") pod \"6a05c7ae-f2a1-4116-9038-ac838fa88af7\" (UID: \"6a05c7ae-f2a1-4116-9038-ac838fa88af7\") " Mar 07 14:42:32 crc kubenswrapper[4943]: I0307 14:42:32.002822 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a05c7ae-f2a1-4116-9038-ac838fa88af7-utilities" (OuterVolumeSpecName: "utilities") pod "6a05c7ae-f2a1-4116-9038-ac838fa88af7" (UID: "6a05c7ae-f2a1-4116-9038-ac838fa88af7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:42:32 crc kubenswrapper[4943]: I0307 14:42:32.007285 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a05c7ae-f2a1-4116-9038-ac838fa88af7-kube-api-access-rf8ld" (OuterVolumeSpecName: "kube-api-access-rf8ld") pod "6a05c7ae-f2a1-4116-9038-ac838fa88af7" (UID: "6a05c7ae-f2a1-4116-9038-ac838fa88af7"). InnerVolumeSpecName "kube-api-access-rf8ld". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:42:32 crc kubenswrapper[4943]: I0307 14:42:32.036243 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a05c7ae-f2a1-4116-9038-ac838fa88af7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6a05c7ae-f2a1-4116-9038-ac838fa88af7" (UID: "6a05c7ae-f2a1-4116-9038-ac838fa88af7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:42:32 crc kubenswrapper[4943]: I0307 14:42:32.104074 4943 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a05c7ae-f2a1-4116-9038-ac838fa88af7-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:32 crc kubenswrapper[4943]: I0307 14:42:32.104098 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rf8ld\" (UniqueName: \"kubernetes.io/projected/6a05c7ae-f2a1-4116-9038-ac838fa88af7-kube-api-access-rf8ld\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:32 crc kubenswrapper[4943]: I0307 14:42:32.104108 4943 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a05c7ae-f2a1-4116-9038-ac838fa88af7-utilities\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:32 crc kubenswrapper[4943]: I0307 14:42:32.451864 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-65d6c4859f-tqqv5" event={"ID":"2225553d-99ab-4be0-8515-190337057bac","Type":"ContainerStarted","Data":"de993e98128a57e41d841f2021a4631d64c5a617a88eea1b8841024cfc7c84b2"} Mar 07 14:42:32 crc kubenswrapper[4943]: I0307 14:42:32.451945 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-65d6c4859f-tqqv5" event={"ID":"2225553d-99ab-4be0-8515-190337057bac","Type":"ContainerStarted","Data":"05b5f181f70bc4b4f9181b0884a852d540a7e9b8be49a9811a0f476c20c617c3"} Mar 07 14:42:32 crc kubenswrapper[4943]: I0307 14:42:32.453071 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-65d6c4859f-tqqv5" Mar 07 14:42:32 crc kubenswrapper[4943]: I0307 14:42:32.454446 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-574f768f98-q28jc" event={"ID":"b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a","Type":"ContainerStarted","Data":"51e6cc808f78e2b17b8de230ede94a6602929050d1899c36e29302d6685c6f18"} Mar 07 14:42:32 crc kubenswrapper[4943]: I0307 14:42:32.454512 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-574f768f98-q28jc" event={"ID":"b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a","Type":"ContainerStarted","Data":"cc04508d21d1e615d72367ea89dac7d015de1f508e987eb4abe91624f3bfe8e4"} Mar 07 14:42:32 crc kubenswrapper[4943]: I0307 14:42:32.454699 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-574f768f98-q28jc" Mar 07 14:42:32 crc kubenswrapper[4943]: I0307 14:42:32.456706 4943 generic.go:334] "Generic (PLEG): container finished" podID="6a05c7ae-f2a1-4116-9038-ac838fa88af7" containerID="195c1f3f178736421b8271566e1f002a2c5619c4109019c84bfd76a1c3e498a6" exitCode=0 Mar 07 14:42:32 crc kubenswrapper[4943]: I0307 14:42:32.456787 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p767n" Mar 07 14:42:32 crc kubenswrapper[4943]: I0307 14:42:32.456790 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p767n" event={"ID":"6a05c7ae-f2a1-4116-9038-ac838fa88af7","Type":"ContainerDied","Data":"195c1f3f178736421b8271566e1f002a2c5619c4109019c84bfd76a1c3e498a6"} Mar 07 14:42:32 crc kubenswrapper[4943]: I0307 14:42:32.456846 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p767n" event={"ID":"6a05c7ae-f2a1-4116-9038-ac838fa88af7","Type":"ContainerDied","Data":"79d319958a42f72270e5dd7104572de0fb9a91005002b405489433c291e60cca"} Mar 07 14:42:32 crc kubenswrapper[4943]: I0307 14:42:32.456876 4943 scope.go:117] "RemoveContainer" containerID="195c1f3f178736421b8271566e1f002a2c5619c4109019c84bfd76a1c3e498a6" Mar 07 14:42:32 crc kubenswrapper[4943]: I0307 14:42:32.456918 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rvkgz" podUID="4ae1cf01-1400-4500-9b84-8435a6541612" containerName="registry-server" containerID="cri-o://a1c0898bb09c451613fd5085b4e13b127657ce13de6cdaff591a3d96bbc8c225" gracePeriod=2 Mar 07 14:42:32 crc kubenswrapper[4943]: I0307 14:42:32.463539 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-65d6c4859f-tqqv5" Mar 07 14:42:32 crc kubenswrapper[4943]: I0307 14:42:32.475850 4943 scope.go:117] "RemoveContainer" containerID="dacb590c8ccf77514c2fe8c0e0bc44da9380e7b5bbfd99d17df8072d8e55f4fe" Mar 07 14:42:32 crc kubenswrapper[4943]: I0307 14:42:32.500968 4943 scope.go:117] "RemoveContainer" containerID="5716e592ac5b4912810aeda7b6fc5b837c8c5c98afe83537ed3b4bea0e1fe36d" Mar 07 14:42:32 crc kubenswrapper[4943]: I0307 14:42:32.505401 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-65d6c4859f-tqqv5" podStartSLOduration=3.505387619 podStartE2EDuration="3.505387619s" podCreationTimestamp="2026-03-07 14:42:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:42:32.48467442 +0000 UTC m=+194.436810918" watchObservedRunningTime="2026-03-07 14:42:32.505387619 +0000 UTC m=+194.457524117" Mar 07 14:42:32 crc kubenswrapper[4943]: I0307 14:42:32.519883 4943 scope.go:117] "RemoveContainer" containerID="195c1f3f178736421b8271566e1f002a2c5619c4109019c84bfd76a1c3e498a6" Mar 07 14:42:32 crc kubenswrapper[4943]: E0307 14:42:32.520457 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"195c1f3f178736421b8271566e1f002a2c5619c4109019c84bfd76a1c3e498a6\": container with ID starting with 195c1f3f178736421b8271566e1f002a2c5619c4109019c84bfd76a1c3e498a6 not found: ID does not exist" containerID="195c1f3f178736421b8271566e1f002a2c5619c4109019c84bfd76a1c3e498a6" Mar 07 14:42:32 crc kubenswrapper[4943]: I0307 14:42:32.520502 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"195c1f3f178736421b8271566e1f002a2c5619c4109019c84bfd76a1c3e498a6"} err="failed to get container status \"195c1f3f178736421b8271566e1f002a2c5619c4109019c84bfd76a1c3e498a6\": rpc error: code = NotFound desc = could not find container \"195c1f3f178736421b8271566e1f002a2c5619c4109019c84bfd76a1c3e498a6\": container with ID starting with 195c1f3f178736421b8271566e1f002a2c5619c4109019c84bfd76a1c3e498a6 not found: ID does not exist" Mar 07 14:42:32 crc kubenswrapper[4943]: I0307 14:42:32.520525 4943 scope.go:117] "RemoveContainer" containerID="dacb590c8ccf77514c2fe8c0e0bc44da9380e7b5bbfd99d17df8072d8e55f4fe" Mar 07 14:42:32 crc kubenswrapper[4943]: E0307 14:42:32.520971 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dacb590c8ccf77514c2fe8c0e0bc44da9380e7b5bbfd99d17df8072d8e55f4fe\": container with ID starting with dacb590c8ccf77514c2fe8c0e0bc44da9380e7b5bbfd99d17df8072d8e55f4fe not found: ID does not exist" containerID="dacb590c8ccf77514c2fe8c0e0bc44da9380e7b5bbfd99d17df8072d8e55f4fe" Mar 07 14:42:32 crc kubenswrapper[4943]: I0307 14:42:32.521004 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dacb590c8ccf77514c2fe8c0e0bc44da9380e7b5bbfd99d17df8072d8e55f4fe"} err="failed to get container status \"dacb590c8ccf77514c2fe8c0e0bc44da9380e7b5bbfd99d17df8072d8e55f4fe\": rpc error: code = NotFound desc = could not find container \"dacb590c8ccf77514c2fe8c0e0bc44da9380e7b5bbfd99d17df8072d8e55f4fe\": container with ID starting with dacb590c8ccf77514c2fe8c0e0bc44da9380e7b5bbfd99d17df8072d8e55f4fe not found: ID does not exist" Mar 07 14:42:32 crc kubenswrapper[4943]: I0307 14:42:32.521023 4943 scope.go:117] "RemoveContainer" containerID="5716e592ac5b4912810aeda7b6fc5b837c8c5c98afe83537ed3b4bea0e1fe36d" Mar 07 14:42:32 crc kubenswrapper[4943]: E0307 14:42:32.523225 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5716e592ac5b4912810aeda7b6fc5b837c8c5c98afe83537ed3b4bea0e1fe36d\": container with ID starting with 5716e592ac5b4912810aeda7b6fc5b837c8c5c98afe83537ed3b4bea0e1fe36d not found: ID does not exist" containerID="5716e592ac5b4912810aeda7b6fc5b837c8c5c98afe83537ed3b4bea0e1fe36d" Mar 07 14:42:32 crc kubenswrapper[4943]: I0307 14:42:32.523254 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5716e592ac5b4912810aeda7b6fc5b837c8c5c98afe83537ed3b4bea0e1fe36d"} err="failed to get container status \"5716e592ac5b4912810aeda7b6fc5b837c8c5c98afe83537ed3b4bea0e1fe36d\": rpc error: code = NotFound desc = could not find container \"5716e592ac5b4912810aeda7b6fc5b837c8c5c98afe83537ed3b4bea0e1fe36d\": container with ID starting with 5716e592ac5b4912810aeda7b6fc5b837c8c5c98afe83537ed3b4bea0e1fe36d not found: ID does not exist" Mar 07 14:42:32 crc kubenswrapper[4943]: I0307 14:42:32.533094 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-574f768f98-q28jc" podStartSLOduration=3.533070866 podStartE2EDuration="3.533070866s" podCreationTimestamp="2026-03-07 14:42:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:42:32.529326986 +0000 UTC m=+194.481463484" watchObservedRunningTime="2026-03-07 14:42:32.533070866 +0000 UTC m=+194.485207374" Mar 07 14:42:32 crc kubenswrapper[4943]: I0307 14:42:32.544805 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-p767n"] Mar 07 14:42:32 crc kubenswrapper[4943]: I0307 14:42:32.549588 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-p767n"] Mar 07 14:42:32 crc kubenswrapper[4943]: I0307 14:42:32.732177 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-574f768f98-q28jc" Mar 07 14:42:32 crc kubenswrapper[4943]: I0307 14:42:32.762420 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a05c7ae-f2a1-4116-9038-ac838fa88af7" path="/var/lib/kubelet/pods/6a05c7ae-f2a1-4116-9038-ac838fa88af7/volumes" Mar 07 14:42:32 crc kubenswrapper[4943]: I0307 14:42:32.848158 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rvkgz" Mar 07 14:42:33 crc kubenswrapper[4943]: I0307 14:42:33.015640 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ae1cf01-1400-4500-9b84-8435a6541612-utilities\") pod \"4ae1cf01-1400-4500-9b84-8435a6541612\" (UID: \"4ae1cf01-1400-4500-9b84-8435a6541612\") " Mar 07 14:42:33 crc kubenswrapper[4943]: I0307 14:42:33.015720 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8p98b\" (UniqueName: \"kubernetes.io/projected/4ae1cf01-1400-4500-9b84-8435a6541612-kube-api-access-8p98b\") pod \"4ae1cf01-1400-4500-9b84-8435a6541612\" (UID: \"4ae1cf01-1400-4500-9b84-8435a6541612\") " Mar 07 14:42:33 crc kubenswrapper[4943]: I0307 14:42:33.015801 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ae1cf01-1400-4500-9b84-8435a6541612-catalog-content\") pod \"4ae1cf01-1400-4500-9b84-8435a6541612\" (UID: \"4ae1cf01-1400-4500-9b84-8435a6541612\") " Mar 07 14:42:33 crc kubenswrapper[4943]: I0307 14:42:33.018122 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ae1cf01-1400-4500-9b84-8435a6541612-utilities" (OuterVolumeSpecName: "utilities") pod "4ae1cf01-1400-4500-9b84-8435a6541612" (UID: "4ae1cf01-1400-4500-9b84-8435a6541612"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:42:33 crc kubenswrapper[4943]: I0307 14:42:33.021970 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ae1cf01-1400-4500-9b84-8435a6541612-kube-api-access-8p98b" (OuterVolumeSpecName: "kube-api-access-8p98b") pod "4ae1cf01-1400-4500-9b84-8435a6541612" (UID: "4ae1cf01-1400-4500-9b84-8435a6541612"). InnerVolumeSpecName "kube-api-access-8p98b". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:42:33 crc kubenswrapper[4943]: I0307 14:42:33.117774 4943 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ae1cf01-1400-4500-9b84-8435a6541612-utilities\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:33 crc kubenswrapper[4943]: I0307 14:42:33.117831 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8p98b\" (UniqueName: \"kubernetes.io/projected/4ae1cf01-1400-4500-9b84-8435a6541612-kube-api-access-8p98b\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:33 crc kubenswrapper[4943]: I0307 14:42:33.173774 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ae1cf01-1400-4500-9b84-8435a6541612-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4ae1cf01-1400-4500-9b84-8435a6541612" (UID: "4ae1cf01-1400-4500-9b84-8435a6541612"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:42:33 crc kubenswrapper[4943]: I0307 14:42:33.219062 4943 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ae1cf01-1400-4500-9b84-8435a6541612-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:33 crc kubenswrapper[4943]: I0307 14:42:33.467741 4943 generic.go:334] "Generic (PLEG): container finished" podID="4ae1cf01-1400-4500-9b84-8435a6541612" containerID="a1c0898bb09c451613fd5085b4e13b127657ce13de6cdaff591a3d96bbc8c225" exitCode=0 Mar 07 14:42:33 crc kubenswrapper[4943]: I0307 14:42:33.467793 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rvkgz" event={"ID":"4ae1cf01-1400-4500-9b84-8435a6541612","Type":"ContainerDied","Data":"a1c0898bb09c451613fd5085b4e13b127657ce13de6cdaff591a3d96bbc8c225"} Mar 07 14:42:33 crc kubenswrapper[4943]: I0307 14:42:33.467849 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rvkgz" event={"ID":"4ae1cf01-1400-4500-9b84-8435a6541612","Type":"ContainerDied","Data":"808f86db1ff79b593ed5ac29dd9cf0bab9698e7d67dec02a4362737858863de2"} Mar 07 14:42:33 crc kubenswrapper[4943]: I0307 14:42:33.467881 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rvkgz" Mar 07 14:42:33 crc kubenswrapper[4943]: I0307 14:42:33.467906 4943 scope.go:117] "RemoveContainer" containerID="a1c0898bb09c451613fd5085b4e13b127657ce13de6cdaff591a3d96bbc8c225" Mar 07 14:42:33 crc kubenswrapper[4943]: I0307 14:42:33.496172 4943 scope.go:117] "RemoveContainer" containerID="a219180b9098cbaad437ff15fcfcb8445f7a4b73c0fe03c40cf487762dee6dad" Mar 07 14:42:33 crc kubenswrapper[4943]: I0307 14:42:33.517786 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rvkgz"] Mar 07 14:42:33 crc kubenswrapper[4943]: I0307 14:42:33.527900 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rvkgz"] Mar 07 14:42:33 crc kubenswrapper[4943]: I0307 14:42:33.549969 4943 scope.go:117] "RemoveContainer" containerID="c348d8eae504da56c00175d59f8486fed8406de64de45371f3818a9673b00289" Mar 07 14:42:33 crc kubenswrapper[4943]: I0307 14:42:33.591761 4943 scope.go:117] "RemoveContainer" containerID="a1c0898bb09c451613fd5085b4e13b127657ce13de6cdaff591a3d96bbc8c225" Mar 07 14:42:33 crc kubenswrapper[4943]: E0307 14:42:33.592415 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1c0898bb09c451613fd5085b4e13b127657ce13de6cdaff591a3d96bbc8c225\": container with ID starting with a1c0898bb09c451613fd5085b4e13b127657ce13de6cdaff591a3d96bbc8c225 not found: ID does not exist" containerID="a1c0898bb09c451613fd5085b4e13b127657ce13de6cdaff591a3d96bbc8c225" Mar 07 14:42:33 crc kubenswrapper[4943]: I0307 14:42:33.592456 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1c0898bb09c451613fd5085b4e13b127657ce13de6cdaff591a3d96bbc8c225"} err="failed to get container status \"a1c0898bb09c451613fd5085b4e13b127657ce13de6cdaff591a3d96bbc8c225\": rpc error: code = NotFound desc = could not find container \"a1c0898bb09c451613fd5085b4e13b127657ce13de6cdaff591a3d96bbc8c225\": container with ID starting with a1c0898bb09c451613fd5085b4e13b127657ce13de6cdaff591a3d96bbc8c225 not found: ID does not exist" Mar 07 14:42:33 crc kubenswrapper[4943]: I0307 14:42:33.592485 4943 scope.go:117] "RemoveContainer" containerID="a219180b9098cbaad437ff15fcfcb8445f7a4b73c0fe03c40cf487762dee6dad" Mar 07 14:42:33 crc kubenswrapper[4943]: E0307 14:42:33.592979 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a219180b9098cbaad437ff15fcfcb8445f7a4b73c0fe03c40cf487762dee6dad\": container with ID starting with a219180b9098cbaad437ff15fcfcb8445f7a4b73c0fe03c40cf487762dee6dad not found: ID does not exist" containerID="a219180b9098cbaad437ff15fcfcb8445f7a4b73c0fe03c40cf487762dee6dad" Mar 07 14:42:33 crc kubenswrapper[4943]: I0307 14:42:33.593013 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a219180b9098cbaad437ff15fcfcb8445f7a4b73c0fe03c40cf487762dee6dad"} err="failed to get container status \"a219180b9098cbaad437ff15fcfcb8445f7a4b73c0fe03c40cf487762dee6dad\": rpc error: code = NotFound desc = could not find container \"a219180b9098cbaad437ff15fcfcb8445f7a4b73c0fe03c40cf487762dee6dad\": container with ID starting with a219180b9098cbaad437ff15fcfcb8445f7a4b73c0fe03c40cf487762dee6dad not found: ID does not exist" Mar 07 14:42:33 crc kubenswrapper[4943]: I0307 14:42:33.593032 4943 scope.go:117] "RemoveContainer" containerID="c348d8eae504da56c00175d59f8486fed8406de64de45371f3818a9673b00289" Mar 07 14:42:33 crc kubenswrapper[4943]: E0307 14:42:33.593887 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c348d8eae504da56c00175d59f8486fed8406de64de45371f3818a9673b00289\": container with ID starting with c348d8eae504da56c00175d59f8486fed8406de64de45371f3818a9673b00289 not found: ID does not exist" containerID="c348d8eae504da56c00175d59f8486fed8406de64de45371f3818a9673b00289" Mar 07 14:42:33 crc kubenswrapper[4943]: I0307 14:42:33.593911 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c348d8eae504da56c00175d59f8486fed8406de64de45371f3818a9673b00289"} err="failed to get container status \"c348d8eae504da56c00175d59f8486fed8406de64de45371f3818a9673b00289\": rpc error: code = NotFound desc = could not find container \"c348d8eae504da56c00175d59f8486fed8406de64de45371f3818a9673b00289\": container with ID starting with c348d8eae504da56c00175d59f8486fed8406de64de45371f3818a9673b00289 not found: ID does not exist" Mar 07 14:42:34 crc kubenswrapper[4943]: I0307 14:42:34.766137 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ae1cf01-1400-4500-9b84-8435a6541612" path="/var/lib/kubelet/pods/4ae1cf01-1400-4500-9b84-8435a6541612/volumes" Mar 07 14:42:36 crc kubenswrapper[4943]: I0307 14:42:36.084105 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dfqzr" Mar 07 14:42:37 crc kubenswrapper[4943]: I0307 14:42:37.974992 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 07 14:42:38 crc kubenswrapper[4943]: I0307 14:42:38.351022 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-cnbpx"] Mar 07 14:42:48 crc kubenswrapper[4943]: I0307 14:42:48.993652 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-65d6c4859f-tqqv5"] Mar 07 14:42:48 crc kubenswrapper[4943]: I0307 14:42:48.994533 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-65d6c4859f-tqqv5" podUID="2225553d-99ab-4be0-8515-190337057bac" containerName="controller-manager" containerID="cri-o://de993e98128a57e41d841f2021a4631d64c5a617a88eea1b8841024cfc7c84b2" gracePeriod=30 Mar 07 14:42:49 crc kubenswrapper[4943]: I0307 14:42:49.088617 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-574f768f98-q28jc"] Mar 07 14:42:49 crc kubenswrapper[4943]: I0307 14:42:49.088848 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-574f768f98-q28jc" podUID="b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a" containerName="route-controller-manager" containerID="cri-o://51e6cc808f78e2b17b8de230ede94a6602929050d1899c36e29302d6685c6f18" gracePeriod=30 Mar 07 14:42:49 crc kubenswrapper[4943]: I0307 14:42:49.616948 4943 generic.go:334] "Generic (PLEG): container finished" podID="2225553d-99ab-4be0-8515-190337057bac" containerID="de993e98128a57e41d841f2021a4631d64c5a617a88eea1b8841024cfc7c84b2" exitCode=0 Mar 07 14:42:49 crc kubenswrapper[4943]: I0307 14:42:49.617140 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-65d6c4859f-tqqv5" event={"ID":"2225553d-99ab-4be0-8515-190337057bac","Type":"ContainerDied","Data":"de993e98128a57e41d841f2021a4631d64c5a617a88eea1b8841024cfc7c84b2"} Mar 07 14:42:49 crc kubenswrapper[4943]: I0307 14:42:49.619688 4943 generic.go:334] "Generic (PLEG): container finished" podID="b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a" containerID="51e6cc808f78e2b17b8de230ede94a6602929050d1899c36e29302d6685c6f18" exitCode=0 Mar 07 14:42:49 crc kubenswrapper[4943]: I0307 14:42:49.619731 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-574f768f98-q28jc" event={"ID":"b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a","Type":"ContainerDied","Data":"51e6cc808f78e2b17b8de230ede94a6602929050d1899c36e29302d6685c6f18"} Mar 07 14:42:49 crc kubenswrapper[4943]: I0307 14:42:49.619762 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-574f768f98-q28jc" event={"ID":"b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a","Type":"ContainerDied","Data":"cc04508d21d1e615d72367ea89dac7d015de1f508e987eb4abe91624f3bfe8e4"} Mar 07 14:42:49 crc kubenswrapper[4943]: I0307 14:42:49.619780 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc04508d21d1e615d72367ea89dac7d015de1f508e987eb4abe91624f3bfe8e4" Mar 07 14:42:49 crc kubenswrapper[4943]: I0307 14:42:49.684797 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-574f768f98-q28jc" Mar 07 14:42:49 crc kubenswrapper[4943]: I0307 14:42:49.691286 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-65d6c4859f-tqqv5" Mar 07 14:42:49 crc kubenswrapper[4943]: I0307 14:42:49.745827 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfpc8\" (UniqueName: \"kubernetes.io/projected/2225553d-99ab-4be0-8515-190337057bac-kube-api-access-cfpc8\") pod \"2225553d-99ab-4be0-8515-190337057bac\" (UID: \"2225553d-99ab-4be0-8515-190337057bac\") " Mar 07 14:42:49 crc kubenswrapper[4943]: I0307 14:42:49.754018 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2225553d-99ab-4be0-8515-190337057bac-kube-api-access-cfpc8" (OuterVolumeSpecName: "kube-api-access-cfpc8") pod "2225553d-99ab-4be0-8515-190337057bac" (UID: "2225553d-99ab-4be0-8515-190337057bac"). InnerVolumeSpecName "kube-api-access-cfpc8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:42:49 crc kubenswrapper[4943]: I0307 14:42:49.847058 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a-config\") pod \"b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a\" (UID: \"b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a\") " Mar 07 14:42:49 crc kubenswrapper[4943]: I0307 14:42:49.847786 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2225553d-99ab-4be0-8515-190337057bac-config\") pod \"2225553d-99ab-4be0-8515-190337057bac\" (UID: \"2225553d-99ab-4be0-8515-190337057bac\") " Mar 07 14:42:49 crc kubenswrapper[4943]: I0307 14:42:49.848054 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mxkrz\" (UniqueName: \"kubernetes.io/projected/b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a-kube-api-access-mxkrz\") pod \"b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a\" (UID: \"b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a\") " Mar 07 14:42:49 crc kubenswrapper[4943]: I0307 14:42:49.848249 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2225553d-99ab-4be0-8515-190337057bac-client-ca\") pod \"2225553d-99ab-4be0-8515-190337057bac\" (UID: \"2225553d-99ab-4be0-8515-190337057bac\") " Mar 07 14:42:49 crc kubenswrapper[4943]: I0307 14:42:49.848431 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2225553d-99ab-4be0-8515-190337057bac-serving-cert\") pod \"2225553d-99ab-4be0-8515-190337057bac\" (UID: \"2225553d-99ab-4be0-8515-190337057bac\") " Mar 07 14:42:49 crc kubenswrapper[4943]: I0307 14:42:49.848642 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a-serving-cert\") pod \"b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a\" (UID: \"b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a\") " Mar 07 14:42:49 crc kubenswrapper[4943]: I0307 14:42:49.848847 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2225553d-99ab-4be0-8515-190337057bac-proxy-ca-bundles\") pod \"2225553d-99ab-4be0-8515-190337057bac\" (UID: \"2225553d-99ab-4be0-8515-190337057bac\") " Mar 07 14:42:49 crc kubenswrapper[4943]: I0307 14:42:49.849037 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a-client-ca\") pod \"b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a\" (UID: \"b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a\") " Mar 07 14:42:49 crc kubenswrapper[4943]: I0307 14:42:49.849216 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a-config" (OuterVolumeSpecName: "config") pod "b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a" (UID: "b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:42:49 crc kubenswrapper[4943]: I0307 14:42:49.849765 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2225553d-99ab-4be0-8515-190337057bac-client-ca" (OuterVolumeSpecName: "client-ca") pod "2225553d-99ab-4be0-8515-190337057bac" (UID: "2225553d-99ab-4be0-8515-190337057bac"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:42:49 crc kubenswrapper[4943]: I0307 14:42:49.849784 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfpc8\" (UniqueName: \"kubernetes.io/projected/2225553d-99ab-4be0-8515-190337057bac-kube-api-access-cfpc8\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:49 crc kubenswrapper[4943]: I0307 14:42:49.849853 4943 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a-config\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:49 crc kubenswrapper[4943]: I0307 14:42:49.849885 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a-client-ca" (OuterVolumeSpecName: "client-ca") pod "b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a" (UID: "b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:42:49 crc kubenswrapper[4943]: I0307 14:42:49.849975 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2225553d-99ab-4be0-8515-190337057bac-config" (OuterVolumeSpecName: "config") pod "2225553d-99ab-4be0-8515-190337057bac" (UID: "2225553d-99ab-4be0-8515-190337057bac"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:42:49 crc kubenswrapper[4943]: I0307 14:42:49.850110 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2225553d-99ab-4be0-8515-190337057bac-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "2225553d-99ab-4be0-8515-190337057bac" (UID: "2225553d-99ab-4be0-8515-190337057bac"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:42:49 crc kubenswrapper[4943]: I0307 14:42:49.853332 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2225553d-99ab-4be0-8515-190337057bac-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "2225553d-99ab-4be0-8515-190337057bac" (UID: "2225553d-99ab-4be0-8515-190337057bac"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:42:49 crc kubenswrapper[4943]: I0307 14:42:49.853830 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a-kube-api-access-mxkrz" (OuterVolumeSpecName: "kube-api-access-mxkrz") pod "b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a" (UID: "b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a"). InnerVolumeSpecName "kube-api-access-mxkrz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:42:49 crc kubenswrapper[4943]: I0307 14:42:49.854405 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a" (UID: "b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:42:49 crc kubenswrapper[4943]: I0307 14:42:49.951194 4943 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2225553d-99ab-4be0-8515-190337057bac-client-ca\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:49 crc kubenswrapper[4943]: I0307 14:42:49.951255 4943 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2225553d-99ab-4be0-8515-190337057bac-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:49 crc kubenswrapper[4943]: I0307 14:42:49.951283 4943 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:49 crc kubenswrapper[4943]: I0307 14:42:49.951309 4943 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2225553d-99ab-4be0-8515-190337057bac-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:49 crc kubenswrapper[4943]: I0307 14:42:49.951339 4943 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a-client-ca\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:49 crc kubenswrapper[4943]: I0307 14:42:49.951362 4943 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2225553d-99ab-4be0-8515-190337057bac-config\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:49 crc kubenswrapper[4943]: I0307 14:42:49.951389 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mxkrz\" (UniqueName: \"kubernetes.io/projected/b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a-kube-api-access-mxkrz\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.629152 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-65d6c4859f-tqqv5" Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.629150 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-574f768f98-q28jc" Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.629154 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-65d6c4859f-tqqv5" event={"ID":"2225553d-99ab-4be0-8515-190337057bac","Type":"ContainerDied","Data":"05b5f181f70bc4b4f9181b0884a852d540a7e9b8be49a9811a0f476c20c617c3"} Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.630887 4943 scope.go:117] "RemoveContainer" containerID="de993e98128a57e41d841f2021a4631d64c5a617a88eea1b8841024cfc7c84b2" Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.680144 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-65d6c4859f-tqqv5"] Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.687002 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-65d6c4859f-tqqv5"] Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.693056 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-574f768f98-q28jc"] Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.696122 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-574f768f98-q28jc"] Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.766346 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2225553d-99ab-4be0-8515-190337057bac" path="/var/lib/kubelet/pods/2225553d-99ab-4be0-8515-190337057bac/volumes" Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.767776 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a" path="/var/lib/kubelet/pods/b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a/volumes" Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.831239 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-86cb58c977-j8f96"] Mar 07 14:42:50 crc kubenswrapper[4943]: E0307 14:42:50.831550 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a05c7ae-f2a1-4116-9038-ac838fa88af7" containerName="extract-content" Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.831568 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a05c7ae-f2a1-4116-9038-ac838fa88af7" containerName="extract-content" Mar 07 14:42:50 crc kubenswrapper[4943]: E0307 14:42:50.831586 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a05c7ae-f2a1-4116-9038-ac838fa88af7" containerName="extract-utilities" Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.831595 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a05c7ae-f2a1-4116-9038-ac838fa88af7" containerName="extract-utilities" Mar 07 14:42:50 crc kubenswrapper[4943]: E0307 14:42:50.831610 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ae1cf01-1400-4500-9b84-8435a6541612" containerName="extract-content" Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.831620 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ae1cf01-1400-4500-9b84-8435a6541612" containerName="extract-content" Mar 07 14:42:50 crc kubenswrapper[4943]: E0307 14:42:50.831634 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a" containerName="route-controller-manager" Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.831644 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a" containerName="route-controller-manager" Mar 07 14:42:50 crc kubenswrapper[4943]: E0307 14:42:50.831665 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2225553d-99ab-4be0-8515-190337057bac" containerName="controller-manager" Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.831674 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="2225553d-99ab-4be0-8515-190337057bac" containerName="controller-manager" Mar 07 14:42:50 crc kubenswrapper[4943]: E0307 14:42:50.831687 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a05c7ae-f2a1-4116-9038-ac838fa88af7" containerName="registry-server" Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.831695 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a05c7ae-f2a1-4116-9038-ac838fa88af7" containerName="registry-server" Mar 07 14:42:50 crc kubenswrapper[4943]: E0307 14:42:50.831709 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ae1cf01-1400-4500-9b84-8435a6541612" containerName="extract-utilities" Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.831718 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ae1cf01-1400-4500-9b84-8435a6541612" containerName="extract-utilities" Mar 07 14:42:50 crc kubenswrapper[4943]: E0307 14:42:50.831730 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ae1cf01-1400-4500-9b84-8435a6541612" containerName="registry-server" Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.831738 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ae1cf01-1400-4500-9b84-8435a6541612" containerName="registry-server" Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.831848 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a05c7ae-f2a1-4116-9038-ac838fa88af7" containerName="registry-server" Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.831863 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ae1cf01-1400-4500-9b84-8435a6541612" containerName="registry-server" Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.831877 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="2225553d-99ab-4be0-8515-190337057bac" containerName="controller-manager" Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.831888 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5e747b4-9fec-42f5-8ffe-dcd7f2bcf28a" containerName="route-controller-manager" Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.832598 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-86cb58c977-j8f96" Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.835900 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.836135 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.836633 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.837088 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.837154 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.840423 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.846312 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7494769d95-sxpjq"] Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.847663 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7494769d95-sxpjq" Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.848742 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.852337 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.852453 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.852654 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-86cb58c977-j8f96"] Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.852718 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.852765 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.853185 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.853449 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.858076 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7494769d95-sxpjq"] Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.867232 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6deeb990-db08-494d-af2b-5be91b4271f6-config\") pod \"route-controller-manager-7494769d95-sxpjq\" (UID: \"6deeb990-db08-494d-af2b-5be91b4271f6\") " pod="openshift-route-controller-manager/route-controller-manager-7494769d95-sxpjq" Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.867304 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svvjz\" (UniqueName: \"kubernetes.io/projected/6deeb990-db08-494d-af2b-5be91b4271f6-kube-api-access-svvjz\") pod \"route-controller-manager-7494769d95-sxpjq\" (UID: \"6deeb990-db08-494d-af2b-5be91b4271f6\") " pod="openshift-route-controller-manager/route-controller-manager-7494769d95-sxpjq" Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.867435 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6deeb990-db08-494d-af2b-5be91b4271f6-client-ca\") pod \"route-controller-manager-7494769d95-sxpjq\" (UID: \"6deeb990-db08-494d-af2b-5be91b4271f6\") " pod="openshift-route-controller-manager/route-controller-manager-7494769d95-sxpjq" Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.867469 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6deeb990-db08-494d-af2b-5be91b4271f6-serving-cert\") pod \"route-controller-manager-7494769d95-sxpjq\" (UID: \"6deeb990-db08-494d-af2b-5be91b4271f6\") " pod="openshift-route-controller-manager/route-controller-manager-7494769d95-sxpjq" Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.969624 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6deeb990-db08-494d-af2b-5be91b4271f6-client-ca\") pod \"route-controller-manager-7494769d95-sxpjq\" (UID: \"6deeb990-db08-494d-af2b-5be91b4271f6\") " pod="openshift-route-controller-manager/route-controller-manager-7494769d95-sxpjq" Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.969717 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6deeb990-db08-494d-af2b-5be91b4271f6-serving-cert\") pod \"route-controller-manager-7494769d95-sxpjq\" (UID: \"6deeb990-db08-494d-af2b-5be91b4271f6\") " pod="openshift-route-controller-manager/route-controller-manager-7494769d95-sxpjq" Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.969762 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/39fa4113-e14f-48e0-9106-52cfeeeba068-client-ca\") pod \"controller-manager-86cb58c977-j8f96\" (UID: \"39fa4113-e14f-48e0-9106-52cfeeeba068\") " pod="openshift-controller-manager/controller-manager-86cb58c977-j8f96" Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.969819 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/39fa4113-e14f-48e0-9106-52cfeeeba068-serving-cert\") pod \"controller-manager-86cb58c977-j8f96\" (UID: \"39fa4113-e14f-48e0-9106-52cfeeeba068\") " pod="openshift-controller-manager/controller-manager-86cb58c977-j8f96" Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.969922 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6deeb990-db08-494d-af2b-5be91b4271f6-config\") pod \"route-controller-manager-7494769d95-sxpjq\" (UID: \"6deeb990-db08-494d-af2b-5be91b4271f6\") " pod="openshift-route-controller-manager/route-controller-manager-7494769d95-sxpjq" Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.970021 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39fa4113-e14f-48e0-9106-52cfeeeba068-config\") pod \"controller-manager-86cb58c977-j8f96\" (UID: \"39fa4113-e14f-48e0-9106-52cfeeeba068\") " pod="openshift-controller-manager/controller-manager-86cb58c977-j8f96" Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.970076 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svvjz\" (UniqueName: \"kubernetes.io/projected/6deeb990-db08-494d-af2b-5be91b4271f6-kube-api-access-svvjz\") pod \"route-controller-manager-7494769d95-sxpjq\" (UID: \"6deeb990-db08-494d-af2b-5be91b4271f6\") " pod="openshift-route-controller-manager/route-controller-manager-7494769d95-sxpjq" Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.970109 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/39fa4113-e14f-48e0-9106-52cfeeeba068-proxy-ca-bundles\") pod \"controller-manager-86cb58c977-j8f96\" (UID: \"39fa4113-e14f-48e0-9106-52cfeeeba068\") " pod="openshift-controller-manager/controller-manager-86cb58c977-j8f96" Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.970148 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7k6q\" (UniqueName: \"kubernetes.io/projected/39fa4113-e14f-48e0-9106-52cfeeeba068-kube-api-access-h7k6q\") pod \"controller-manager-86cb58c977-j8f96\" (UID: \"39fa4113-e14f-48e0-9106-52cfeeeba068\") " pod="openshift-controller-manager/controller-manager-86cb58c977-j8f96" Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.970877 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6deeb990-db08-494d-af2b-5be91b4271f6-client-ca\") pod \"route-controller-manager-7494769d95-sxpjq\" (UID: \"6deeb990-db08-494d-af2b-5be91b4271f6\") " pod="openshift-route-controller-manager/route-controller-manager-7494769d95-sxpjq" Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.970977 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6deeb990-db08-494d-af2b-5be91b4271f6-config\") pod \"route-controller-manager-7494769d95-sxpjq\" (UID: \"6deeb990-db08-494d-af2b-5be91b4271f6\") " pod="openshift-route-controller-manager/route-controller-manager-7494769d95-sxpjq" Mar 07 14:42:50 crc kubenswrapper[4943]: I0307 14:42:50.976797 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6deeb990-db08-494d-af2b-5be91b4271f6-serving-cert\") pod \"route-controller-manager-7494769d95-sxpjq\" (UID: \"6deeb990-db08-494d-af2b-5be91b4271f6\") " pod="openshift-route-controller-manager/route-controller-manager-7494769d95-sxpjq" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.007741 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svvjz\" (UniqueName: \"kubernetes.io/projected/6deeb990-db08-494d-af2b-5be91b4271f6-kube-api-access-svvjz\") pod \"route-controller-manager-7494769d95-sxpjq\" (UID: \"6deeb990-db08-494d-af2b-5be91b4271f6\") " pod="openshift-route-controller-manager/route-controller-manager-7494769d95-sxpjq" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.070853 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/39fa4113-e14f-48e0-9106-52cfeeeba068-client-ca\") pod \"controller-manager-86cb58c977-j8f96\" (UID: \"39fa4113-e14f-48e0-9106-52cfeeeba068\") " pod="openshift-controller-manager/controller-manager-86cb58c977-j8f96" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.070957 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/39fa4113-e14f-48e0-9106-52cfeeeba068-serving-cert\") pod \"controller-manager-86cb58c977-j8f96\" (UID: \"39fa4113-e14f-48e0-9106-52cfeeeba068\") " pod="openshift-controller-manager/controller-manager-86cb58c977-j8f96" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.071055 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39fa4113-e14f-48e0-9106-52cfeeeba068-config\") pod \"controller-manager-86cb58c977-j8f96\" (UID: \"39fa4113-e14f-48e0-9106-52cfeeeba068\") " pod="openshift-controller-manager/controller-manager-86cb58c977-j8f96" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.071104 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/39fa4113-e14f-48e0-9106-52cfeeeba068-proxy-ca-bundles\") pod \"controller-manager-86cb58c977-j8f96\" (UID: \"39fa4113-e14f-48e0-9106-52cfeeeba068\") " pod="openshift-controller-manager/controller-manager-86cb58c977-j8f96" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.071144 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7k6q\" (UniqueName: \"kubernetes.io/projected/39fa4113-e14f-48e0-9106-52cfeeeba068-kube-api-access-h7k6q\") pod \"controller-manager-86cb58c977-j8f96\" (UID: \"39fa4113-e14f-48e0-9106-52cfeeeba068\") " pod="openshift-controller-manager/controller-manager-86cb58c977-j8f96" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.071974 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/39fa4113-e14f-48e0-9106-52cfeeeba068-client-ca\") pod \"controller-manager-86cb58c977-j8f96\" (UID: \"39fa4113-e14f-48e0-9106-52cfeeeba068\") " pod="openshift-controller-manager/controller-manager-86cb58c977-j8f96" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.073597 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/39fa4113-e14f-48e0-9106-52cfeeeba068-proxy-ca-bundles\") pod \"controller-manager-86cb58c977-j8f96\" (UID: \"39fa4113-e14f-48e0-9106-52cfeeeba068\") " pod="openshift-controller-manager/controller-manager-86cb58c977-j8f96" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.073728 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39fa4113-e14f-48e0-9106-52cfeeeba068-config\") pod \"controller-manager-86cb58c977-j8f96\" (UID: \"39fa4113-e14f-48e0-9106-52cfeeeba068\") " pod="openshift-controller-manager/controller-manager-86cb58c977-j8f96" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.076541 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/39fa4113-e14f-48e0-9106-52cfeeeba068-serving-cert\") pod \"controller-manager-86cb58c977-j8f96\" (UID: \"39fa4113-e14f-48e0-9106-52cfeeeba068\") " pod="openshift-controller-manager/controller-manager-86cb58c977-j8f96" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.099621 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7k6q\" (UniqueName: \"kubernetes.io/projected/39fa4113-e14f-48e0-9106-52cfeeeba068-kube-api-access-h7k6q\") pod \"controller-manager-86cb58c977-j8f96\" (UID: \"39fa4113-e14f-48e0-9106-52cfeeeba068\") " pod="openshift-controller-manager/controller-manager-86cb58c977-j8f96" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.177059 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-86cb58c977-j8f96" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.190323 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7494769d95-sxpjq" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.379470 4943 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.380495 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.380866 4943 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.385264 4943 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 07 14:42:51 crc kubenswrapper[4943]: E0307 14:42:51.386742 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.386771 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 07 14:42:51 crc kubenswrapper[4943]: E0307 14:42:51.386789 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.386832 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 07 14:42:51 crc kubenswrapper[4943]: E0307 14:42:51.386853 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.386866 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Mar 07 14:42:51 crc kubenswrapper[4943]: E0307 14:42:51.386910 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.386955 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 07 14:42:51 crc kubenswrapper[4943]: E0307 14:42:51.386982 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.386996 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 07 14:42:51 crc kubenswrapper[4943]: E0307 14:42:51.387044 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.387059 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 07 14:42:51 crc kubenswrapper[4943]: E0307 14:42:51.387073 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.387086 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 07 14:42:51 crc kubenswrapper[4943]: E0307 14:42:51.387130 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.387142 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.387396 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.387424 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.387467 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.387482 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.387494 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.387513 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.387554 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 07 14:42:51 crc kubenswrapper[4943]: E0307 14:42:51.387761 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.387806 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 07 14:42:51 crc kubenswrapper[4943]: E0307 14:42:51.387827 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.387839 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.388104 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.388152 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.388353 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://6084025a1e017c5e63892cdc0cc359ad0297878a6eb80cbf8d4d620bb5f2f681" gracePeriod=15 Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.388453 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://5f9507997ae006ce9864d88b639c5c782aaed107dbd59b42ea0bb34e1606bf0c" gracePeriod=15 Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.388487 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://36053e9d136906da3b2eb3d258bf5b73bb956ea6141497e74f1f18c8f9a937a9" gracePeriod=15 Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.388407 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://69d5ad495f66b4bdb2e419c05c8a8060bfc94df1a9c5cb3156bed86ba312562b" gracePeriod=15 Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.388554 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://46a2cc9aec0e9e4a54e8fa06e6eeaec8ba29e88d0a0672c00f9a01b439ad4331" gracePeriod=15 Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.437384 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.477513 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.477571 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.477598 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.477622 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.477667 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.477690 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.477726 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.477745 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.578671 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.578730 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.578766 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.578783 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.578797 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.578801 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.578815 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.578855 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.578883 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.578952 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.578941 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.579045 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.579132 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.579155 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.579235 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.579354 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.640390 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.642151 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.643105 4943 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="36053e9d136906da3b2eb3d258bf5b73bb956ea6141497e74f1f18c8f9a937a9" exitCode=0 Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.643156 4943 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="69d5ad495f66b4bdb2e419c05c8a8060bfc94df1a9c5cb3156bed86ba312562b" exitCode=0 Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.643178 4943 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="5f9507997ae006ce9864d88b639c5c782aaed107dbd59b42ea0bb34e1606bf0c" exitCode=0 Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.643195 4943 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="46a2cc9aec0e9e4a54e8fa06e6eeaec8ba29e88d0a0672c00f9a01b439ad4331" exitCode=2 Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.643297 4943 scope.go:117] "RemoveContainer" containerID="ed8d27f5a369a691493bda9dbd6e03c098931b8dab637bbd9bacac6fe7979c80" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.647109 4943 generic.go:334] "Generic (PLEG): container finished" podID="979a9c5a-4f69-4b26-8127-f760a697ae91" containerID="ed8efad2a01435fb5db9bfd68f4303f464293e51ca1a7be35011ad53a6dfb3bf" exitCode=0 Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.647146 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"979a9c5a-4f69-4b26-8127-f760a697ae91","Type":"ContainerDied","Data":"ed8efad2a01435fb5db9bfd68f4303f464293e51ca1a7be35011ad53a6dfb3bf"} Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.648160 4943 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.75:6443: connect: connection refused" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.648734 4943 status_manager.go:851] "Failed to get status for pod" podUID="979a9c5a-4f69-4b26-8127-f760a697ae91" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.75:6443: connect: connection refused" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.649219 4943 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.75:6443: connect: connection refused" Mar 07 14:42:51 crc kubenswrapper[4943]: I0307 14:42:51.717722 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 07 14:42:51 crc kubenswrapper[4943]: W0307 14:42:51.747143 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-49c35db207dc324b23f2e574d4ed6745dbe5174671b91d83858d3aeffcf1ff1a WatchSource:0}: Error finding container 49c35db207dc324b23f2e574d4ed6745dbe5174671b91d83858d3aeffcf1ff1a: Status 404 returned error can't find the container with id 49c35db207dc324b23f2e574d4ed6745dbe5174671b91d83858d3aeffcf1ff1a Mar 07 14:42:51 crc kubenswrapper[4943]: E0307 14:42:51.751333 4943 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.75:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189a963845699772 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:42:51.750594418 +0000 UTC m=+213.702730956,LastTimestamp:2026-03-07 14:42:51.750594418 +0000 UTC m=+213.702730956,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:42:51 crc kubenswrapper[4943]: E0307 14:42:51.976848 4943 log.go:32] "RunPodSandbox from runtime service failed" err=< Mar 07 14:42:51 crc kubenswrapper[4943]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_route-controller-manager-7494769d95-sxpjq_openshift-route-controller-manager_6deeb990-db08-494d-af2b-5be91b4271f6_0(ec88c5bccdc6dc648d068e4c3f4ad00738596c1e4b57bc550b4aaa438505bc9e): error adding pod openshift-route-controller-manager_route-controller-manager-7494769d95-sxpjq to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"ec88c5bccdc6dc648d068e4c3f4ad00738596c1e4b57bc550b4aaa438505bc9e" Netns:"/var/run/netns/8777e4da-ef48-4107-b9e5-1701826e5598" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-route-controller-manager;K8S_POD_NAME=route-controller-manager-7494769d95-sxpjq;K8S_POD_INFRA_CONTAINER_ID=ec88c5bccdc6dc648d068e4c3f4ad00738596c1e4b57bc550b4aaa438505bc9e;K8S_POD_UID=6deeb990-db08-494d-af2b-5be91b4271f6" Path:"" ERRORED: error configuring pod [openshift-route-controller-manager/route-controller-manager-7494769d95-sxpjq] networking: Multus: [openshift-route-controller-manager/route-controller-manager-7494769d95-sxpjq/6deeb990-db08-494d-af2b-5be91b4271f6]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod route-controller-manager-7494769d95-sxpjq in out of cluster comm: SetNetworkStatus: failed to update the pod route-controller-manager-7494769d95-sxpjq in out of cluster comm: status update failed for pod /: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-7494769d95-sxpjq?timeout=1m0s": dial tcp 38.102.83.75:6443: connect: connection refused Mar 07 14:42:51 crc kubenswrapper[4943]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Mar 07 14:42:51 crc kubenswrapper[4943]: > Mar 07 14:42:51 crc kubenswrapper[4943]: E0307 14:42:51.977444 4943 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Mar 07 14:42:51 crc kubenswrapper[4943]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_route-controller-manager-7494769d95-sxpjq_openshift-route-controller-manager_6deeb990-db08-494d-af2b-5be91b4271f6_0(ec88c5bccdc6dc648d068e4c3f4ad00738596c1e4b57bc550b4aaa438505bc9e): error adding pod openshift-route-controller-manager_route-controller-manager-7494769d95-sxpjq to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"ec88c5bccdc6dc648d068e4c3f4ad00738596c1e4b57bc550b4aaa438505bc9e" Netns:"/var/run/netns/8777e4da-ef48-4107-b9e5-1701826e5598" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-route-controller-manager;K8S_POD_NAME=route-controller-manager-7494769d95-sxpjq;K8S_POD_INFRA_CONTAINER_ID=ec88c5bccdc6dc648d068e4c3f4ad00738596c1e4b57bc550b4aaa438505bc9e;K8S_POD_UID=6deeb990-db08-494d-af2b-5be91b4271f6" Path:"" ERRORED: error configuring pod [openshift-route-controller-manager/route-controller-manager-7494769d95-sxpjq] networking: Multus: [openshift-route-controller-manager/route-controller-manager-7494769d95-sxpjq/6deeb990-db08-494d-af2b-5be91b4271f6]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod route-controller-manager-7494769d95-sxpjq in out of cluster comm: SetNetworkStatus: failed to update the pod route-controller-manager-7494769d95-sxpjq in out of cluster comm: status update failed for pod /: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-7494769d95-sxpjq?timeout=1m0s": dial tcp 38.102.83.75:6443: connect: connection refused Mar 07 14:42:51 crc kubenswrapper[4943]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Mar 07 14:42:51 crc kubenswrapper[4943]: > pod="openshift-route-controller-manager/route-controller-manager-7494769d95-sxpjq" Mar 07 14:42:51 crc kubenswrapper[4943]: E0307 14:42:51.977480 4943 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Mar 07 14:42:51 crc kubenswrapper[4943]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_route-controller-manager-7494769d95-sxpjq_openshift-route-controller-manager_6deeb990-db08-494d-af2b-5be91b4271f6_0(ec88c5bccdc6dc648d068e4c3f4ad00738596c1e4b57bc550b4aaa438505bc9e): error adding pod openshift-route-controller-manager_route-controller-manager-7494769d95-sxpjq to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"ec88c5bccdc6dc648d068e4c3f4ad00738596c1e4b57bc550b4aaa438505bc9e" Netns:"/var/run/netns/8777e4da-ef48-4107-b9e5-1701826e5598" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-route-controller-manager;K8S_POD_NAME=route-controller-manager-7494769d95-sxpjq;K8S_POD_INFRA_CONTAINER_ID=ec88c5bccdc6dc648d068e4c3f4ad00738596c1e4b57bc550b4aaa438505bc9e;K8S_POD_UID=6deeb990-db08-494d-af2b-5be91b4271f6" Path:"" ERRORED: error configuring pod [openshift-route-controller-manager/route-controller-manager-7494769d95-sxpjq] networking: Multus: [openshift-route-controller-manager/route-controller-manager-7494769d95-sxpjq/6deeb990-db08-494d-af2b-5be91b4271f6]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod route-controller-manager-7494769d95-sxpjq in out of cluster comm: SetNetworkStatus: failed to update the pod route-controller-manager-7494769d95-sxpjq in out of cluster comm: status update failed for pod /: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-7494769d95-sxpjq?timeout=1m0s": dial tcp 38.102.83.75:6443: connect: connection refused Mar 07 14:42:51 crc kubenswrapper[4943]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Mar 07 14:42:51 crc kubenswrapper[4943]: > pod="openshift-route-controller-manager/route-controller-manager-7494769d95-sxpjq" Mar 07 14:42:51 crc kubenswrapper[4943]: E0307 14:42:51.977598 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"route-controller-manager-7494769d95-sxpjq_openshift-route-controller-manager(6deeb990-db08-494d-af2b-5be91b4271f6)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"route-controller-manager-7494769d95-sxpjq_openshift-route-controller-manager(6deeb990-db08-494d-af2b-5be91b4271f6)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_route-controller-manager-7494769d95-sxpjq_openshift-route-controller-manager_6deeb990-db08-494d-af2b-5be91b4271f6_0(ec88c5bccdc6dc648d068e4c3f4ad00738596c1e4b57bc550b4aaa438505bc9e): error adding pod openshift-route-controller-manager_route-controller-manager-7494769d95-sxpjq to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"ec88c5bccdc6dc648d068e4c3f4ad00738596c1e4b57bc550b4aaa438505bc9e\\\" Netns:\\\"/var/run/netns/8777e4da-ef48-4107-b9e5-1701826e5598\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-route-controller-manager;K8S_POD_NAME=route-controller-manager-7494769d95-sxpjq;K8S_POD_INFRA_CONTAINER_ID=ec88c5bccdc6dc648d068e4c3f4ad00738596c1e4b57bc550b4aaa438505bc9e;K8S_POD_UID=6deeb990-db08-494d-af2b-5be91b4271f6\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-route-controller-manager/route-controller-manager-7494769d95-sxpjq] networking: Multus: [openshift-route-controller-manager/route-controller-manager-7494769d95-sxpjq/6deeb990-db08-494d-af2b-5be91b4271f6]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod route-controller-manager-7494769d95-sxpjq in out of cluster comm: SetNetworkStatus: failed to update the pod route-controller-manager-7494769d95-sxpjq in out of cluster comm: status update failed for pod /: Get \\\"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-7494769d95-sxpjq?timeout=1m0s\\\": dial tcp 38.102.83.75:6443: connect: connection refused\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-route-controller-manager/route-controller-manager-7494769d95-sxpjq" podUID="6deeb990-db08-494d-af2b-5be91b4271f6" Mar 07 14:42:52 crc kubenswrapper[4943]: E0307 14:42:52.070083 4943 log.go:32] "RunPodSandbox from runtime service failed" err=< Mar 07 14:42:52 crc kubenswrapper[4943]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_controller-manager-86cb58c977-j8f96_openshift-controller-manager_39fa4113-e14f-48e0-9106-52cfeeeba068_0(11e95eae34aa9eb5a747d1d468eb9ef2a9fab9ee90e3d0a1277d8b059973c3c4): error adding pod openshift-controller-manager_controller-manager-86cb58c977-j8f96 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"11e95eae34aa9eb5a747d1d468eb9ef2a9fab9ee90e3d0a1277d8b059973c3c4" Netns:"/var/run/netns/c21d1534-4cd9-416d-b7f2-6c73545099fa" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-controller-manager;K8S_POD_NAME=controller-manager-86cb58c977-j8f96;K8S_POD_INFRA_CONTAINER_ID=11e95eae34aa9eb5a747d1d468eb9ef2a9fab9ee90e3d0a1277d8b059973c3c4;K8S_POD_UID=39fa4113-e14f-48e0-9106-52cfeeeba068" Path:"" ERRORED: error configuring pod [openshift-controller-manager/controller-manager-86cb58c977-j8f96] networking: Multus: [openshift-controller-manager/controller-manager-86cb58c977-j8f96/39fa4113-e14f-48e0-9106-52cfeeeba068]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod controller-manager-86cb58c977-j8f96 in out of cluster comm: SetNetworkStatus: failed to update the pod controller-manager-86cb58c977-j8f96 in out of cluster comm: status update failed for pod /: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-86cb58c977-j8f96?timeout=1m0s": dial tcp 38.102.83.75:6443: connect: connection refused Mar 07 14:42:52 crc kubenswrapper[4943]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Mar 07 14:42:52 crc kubenswrapper[4943]: > Mar 07 14:42:52 crc kubenswrapper[4943]: E0307 14:42:52.070176 4943 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Mar 07 14:42:52 crc kubenswrapper[4943]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_controller-manager-86cb58c977-j8f96_openshift-controller-manager_39fa4113-e14f-48e0-9106-52cfeeeba068_0(11e95eae34aa9eb5a747d1d468eb9ef2a9fab9ee90e3d0a1277d8b059973c3c4): error adding pod openshift-controller-manager_controller-manager-86cb58c977-j8f96 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"11e95eae34aa9eb5a747d1d468eb9ef2a9fab9ee90e3d0a1277d8b059973c3c4" Netns:"/var/run/netns/c21d1534-4cd9-416d-b7f2-6c73545099fa" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-controller-manager;K8S_POD_NAME=controller-manager-86cb58c977-j8f96;K8S_POD_INFRA_CONTAINER_ID=11e95eae34aa9eb5a747d1d468eb9ef2a9fab9ee90e3d0a1277d8b059973c3c4;K8S_POD_UID=39fa4113-e14f-48e0-9106-52cfeeeba068" Path:"" ERRORED: error configuring pod [openshift-controller-manager/controller-manager-86cb58c977-j8f96] networking: Multus: [openshift-controller-manager/controller-manager-86cb58c977-j8f96/39fa4113-e14f-48e0-9106-52cfeeeba068]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod controller-manager-86cb58c977-j8f96 in out of cluster comm: SetNetworkStatus: failed to update the pod controller-manager-86cb58c977-j8f96 in out of cluster comm: status update failed for pod /: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-86cb58c977-j8f96?timeout=1m0s": dial tcp 38.102.83.75:6443: connect: connection refused Mar 07 14:42:52 crc kubenswrapper[4943]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Mar 07 14:42:52 crc kubenswrapper[4943]: > pod="openshift-controller-manager/controller-manager-86cb58c977-j8f96" Mar 07 14:42:52 crc kubenswrapper[4943]: E0307 14:42:52.070197 4943 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Mar 07 14:42:52 crc kubenswrapper[4943]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_controller-manager-86cb58c977-j8f96_openshift-controller-manager_39fa4113-e14f-48e0-9106-52cfeeeba068_0(11e95eae34aa9eb5a747d1d468eb9ef2a9fab9ee90e3d0a1277d8b059973c3c4): error adding pod openshift-controller-manager_controller-manager-86cb58c977-j8f96 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"11e95eae34aa9eb5a747d1d468eb9ef2a9fab9ee90e3d0a1277d8b059973c3c4" Netns:"/var/run/netns/c21d1534-4cd9-416d-b7f2-6c73545099fa" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-controller-manager;K8S_POD_NAME=controller-manager-86cb58c977-j8f96;K8S_POD_INFRA_CONTAINER_ID=11e95eae34aa9eb5a747d1d468eb9ef2a9fab9ee90e3d0a1277d8b059973c3c4;K8S_POD_UID=39fa4113-e14f-48e0-9106-52cfeeeba068" Path:"" ERRORED: error configuring pod [openshift-controller-manager/controller-manager-86cb58c977-j8f96] networking: Multus: [openshift-controller-manager/controller-manager-86cb58c977-j8f96/39fa4113-e14f-48e0-9106-52cfeeeba068]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod controller-manager-86cb58c977-j8f96 in out of cluster comm: SetNetworkStatus: failed to update the pod controller-manager-86cb58c977-j8f96 in out of cluster comm: status update failed for pod /: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-86cb58c977-j8f96?timeout=1m0s": dial tcp 38.102.83.75:6443: connect: connection refused Mar 07 14:42:52 crc kubenswrapper[4943]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Mar 07 14:42:52 crc kubenswrapper[4943]: > pod="openshift-controller-manager/controller-manager-86cb58c977-j8f96" Mar 07 14:42:52 crc kubenswrapper[4943]: E0307 14:42:52.070371 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"controller-manager-86cb58c977-j8f96_openshift-controller-manager(39fa4113-e14f-48e0-9106-52cfeeeba068)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"controller-manager-86cb58c977-j8f96_openshift-controller-manager(39fa4113-e14f-48e0-9106-52cfeeeba068)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_controller-manager-86cb58c977-j8f96_openshift-controller-manager_39fa4113-e14f-48e0-9106-52cfeeeba068_0(11e95eae34aa9eb5a747d1d468eb9ef2a9fab9ee90e3d0a1277d8b059973c3c4): error adding pod openshift-controller-manager_controller-manager-86cb58c977-j8f96 to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"11e95eae34aa9eb5a747d1d468eb9ef2a9fab9ee90e3d0a1277d8b059973c3c4\\\" Netns:\\\"/var/run/netns/c21d1534-4cd9-416d-b7f2-6c73545099fa\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-controller-manager;K8S_POD_NAME=controller-manager-86cb58c977-j8f96;K8S_POD_INFRA_CONTAINER_ID=11e95eae34aa9eb5a747d1d468eb9ef2a9fab9ee90e3d0a1277d8b059973c3c4;K8S_POD_UID=39fa4113-e14f-48e0-9106-52cfeeeba068\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-controller-manager/controller-manager-86cb58c977-j8f96] networking: Multus: [openshift-controller-manager/controller-manager-86cb58c977-j8f96/39fa4113-e14f-48e0-9106-52cfeeeba068]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod controller-manager-86cb58c977-j8f96 in out of cluster comm: SetNetworkStatus: failed to update the pod controller-manager-86cb58c977-j8f96 in out of cluster comm: status update failed for pod /: Get \\\"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-86cb58c977-j8f96?timeout=1m0s\\\": dial tcp 38.102.83.75:6443: connect: connection refused\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-controller-manager/controller-manager-86cb58c977-j8f96" podUID="39fa4113-e14f-48e0-9106-52cfeeeba068" Mar 07 14:42:52 crc kubenswrapper[4943]: I0307 14:42:52.335958 4943 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:6443/readyz\": dial tcp 192.168.126.11:6443: connect: connection refused" start-of-body= Mar 07 14:42:52 crc kubenswrapper[4943]: I0307 14:42:52.336098 4943 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/readyz\": dial tcp 192.168.126.11:6443: connect: connection refused" Mar 07 14:42:52 crc kubenswrapper[4943]: I0307 14:42:52.660646 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"f1df75a5adbef71c71f3333fa02b2226668f211c29e27ab17fe30a7a46310f93"} Mar 07 14:42:52 crc kubenswrapper[4943]: I0307 14:42:52.660728 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"49c35db207dc324b23f2e574d4ed6745dbe5174671b91d83858d3aeffcf1ff1a"} Mar 07 14:42:52 crc kubenswrapper[4943]: I0307 14:42:52.661894 4943 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.75:6443: connect: connection refused" Mar 07 14:42:52 crc kubenswrapper[4943]: I0307 14:42:52.662891 4943 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.75:6443: connect: connection refused" Mar 07 14:42:52 crc kubenswrapper[4943]: I0307 14:42:52.663469 4943 status_manager.go:851] "Failed to get status for pod" podUID="979a9c5a-4f69-4b26-8127-f760a697ae91" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.75:6443: connect: connection refused" Mar 07 14:42:52 crc kubenswrapper[4943]: I0307 14:42:52.665911 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 07 14:42:52 crc kubenswrapper[4943]: I0307 14:42:52.666821 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7494769d95-sxpjq" Mar 07 14:42:52 crc kubenswrapper[4943]: I0307 14:42:52.666858 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-86cb58c977-j8f96" Mar 07 14:42:52 crc kubenswrapper[4943]: I0307 14:42:52.667512 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7494769d95-sxpjq" Mar 07 14:42:52 crc kubenswrapper[4943]: I0307 14:42:52.667527 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-86cb58c977-j8f96" Mar 07 14:42:52 crc kubenswrapper[4943]: E0307 14:42:52.996689 4943 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.75:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189a963845699772 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:42:51.750594418 +0000 UTC m=+213.702730956,LastTimestamp:2026-03-07 14:42:51.750594418 +0000 UTC m=+213.702730956,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:42:53 crc kubenswrapper[4943]: I0307 14:42:53.022247 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 07 14:42:53 crc kubenswrapper[4943]: I0307 14:42:53.022910 4943 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.75:6443: connect: connection refused" Mar 07 14:42:53 crc kubenswrapper[4943]: I0307 14:42:53.023349 4943 status_manager.go:851] "Failed to get status for pod" podUID="979a9c5a-4f69-4b26-8127-f760a697ae91" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.75:6443: connect: connection refused" Mar 07 14:42:53 crc kubenswrapper[4943]: I0307 14:42:53.209725 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/979a9c5a-4f69-4b26-8127-f760a697ae91-var-lock\") pod \"979a9c5a-4f69-4b26-8127-f760a697ae91\" (UID: \"979a9c5a-4f69-4b26-8127-f760a697ae91\") " Mar 07 14:42:53 crc kubenswrapper[4943]: I0307 14:42:53.209858 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/979a9c5a-4f69-4b26-8127-f760a697ae91-kube-api-access\") pod \"979a9c5a-4f69-4b26-8127-f760a697ae91\" (UID: \"979a9c5a-4f69-4b26-8127-f760a697ae91\") " Mar 07 14:42:53 crc kubenswrapper[4943]: I0307 14:42:53.209996 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/979a9c5a-4f69-4b26-8127-f760a697ae91-kubelet-dir\") pod \"979a9c5a-4f69-4b26-8127-f760a697ae91\" (UID: \"979a9c5a-4f69-4b26-8127-f760a697ae91\") " Mar 07 14:42:53 crc kubenswrapper[4943]: I0307 14:42:53.210055 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/979a9c5a-4f69-4b26-8127-f760a697ae91-var-lock" (OuterVolumeSpecName: "var-lock") pod "979a9c5a-4f69-4b26-8127-f760a697ae91" (UID: "979a9c5a-4f69-4b26-8127-f760a697ae91"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 07 14:42:53 crc kubenswrapper[4943]: I0307 14:42:53.210173 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/979a9c5a-4f69-4b26-8127-f760a697ae91-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "979a9c5a-4f69-4b26-8127-f760a697ae91" (UID: "979a9c5a-4f69-4b26-8127-f760a697ae91"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 07 14:42:53 crc kubenswrapper[4943]: I0307 14:42:53.211349 4943 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/979a9c5a-4f69-4b26-8127-f760a697ae91-var-lock\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:53 crc kubenswrapper[4943]: I0307 14:42:53.211411 4943 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/979a9c5a-4f69-4b26-8127-f760a697ae91-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:53 crc kubenswrapper[4943]: I0307 14:42:53.224220 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/979a9c5a-4f69-4b26-8127-f760a697ae91-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "979a9c5a-4f69-4b26-8127-f760a697ae91" (UID: "979a9c5a-4f69-4b26-8127-f760a697ae91"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:42:53 crc kubenswrapper[4943]: I0307 14:42:53.313283 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/979a9c5a-4f69-4b26-8127-f760a697ae91-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:53 crc kubenswrapper[4943]: E0307 14:42:53.592733 4943 log.go:32] "RunPodSandbox from runtime service failed" err=< Mar 07 14:42:53 crc kubenswrapper[4943]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_route-controller-manager-7494769d95-sxpjq_openshift-route-controller-manager_6deeb990-db08-494d-af2b-5be91b4271f6_0(3c9a9e8d187e4b56ba5823996173487ae9a6573c781a9843c3176df7259424c3): error adding pod openshift-route-controller-manager_route-controller-manager-7494769d95-sxpjq to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"3c9a9e8d187e4b56ba5823996173487ae9a6573c781a9843c3176df7259424c3" Netns:"/var/run/netns/afd2bc21-9993-4cad-a35c-a0801523650d" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-route-controller-manager;K8S_POD_NAME=route-controller-manager-7494769d95-sxpjq;K8S_POD_INFRA_CONTAINER_ID=3c9a9e8d187e4b56ba5823996173487ae9a6573c781a9843c3176df7259424c3;K8S_POD_UID=6deeb990-db08-494d-af2b-5be91b4271f6" Path:"" ERRORED: error configuring pod [openshift-route-controller-manager/route-controller-manager-7494769d95-sxpjq] networking: Multus: [openshift-route-controller-manager/route-controller-manager-7494769d95-sxpjq/6deeb990-db08-494d-af2b-5be91b4271f6]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod route-controller-manager-7494769d95-sxpjq in out of cluster comm: SetNetworkStatus: failed to update the pod route-controller-manager-7494769d95-sxpjq in out of cluster comm: status update failed for pod /: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-7494769d95-sxpjq?timeout=1m0s": dial tcp 38.102.83.75:6443: connect: connection refused Mar 07 14:42:53 crc kubenswrapper[4943]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Mar 07 14:42:53 crc kubenswrapper[4943]: > Mar 07 14:42:53 crc kubenswrapper[4943]: E0307 14:42:53.593161 4943 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Mar 07 14:42:53 crc kubenswrapper[4943]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_route-controller-manager-7494769d95-sxpjq_openshift-route-controller-manager_6deeb990-db08-494d-af2b-5be91b4271f6_0(3c9a9e8d187e4b56ba5823996173487ae9a6573c781a9843c3176df7259424c3): error adding pod openshift-route-controller-manager_route-controller-manager-7494769d95-sxpjq to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"3c9a9e8d187e4b56ba5823996173487ae9a6573c781a9843c3176df7259424c3" Netns:"/var/run/netns/afd2bc21-9993-4cad-a35c-a0801523650d" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-route-controller-manager;K8S_POD_NAME=route-controller-manager-7494769d95-sxpjq;K8S_POD_INFRA_CONTAINER_ID=3c9a9e8d187e4b56ba5823996173487ae9a6573c781a9843c3176df7259424c3;K8S_POD_UID=6deeb990-db08-494d-af2b-5be91b4271f6" Path:"" ERRORED: error configuring pod [openshift-route-controller-manager/route-controller-manager-7494769d95-sxpjq] networking: Multus: [openshift-route-controller-manager/route-controller-manager-7494769d95-sxpjq/6deeb990-db08-494d-af2b-5be91b4271f6]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod route-controller-manager-7494769d95-sxpjq in out of cluster comm: SetNetworkStatus: failed to update the pod route-controller-manager-7494769d95-sxpjq in out of cluster comm: status update failed for pod /: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-7494769d95-sxpjq?timeout=1m0s": dial tcp 38.102.83.75:6443: connect: connection refused Mar 07 14:42:53 crc kubenswrapper[4943]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Mar 07 14:42:53 crc kubenswrapper[4943]: > pod="openshift-route-controller-manager/route-controller-manager-7494769d95-sxpjq" Mar 07 14:42:53 crc kubenswrapper[4943]: E0307 14:42:53.593187 4943 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Mar 07 14:42:53 crc kubenswrapper[4943]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_route-controller-manager-7494769d95-sxpjq_openshift-route-controller-manager_6deeb990-db08-494d-af2b-5be91b4271f6_0(3c9a9e8d187e4b56ba5823996173487ae9a6573c781a9843c3176df7259424c3): error adding pod openshift-route-controller-manager_route-controller-manager-7494769d95-sxpjq to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"3c9a9e8d187e4b56ba5823996173487ae9a6573c781a9843c3176df7259424c3" Netns:"/var/run/netns/afd2bc21-9993-4cad-a35c-a0801523650d" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-route-controller-manager;K8S_POD_NAME=route-controller-manager-7494769d95-sxpjq;K8S_POD_INFRA_CONTAINER_ID=3c9a9e8d187e4b56ba5823996173487ae9a6573c781a9843c3176df7259424c3;K8S_POD_UID=6deeb990-db08-494d-af2b-5be91b4271f6" Path:"" ERRORED: error configuring pod [openshift-route-controller-manager/route-controller-manager-7494769d95-sxpjq] networking: Multus: [openshift-route-controller-manager/route-controller-manager-7494769d95-sxpjq/6deeb990-db08-494d-af2b-5be91b4271f6]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod route-controller-manager-7494769d95-sxpjq in out of cluster comm: SetNetworkStatus: failed to update the pod route-controller-manager-7494769d95-sxpjq in out of cluster comm: status update failed for pod /: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-7494769d95-sxpjq?timeout=1m0s": dial tcp 38.102.83.75:6443: connect: connection refused Mar 07 14:42:53 crc kubenswrapper[4943]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Mar 07 14:42:53 crc kubenswrapper[4943]: > pod="openshift-route-controller-manager/route-controller-manager-7494769d95-sxpjq" Mar 07 14:42:53 crc kubenswrapper[4943]: E0307 14:42:53.593261 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"route-controller-manager-7494769d95-sxpjq_openshift-route-controller-manager(6deeb990-db08-494d-af2b-5be91b4271f6)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"route-controller-manager-7494769d95-sxpjq_openshift-route-controller-manager(6deeb990-db08-494d-af2b-5be91b4271f6)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_route-controller-manager-7494769d95-sxpjq_openshift-route-controller-manager_6deeb990-db08-494d-af2b-5be91b4271f6_0(3c9a9e8d187e4b56ba5823996173487ae9a6573c781a9843c3176df7259424c3): error adding pod openshift-route-controller-manager_route-controller-manager-7494769d95-sxpjq to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"3c9a9e8d187e4b56ba5823996173487ae9a6573c781a9843c3176df7259424c3\\\" Netns:\\\"/var/run/netns/afd2bc21-9993-4cad-a35c-a0801523650d\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-route-controller-manager;K8S_POD_NAME=route-controller-manager-7494769d95-sxpjq;K8S_POD_INFRA_CONTAINER_ID=3c9a9e8d187e4b56ba5823996173487ae9a6573c781a9843c3176df7259424c3;K8S_POD_UID=6deeb990-db08-494d-af2b-5be91b4271f6\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-route-controller-manager/route-controller-manager-7494769d95-sxpjq] networking: Multus: [openshift-route-controller-manager/route-controller-manager-7494769d95-sxpjq/6deeb990-db08-494d-af2b-5be91b4271f6]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod route-controller-manager-7494769d95-sxpjq in out of cluster comm: SetNetworkStatus: failed to update the pod route-controller-manager-7494769d95-sxpjq in out of cluster comm: status update failed for pod /: Get \\\"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-7494769d95-sxpjq?timeout=1m0s\\\": dial tcp 38.102.83.75:6443: connect: connection refused\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-route-controller-manager/route-controller-manager-7494769d95-sxpjq" podUID="6deeb990-db08-494d-af2b-5be91b4271f6" Mar 07 14:42:53 crc kubenswrapper[4943]: E0307 14:42:53.664074 4943 log.go:32] "RunPodSandbox from runtime service failed" err=< Mar 07 14:42:53 crc kubenswrapper[4943]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_controller-manager-86cb58c977-j8f96_openshift-controller-manager_39fa4113-e14f-48e0-9106-52cfeeeba068_0(a227c10cead5247446cf4478e191053397273b8a834b69da6265db9ee5dea58a): error adding pod openshift-controller-manager_controller-manager-86cb58c977-j8f96 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"a227c10cead5247446cf4478e191053397273b8a834b69da6265db9ee5dea58a" Netns:"/var/run/netns/13a205b8-45f5-4f53-9eee-2f8e6a0b9a75" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-controller-manager;K8S_POD_NAME=controller-manager-86cb58c977-j8f96;K8S_POD_INFRA_CONTAINER_ID=a227c10cead5247446cf4478e191053397273b8a834b69da6265db9ee5dea58a;K8S_POD_UID=39fa4113-e14f-48e0-9106-52cfeeeba068" Path:"" ERRORED: error configuring pod [openshift-controller-manager/controller-manager-86cb58c977-j8f96] networking: Multus: [openshift-controller-manager/controller-manager-86cb58c977-j8f96/39fa4113-e14f-48e0-9106-52cfeeeba068]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod controller-manager-86cb58c977-j8f96 in out of cluster comm: SetNetworkStatus: failed to update the pod controller-manager-86cb58c977-j8f96 in out of cluster comm: status update failed for pod /: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-86cb58c977-j8f96?timeout=1m0s": dial tcp 38.102.83.75:6443: connect: connection refused Mar 07 14:42:53 crc kubenswrapper[4943]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Mar 07 14:42:53 crc kubenswrapper[4943]: > Mar 07 14:42:53 crc kubenswrapper[4943]: E0307 14:42:53.664143 4943 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Mar 07 14:42:53 crc kubenswrapper[4943]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_controller-manager-86cb58c977-j8f96_openshift-controller-manager_39fa4113-e14f-48e0-9106-52cfeeeba068_0(a227c10cead5247446cf4478e191053397273b8a834b69da6265db9ee5dea58a): error adding pod openshift-controller-manager_controller-manager-86cb58c977-j8f96 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"a227c10cead5247446cf4478e191053397273b8a834b69da6265db9ee5dea58a" Netns:"/var/run/netns/13a205b8-45f5-4f53-9eee-2f8e6a0b9a75" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-controller-manager;K8S_POD_NAME=controller-manager-86cb58c977-j8f96;K8S_POD_INFRA_CONTAINER_ID=a227c10cead5247446cf4478e191053397273b8a834b69da6265db9ee5dea58a;K8S_POD_UID=39fa4113-e14f-48e0-9106-52cfeeeba068" Path:"" ERRORED: error configuring pod [openshift-controller-manager/controller-manager-86cb58c977-j8f96] networking: Multus: [openshift-controller-manager/controller-manager-86cb58c977-j8f96/39fa4113-e14f-48e0-9106-52cfeeeba068]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod controller-manager-86cb58c977-j8f96 in out of cluster comm: SetNetworkStatus: failed to update the pod controller-manager-86cb58c977-j8f96 in out of cluster comm: status update failed for pod /: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-86cb58c977-j8f96?timeout=1m0s": dial tcp 38.102.83.75:6443: connect: connection refused Mar 07 14:42:53 crc kubenswrapper[4943]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Mar 07 14:42:53 crc kubenswrapper[4943]: > pod="openshift-controller-manager/controller-manager-86cb58c977-j8f96" Mar 07 14:42:53 crc kubenswrapper[4943]: E0307 14:42:53.664164 4943 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Mar 07 14:42:53 crc kubenswrapper[4943]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_controller-manager-86cb58c977-j8f96_openshift-controller-manager_39fa4113-e14f-48e0-9106-52cfeeeba068_0(a227c10cead5247446cf4478e191053397273b8a834b69da6265db9ee5dea58a): error adding pod openshift-controller-manager_controller-manager-86cb58c977-j8f96 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"a227c10cead5247446cf4478e191053397273b8a834b69da6265db9ee5dea58a" Netns:"/var/run/netns/13a205b8-45f5-4f53-9eee-2f8e6a0b9a75" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-controller-manager;K8S_POD_NAME=controller-manager-86cb58c977-j8f96;K8S_POD_INFRA_CONTAINER_ID=a227c10cead5247446cf4478e191053397273b8a834b69da6265db9ee5dea58a;K8S_POD_UID=39fa4113-e14f-48e0-9106-52cfeeeba068" Path:"" ERRORED: error configuring pod [openshift-controller-manager/controller-manager-86cb58c977-j8f96] networking: Multus: [openshift-controller-manager/controller-manager-86cb58c977-j8f96/39fa4113-e14f-48e0-9106-52cfeeeba068]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod controller-manager-86cb58c977-j8f96 in out of cluster comm: SetNetworkStatus: failed to update the pod controller-manager-86cb58c977-j8f96 in out of cluster comm: status update failed for pod /: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-86cb58c977-j8f96?timeout=1m0s": dial tcp 38.102.83.75:6443: connect: connection refused Mar 07 14:42:53 crc kubenswrapper[4943]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Mar 07 14:42:53 crc kubenswrapper[4943]: > pod="openshift-controller-manager/controller-manager-86cb58c977-j8f96" Mar 07 14:42:53 crc kubenswrapper[4943]: E0307 14:42:53.664218 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"controller-manager-86cb58c977-j8f96_openshift-controller-manager(39fa4113-e14f-48e0-9106-52cfeeeba068)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"controller-manager-86cb58c977-j8f96_openshift-controller-manager(39fa4113-e14f-48e0-9106-52cfeeeba068)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_controller-manager-86cb58c977-j8f96_openshift-controller-manager_39fa4113-e14f-48e0-9106-52cfeeeba068_0(a227c10cead5247446cf4478e191053397273b8a834b69da6265db9ee5dea58a): error adding pod openshift-controller-manager_controller-manager-86cb58c977-j8f96 to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"a227c10cead5247446cf4478e191053397273b8a834b69da6265db9ee5dea58a\\\" Netns:\\\"/var/run/netns/13a205b8-45f5-4f53-9eee-2f8e6a0b9a75\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-controller-manager;K8S_POD_NAME=controller-manager-86cb58c977-j8f96;K8S_POD_INFRA_CONTAINER_ID=a227c10cead5247446cf4478e191053397273b8a834b69da6265db9ee5dea58a;K8S_POD_UID=39fa4113-e14f-48e0-9106-52cfeeeba068\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-controller-manager/controller-manager-86cb58c977-j8f96] networking: Multus: [openshift-controller-manager/controller-manager-86cb58c977-j8f96/39fa4113-e14f-48e0-9106-52cfeeeba068]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod controller-manager-86cb58c977-j8f96 in out of cluster comm: SetNetworkStatus: failed to update the pod controller-manager-86cb58c977-j8f96 in out of cluster comm: status update failed for pod /: Get \\\"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-86cb58c977-j8f96?timeout=1m0s\\\": dial tcp 38.102.83.75:6443: connect: connection refused\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-controller-manager/controller-manager-86cb58c977-j8f96" podUID="39fa4113-e14f-48e0-9106-52cfeeeba068" Mar 07 14:42:53 crc kubenswrapper[4943]: I0307 14:42:53.679500 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 07 14:42:53 crc kubenswrapper[4943]: I0307 14:42:53.682780 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"979a9c5a-4f69-4b26-8127-f760a697ae91","Type":"ContainerDied","Data":"b5d195dd1186d19e7481f63b38ef59140069d31599dbe38908f4898eae4c0701"} Mar 07 14:42:53 crc kubenswrapper[4943]: I0307 14:42:53.682852 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b5d195dd1186d19e7481f63b38ef59140069d31599dbe38908f4898eae4c0701" Mar 07 14:42:53 crc kubenswrapper[4943]: I0307 14:42:53.747100 4943 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.75:6443: connect: connection refused" Mar 07 14:42:53 crc kubenswrapper[4943]: I0307 14:42:53.747669 4943 status_manager.go:851] "Failed to get status for pod" podUID="979a9c5a-4f69-4b26-8127-f760a697ae91" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.75:6443: connect: connection refused" Mar 07 14:42:53 crc kubenswrapper[4943]: I0307 14:42:53.761047 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 07 14:42:53 crc kubenswrapper[4943]: I0307 14:42:53.761980 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 07 14:42:53 crc kubenswrapper[4943]: I0307 14:42:53.762435 4943 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.75:6443: connect: connection refused" Mar 07 14:42:53 crc kubenswrapper[4943]: I0307 14:42:53.763000 4943 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.75:6443: connect: connection refused" Mar 07 14:42:53 crc kubenswrapper[4943]: I0307 14:42:53.763411 4943 status_manager.go:851] "Failed to get status for pod" podUID="979a9c5a-4f69-4b26-8127-f760a697ae91" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.75:6443: connect: connection refused" Mar 07 14:42:53 crc kubenswrapper[4943]: I0307 14:42:53.923237 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 07 14:42:53 crc kubenswrapper[4943]: I0307 14:42:53.923318 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 07 14:42:53 crc kubenswrapper[4943]: I0307 14:42:53.923424 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 07 14:42:53 crc kubenswrapper[4943]: I0307 14:42:53.923438 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 07 14:42:53 crc kubenswrapper[4943]: I0307 14:42:53.923499 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 07 14:42:53 crc kubenswrapper[4943]: I0307 14:42:53.923568 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 07 14:42:53 crc kubenswrapper[4943]: I0307 14:42:53.924128 4943 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:53 crc kubenswrapper[4943]: I0307 14:42:53.924159 4943 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:53 crc kubenswrapper[4943]: I0307 14:42:53.924179 4943 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 07 14:42:54 crc kubenswrapper[4943]: I0307 14:42:54.690585 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 07 14:42:54 crc kubenswrapper[4943]: I0307 14:42:54.691367 4943 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="6084025a1e017c5e63892cdc0cc359ad0297878a6eb80cbf8d4d620bb5f2f681" exitCode=0 Mar 07 14:42:54 crc kubenswrapper[4943]: I0307 14:42:54.691453 4943 scope.go:117] "RemoveContainer" containerID="36053e9d136906da3b2eb3d258bf5b73bb956ea6141497e74f1f18c8f9a937a9" Mar 07 14:42:54 crc kubenswrapper[4943]: I0307 14:42:54.691493 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 07 14:42:54 crc kubenswrapper[4943]: I0307 14:42:54.711430 4943 scope.go:117] "RemoveContainer" containerID="69d5ad495f66b4bdb2e419c05c8a8060bfc94df1a9c5cb3156bed86ba312562b" Mar 07 14:42:54 crc kubenswrapper[4943]: I0307 14:42:54.721659 4943 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.75:6443: connect: connection refused" Mar 07 14:42:54 crc kubenswrapper[4943]: I0307 14:42:54.722031 4943 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.75:6443: connect: connection refused" Mar 07 14:42:54 crc kubenswrapper[4943]: I0307 14:42:54.722400 4943 status_manager.go:851] "Failed to get status for pod" podUID="979a9c5a-4f69-4b26-8127-f760a697ae91" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.75:6443: connect: connection refused" Mar 07 14:42:54 crc kubenswrapper[4943]: I0307 14:42:54.735913 4943 scope.go:117] "RemoveContainer" containerID="5f9507997ae006ce9864d88b639c5c782aaed107dbd59b42ea0bb34e1606bf0c" Mar 07 14:42:54 crc kubenswrapper[4943]: I0307 14:42:54.758496 4943 scope.go:117] "RemoveContainer" containerID="46a2cc9aec0e9e4a54e8fa06e6eeaec8ba29e88d0a0672c00f9a01b439ad4331" Mar 07 14:42:54 crc kubenswrapper[4943]: I0307 14:42:54.763578 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Mar 07 14:42:54 crc kubenswrapper[4943]: I0307 14:42:54.777678 4943 scope.go:117] "RemoveContainer" containerID="6084025a1e017c5e63892cdc0cc359ad0297878a6eb80cbf8d4d620bb5f2f681" Mar 07 14:42:54 crc kubenswrapper[4943]: I0307 14:42:54.800797 4943 scope.go:117] "RemoveContainer" containerID="87c2d7e4180bc1e9383182a52ba43dd30e41d4e08dd229a29f831abe304722a4" Mar 07 14:42:54 crc kubenswrapper[4943]: I0307 14:42:54.832613 4943 scope.go:117] "RemoveContainer" containerID="36053e9d136906da3b2eb3d258bf5b73bb956ea6141497e74f1f18c8f9a937a9" Mar 07 14:42:54 crc kubenswrapper[4943]: E0307 14:42:54.833009 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36053e9d136906da3b2eb3d258bf5b73bb956ea6141497e74f1f18c8f9a937a9\": container with ID starting with 36053e9d136906da3b2eb3d258bf5b73bb956ea6141497e74f1f18c8f9a937a9 not found: ID does not exist" containerID="36053e9d136906da3b2eb3d258bf5b73bb956ea6141497e74f1f18c8f9a937a9" Mar 07 14:42:54 crc kubenswrapper[4943]: I0307 14:42:54.833039 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36053e9d136906da3b2eb3d258bf5b73bb956ea6141497e74f1f18c8f9a937a9"} err="failed to get container status \"36053e9d136906da3b2eb3d258bf5b73bb956ea6141497e74f1f18c8f9a937a9\": rpc error: code = NotFound desc = could not find container \"36053e9d136906da3b2eb3d258bf5b73bb956ea6141497e74f1f18c8f9a937a9\": container with ID starting with 36053e9d136906da3b2eb3d258bf5b73bb956ea6141497e74f1f18c8f9a937a9 not found: ID does not exist" Mar 07 14:42:54 crc kubenswrapper[4943]: I0307 14:42:54.833068 4943 scope.go:117] "RemoveContainer" containerID="69d5ad495f66b4bdb2e419c05c8a8060bfc94df1a9c5cb3156bed86ba312562b" Mar 07 14:42:54 crc kubenswrapper[4943]: E0307 14:42:54.833421 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69d5ad495f66b4bdb2e419c05c8a8060bfc94df1a9c5cb3156bed86ba312562b\": container with ID starting with 69d5ad495f66b4bdb2e419c05c8a8060bfc94df1a9c5cb3156bed86ba312562b not found: ID does not exist" containerID="69d5ad495f66b4bdb2e419c05c8a8060bfc94df1a9c5cb3156bed86ba312562b" Mar 07 14:42:54 crc kubenswrapper[4943]: I0307 14:42:54.833474 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69d5ad495f66b4bdb2e419c05c8a8060bfc94df1a9c5cb3156bed86ba312562b"} err="failed to get container status \"69d5ad495f66b4bdb2e419c05c8a8060bfc94df1a9c5cb3156bed86ba312562b\": rpc error: code = NotFound desc = could not find container \"69d5ad495f66b4bdb2e419c05c8a8060bfc94df1a9c5cb3156bed86ba312562b\": container with ID starting with 69d5ad495f66b4bdb2e419c05c8a8060bfc94df1a9c5cb3156bed86ba312562b not found: ID does not exist" Mar 07 14:42:54 crc kubenswrapper[4943]: I0307 14:42:54.833743 4943 scope.go:117] "RemoveContainer" containerID="5f9507997ae006ce9864d88b639c5c782aaed107dbd59b42ea0bb34e1606bf0c" Mar 07 14:42:54 crc kubenswrapper[4943]: E0307 14:42:54.834360 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f9507997ae006ce9864d88b639c5c782aaed107dbd59b42ea0bb34e1606bf0c\": container with ID starting with 5f9507997ae006ce9864d88b639c5c782aaed107dbd59b42ea0bb34e1606bf0c not found: ID does not exist" containerID="5f9507997ae006ce9864d88b639c5c782aaed107dbd59b42ea0bb34e1606bf0c" Mar 07 14:42:54 crc kubenswrapper[4943]: I0307 14:42:54.834392 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f9507997ae006ce9864d88b639c5c782aaed107dbd59b42ea0bb34e1606bf0c"} err="failed to get container status \"5f9507997ae006ce9864d88b639c5c782aaed107dbd59b42ea0bb34e1606bf0c\": rpc error: code = NotFound desc = could not find container \"5f9507997ae006ce9864d88b639c5c782aaed107dbd59b42ea0bb34e1606bf0c\": container with ID starting with 5f9507997ae006ce9864d88b639c5c782aaed107dbd59b42ea0bb34e1606bf0c not found: ID does not exist" Mar 07 14:42:54 crc kubenswrapper[4943]: I0307 14:42:54.834407 4943 scope.go:117] "RemoveContainer" containerID="46a2cc9aec0e9e4a54e8fa06e6eeaec8ba29e88d0a0672c00f9a01b439ad4331" Mar 07 14:42:54 crc kubenswrapper[4943]: E0307 14:42:54.834773 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46a2cc9aec0e9e4a54e8fa06e6eeaec8ba29e88d0a0672c00f9a01b439ad4331\": container with ID starting with 46a2cc9aec0e9e4a54e8fa06e6eeaec8ba29e88d0a0672c00f9a01b439ad4331 not found: ID does not exist" containerID="46a2cc9aec0e9e4a54e8fa06e6eeaec8ba29e88d0a0672c00f9a01b439ad4331" Mar 07 14:42:54 crc kubenswrapper[4943]: I0307 14:42:54.834856 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46a2cc9aec0e9e4a54e8fa06e6eeaec8ba29e88d0a0672c00f9a01b439ad4331"} err="failed to get container status \"46a2cc9aec0e9e4a54e8fa06e6eeaec8ba29e88d0a0672c00f9a01b439ad4331\": rpc error: code = NotFound desc = could not find container \"46a2cc9aec0e9e4a54e8fa06e6eeaec8ba29e88d0a0672c00f9a01b439ad4331\": container with ID starting with 46a2cc9aec0e9e4a54e8fa06e6eeaec8ba29e88d0a0672c00f9a01b439ad4331 not found: ID does not exist" Mar 07 14:42:54 crc kubenswrapper[4943]: I0307 14:42:54.834888 4943 scope.go:117] "RemoveContainer" containerID="6084025a1e017c5e63892cdc0cc359ad0297878a6eb80cbf8d4d620bb5f2f681" Mar 07 14:42:54 crc kubenswrapper[4943]: E0307 14:42:54.835987 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6084025a1e017c5e63892cdc0cc359ad0297878a6eb80cbf8d4d620bb5f2f681\": container with ID starting with 6084025a1e017c5e63892cdc0cc359ad0297878a6eb80cbf8d4d620bb5f2f681 not found: ID does not exist" containerID="6084025a1e017c5e63892cdc0cc359ad0297878a6eb80cbf8d4d620bb5f2f681" Mar 07 14:42:54 crc kubenswrapper[4943]: E0307 14:42:54.836043 4943 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.75:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" volumeName="registry-storage" Mar 07 14:42:54 crc kubenswrapper[4943]: I0307 14:42:54.836060 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6084025a1e017c5e63892cdc0cc359ad0297878a6eb80cbf8d4d620bb5f2f681"} err="failed to get container status \"6084025a1e017c5e63892cdc0cc359ad0297878a6eb80cbf8d4d620bb5f2f681\": rpc error: code = NotFound desc = could not find container \"6084025a1e017c5e63892cdc0cc359ad0297878a6eb80cbf8d4d620bb5f2f681\": container with ID starting with 6084025a1e017c5e63892cdc0cc359ad0297878a6eb80cbf8d4d620bb5f2f681 not found: ID does not exist" Mar 07 14:42:54 crc kubenswrapper[4943]: I0307 14:42:54.836216 4943 scope.go:117] "RemoveContainer" containerID="87c2d7e4180bc1e9383182a52ba43dd30e41d4e08dd229a29f831abe304722a4" Mar 07 14:42:54 crc kubenswrapper[4943]: E0307 14:42:54.837203 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87c2d7e4180bc1e9383182a52ba43dd30e41d4e08dd229a29f831abe304722a4\": container with ID starting with 87c2d7e4180bc1e9383182a52ba43dd30e41d4e08dd229a29f831abe304722a4 not found: ID does not exist" containerID="87c2d7e4180bc1e9383182a52ba43dd30e41d4e08dd229a29f831abe304722a4" Mar 07 14:42:54 crc kubenswrapper[4943]: I0307 14:42:54.837249 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87c2d7e4180bc1e9383182a52ba43dd30e41d4e08dd229a29f831abe304722a4"} err="failed to get container status \"87c2d7e4180bc1e9383182a52ba43dd30e41d4e08dd229a29f831abe304722a4\": rpc error: code = NotFound desc = could not find container \"87c2d7e4180bc1e9383182a52ba43dd30e41d4e08dd229a29f831abe304722a4\": container with ID starting with 87c2d7e4180bc1e9383182a52ba43dd30e41d4e08dd229a29f831abe304722a4 not found: ID does not exist" Mar 07 14:42:58 crc kubenswrapper[4943]: E0307 14:42:58.681807 4943 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.75:6443: connect: connection refused" Mar 07 14:42:58 crc kubenswrapper[4943]: E0307 14:42:58.683767 4943 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.75:6443: connect: connection refused" Mar 07 14:42:58 crc kubenswrapper[4943]: E0307 14:42:58.684410 4943 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.75:6443: connect: connection refused" Mar 07 14:42:58 crc kubenswrapper[4943]: E0307 14:42:58.685143 4943 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.75:6443: connect: connection refused" Mar 07 14:42:58 crc kubenswrapper[4943]: E0307 14:42:58.685446 4943 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.75:6443: connect: connection refused" Mar 07 14:42:58 crc kubenswrapper[4943]: I0307 14:42:58.685561 4943 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Mar 07 14:42:58 crc kubenswrapper[4943]: E0307 14:42:58.685828 4943 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.75:6443: connect: connection refused" interval="200ms" Mar 07 14:42:58 crc kubenswrapper[4943]: I0307 14:42:58.762254 4943 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.75:6443: connect: connection refused" Mar 07 14:42:58 crc kubenswrapper[4943]: I0307 14:42:58.762914 4943 status_manager.go:851] "Failed to get status for pod" podUID="979a9c5a-4f69-4b26-8127-f760a697ae91" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.75:6443: connect: connection refused" Mar 07 14:42:58 crc kubenswrapper[4943]: E0307 14:42:58.886857 4943 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.75:6443: connect: connection refused" interval="400ms" Mar 07 14:42:59 crc kubenswrapper[4943]: E0307 14:42:59.288108 4943 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.75:6443: connect: connection refused" interval="800ms" Mar 07 14:43:00 crc kubenswrapper[4943]: E0307 14:43:00.089617 4943 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.75:6443: connect: connection refused" interval="1.6s" Mar 07 14:43:01 crc kubenswrapper[4943]: E0307 14:43:01.695284 4943 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.75:6443: connect: connection refused" interval="3.2s" Mar 07 14:43:02 crc kubenswrapper[4943]: I0307 14:43:02.755409 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 07 14:43:02 crc kubenswrapper[4943]: I0307 14:43:02.757969 4943 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.75:6443: connect: connection refused" Mar 07 14:43:02 crc kubenswrapper[4943]: I0307 14:43:02.758589 4943 status_manager.go:851] "Failed to get status for pod" podUID="979a9c5a-4f69-4b26-8127-f760a697ae91" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.75:6443: connect: connection refused" Mar 07 14:43:02 crc kubenswrapper[4943]: I0307 14:43:02.777566 4943 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="107a7da3-71d2-4990-815e-9906bb3ea525" Mar 07 14:43:02 crc kubenswrapper[4943]: I0307 14:43:02.777633 4943 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="107a7da3-71d2-4990-815e-9906bb3ea525" Mar 07 14:43:02 crc kubenswrapper[4943]: E0307 14:43:02.778400 4943 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.75:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 07 14:43:02 crc kubenswrapper[4943]: I0307 14:43:02.779351 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 07 14:43:02 crc kubenswrapper[4943]: W0307 14:43:02.812539 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-5c76a812ee8911d045d6d47c713765bd577e7d07bd63ef3b80f9d61bc3cd502f WatchSource:0}: Error finding container 5c76a812ee8911d045d6d47c713765bd577e7d07bd63ef3b80f9d61bc3cd502f: Status 404 returned error can't find the container with id 5c76a812ee8911d045d6d47c713765bd577e7d07bd63ef3b80f9d61bc3cd502f Mar 07 14:43:02 crc kubenswrapper[4943]: E0307 14:43:02.997956 4943 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.75:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189a963845699772 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-07 14:42:51.750594418 +0000 UTC m=+213.702730956,LastTimestamp:2026-03-07 14:42:51.750594418 +0000 UTC m=+213.702730956,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 07 14:43:03 crc kubenswrapper[4943]: I0307 14:43:03.378811 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" podUID="31ce0220-0f53-4d63-aa7b-99357e9c6b04" containerName="oauth-openshift" containerID="cri-o://64735be97cb4b56fae79aae00df6781ca6b2bc8a37a772cbedb7b9239d45514d" gracePeriod=15 Mar 07 14:43:03 crc kubenswrapper[4943]: I0307 14:43:03.754840 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"ebc2a89835d6841bd089097eac67f8ac5a88e800e8e84bff62aed296856897fd"} Mar 07 14:43:03 crc kubenswrapper[4943]: I0307 14:43:03.755628 4943 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="107a7da3-71d2-4990-815e-9906bb3ea525" Mar 07 14:43:03 crc kubenswrapper[4943]: I0307 14:43:03.755654 4943 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="107a7da3-71d2-4990-815e-9906bb3ea525" Mar 07 14:43:03 crc kubenswrapper[4943]: I0307 14:43:03.754788 4943 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="ebc2a89835d6841bd089097eac67f8ac5a88e800e8e84bff62aed296856897fd" exitCode=0 Mar 07 14:43:03 crc kubenswrapper[4943]: I0307 14:43:03.756201 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"5c76a812ee8911d045d6d47c713765bd577e7d07bd63ef3b80f9d61bc3cd502f"} Mar 07 14:43:03 crc kubenswrapper[4943]: E0307 14:43:03.756273 4943 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.75:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 07 14:43:03 crc kubenswrapper[4943]: I0307 14:43:03.756337 4943 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.75:6443: connect: connection refused" Mar 07 14:43:03 crc kubenswrapper[4943]: I0307 14:43:03.756837 4943 status_manager.go:851] "Failed to get status for pod" podUID="979a9c5a-4f69-4b26-8127-f760a697ae91" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.75:6443: connect: connection refused" Mar 07 14:43:03 crc kubenswrapper[4943]: I0307 14:43:03.759157 4943 generic.go:334] "Generic (PLEG): container finished" podID="31ce0220-0f53-4d63-aa7b-99357e9c6b04" containerID="64735be97cb4b56fae79aae00df6781ca6b2bc8a37a772cbedb7b9239d45514d" exitCode=0 Mar 07 14:43:03 crc kubenswrapper[4943]: I0307 14:43:03.759212 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" event={"ID":"31ce0220-0f53-4d63-aa7b-99357e9c6b04","Type":"ContainerDied","Data":"64735be97cb4b56fae79aae00df6781ca6b2bc8a37a772cbedb7b9239d45514d"} Mar 07 14:43:03 crc kubenswrapper[4943]: I0307 14:43:03.906807 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:43:03 crc kubenswrapper[4943]: I0307 14:43:03.907442 4943 status_manager.go:851] "Failed to get status for pod" podUID="31ce0220-0f53-4d63-aa7b-99357e9c6b04" pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-cnbpx\": dial tcp 38.102.83.75:6443: connect: connection refused" Mar 07 14:43:03 crc kubenswrapper[4943]: I0307 14:43:03.907753 4943 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.75:6443: connect: connection refused" Mar 07 14:43:03 crc kubenswrapper[4943]: I0307 14:43:03.908276 4943 status_manager.go:851] "Failed to get status for pod" podUID="979a9c5a-4f69-4b26-8127-f760a697ae91" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.75:6443: connect: connection refused" Mar 07 14:43:04 crc kubenswrapper[4943]: I0307 14:43:04.057486 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-router-certs\") pod \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " Mar 07 14:43:04 crc kubenswrapper[4943]: I0307 14:43:04.057547 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-user-template-provider-selection\") pod \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " Mar 07 14:43:04 crc kubenswrapper[4943]: I0307 14:43:04.057595 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4pr5v\" (UniqueName: \"kubernetes.io/projected/31ce0220-0f53-4d63-aa7b-99357e9c6b04-kube-api-access-4pr5v\") pod \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " Mar 07 14:43:04 crc kubenswrapper[4943]: I0307 14:43:04.057642 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/31ce0220-0f53-4d63-aa7b-99357e9c6b04-audit-policies\") pod \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " Mar 07 14:43:04 crc kubenswrapper[4943]: I0307 14:43:04.057673 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-session\") pod \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " Mar 07 14:43:04 crc kubenswrapper[4943]: I0307 14:43:04.057729 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-user-idp-0-file-data\") pod \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " Mar 07 14:43:04 crc kubenswrapper[4943]: I0307 14:43:04.057768 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-serving-cert\") pod \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " Mar 07 14:43:04 crc kubenswrapper[4943]: I0307 14:43:04.057809 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/31ce0220-0f53-4d63-aa7b-99357e9c6b04-audit-dir\") pod \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " Mar 07 14:43:04 crc kubenswrapper[4943]: I0307 14:43:04.057844 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-cliconfig\") pod \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " Mar 07 14:43:04 crc kubenswrapper[4943]: I0307 14:43:04.057896 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-service-ca\") pod \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " Mar 07 14:43:04 crc kubenswrapper[4943]: I0307 14:43:04.057960 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-user-template-login\") pod \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " Mar 07 14:43:04 crc kubenswrapper[4943]: I0307 14:43:04.057994 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-trusted-ca-bundle\") pod \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " Mar 07 14:43:04 crc kubenswrapper[4943]: I0307 14:43:04.058033 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-user-template-error\") pod \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " Mar 07 14:43:04 crc kubenswrapper[4943]: I0307 14:43:04.058069 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-ocp-branding-template\") pod \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\" (UID: \"31ce0220-0f53-4d63-aa7b-99357e9c6b04\") " Mar 07 14:43:04 crc kubenswrapper[4943]: I0307 14:43:04.058604 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/31ce0220-0f53-4d63-aa7b-99357e9c6b04-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "31ce0220-0f53-4d63-aa7b-99357e9c6b04" (UID: "31ce0220-0f53-4d63-aa7b-99357e9c6b04"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 07 14:43:04 crc kubenswrapper[4943]: I0307 14:43:04.060242 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31ce0220-0f53-4d63-aa7b-99357e9c6b04-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "31ce0220-0f53-4d63-aa7b-99357e9c6b04" (UID: "31ce0220-0f53-4d63-aa7b-99357e9c6b04"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:43:04 crc kubenswrapper[4943]: I0307 14:43:04.060301 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "31ce0220-0f53-4d63-aa7b-99357e9c6b04" (UID: "31ce0220-0f53-4d63-aa7b-99357e9c6b04"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:43:04 crc kubenswrapper[4943]: I0307 14:43:04.060505 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "31ce0220-0f53-4d63-aa7b-99357e9c6b04" (UID: "31ce0220-0f53-4d63-aa7b-99357e9c6b04"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:43:04 crc kubenswrapper[4943]: I0307 14:43:04.061225 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "31ce0220-0f53-4d63-aa7b-99357e9c6b04" (UID: "31ce0220-0f53-4d63-aa7b-99357e9c6b04"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:43:04 crc kubenswrapper[4943]: I0307 14:43:04.065973 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31ce0220-0f53-4d63-aa7b-99357e9c6b04-kube-api-access-4pr5v" (OuterVolumeSpecName: "kube-api-access-4pr5v") pod "31ce0220-0f53-4d63-aa7b-99357e9c6b04" (UID: "31ce0220-0f53-4d63-aa7b-99357e9c6b04"). InnerVolumeSpecName "kube-api-access-4pr5v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:43:04 crc kubenswrapper[4943]: I0307 14:43:04.067169 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "31ce0220-0f53-4d63-aa7b-99357e9c6b04" (UID: "31ce0220-0f53-4d63-aa7b-99357e9c6b04"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:43:04 crc kubenswrapper[4943]: I0307 14:43:04.067818 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "31ce0220-0f53-4d63-aa7b-99357e9c6b04" (UID: "31ce0220-0f53-4d63-aa7b-99357e9c6b04"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:43:04 crc kubenswrapper[4943]: I0307 14:43:04.068062 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "31ce0220-0f53-4d63-aa7b-99357e9c6b04" (UID: "31ce0220-0f53-4d63-aa7b-99357e9c6b04"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:43:04 crc kubenswrapper[4943]: I0307 14:43:04.068431 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "31ce0220-0f53-4d63-aa7b-99357e9c6b04" (UID: "31ce0220-0f53-4d63-aa7b-99357e9c6b04"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:43:04 crc kubenswrapper[4943]: I0307 14:43:04.068987 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "31ce0220-0f53-4d63-aa7b-99357e9c6b04" (UID: "31ce0220-0f53-4d63-aa7b-99357e9c6b04"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:43:04 crc kubenswrapper[4943]: I0307 14:43:04.071291 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "31ce0220-0f53-4d63-aa7b-99357e9c6b04" (UID: "31ce0220-0f53-4d63-aa7b-99357e9c6b04"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:43:04 crc kubenswrapper[4943]: I0307 14:43:04.079260 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "31ce0220-0f53-4d63-aa7b-99357e9c6b04" (UID: "31ce0220-0f53-4d63-aa7b-99357e9c6b04"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:43:04 crc kubenswrapper[4943]: I0307 14:43:04.079883 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "31ce0220-0f53-4d63-aa7b-99357e9c6b04" (UID: "31ce0220-0f53-4d63-aa7b-99357e9c6b04"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:43:04 crc kubenswrapper[4943]: I0307 14:43:04.159860 4943 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Mar 07 14:43:04 crc kubenswrapper[4943]: I0307 14:43:04.159995 4943 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Mar 07 14:43:04 crc kubenswrapper[4943]: I0307 14:43:04.160517 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4pr5v\" (UniqueName: \"kubernetes.io/projected/31ce0220-0f53-4d63-aa7b-99357e9c6b04-kube-api-access-4pr5v\") on node \"crc\" DevicePath \"\"" Mar 07 14:43:04 crc kubenswrapper[4943]: I0307 14:43:04.161130 4943 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/31ce0220-0f53-4d63-aa7b-99357e9c6b04-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 07 14:43:04 crc kubenswrapper[4943]: I0307 14:43:04.161197 4943 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Mar 07 14:43:04 crc kubenswrapper[4943]: I0307 14:43:04.161226 4943 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Mar 07 14:43:04 crc kubenswrapper[4943]: I0307 14:43:04.161246 4943 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 07 14:43:04 crc kubenswrapper[4943]: I0307 14:43:04.161290 4943 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/31ce0220-0f53-4d63-aa7b-99357e9c6b04-audit-dir\") on node \"crc\" DevicePath \"\"" Mar 07 14:43:04 crc kubenswrapper[4943]: I0307 14:43:04.161312 4943 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Mar 07 14:43:04 crc kubenswrapper[4943]: I0307 14:43:04.161333 4943 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Mar 07 14:43:04 crc kubenswrapper[4943]: I0307 14:43:04.161353 4943 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Mar 07 14:43:04 crc kubenswrapper[4943]: I0307 14:43:04.161382 4943 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 07 14:43:04 crc kubenswrapper[4943]: I0307 14:43:04.161401 4943 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Mar 07 14:43:04 crc kubenswrapper[4943]: I0307 14:43:04.161421 4943 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/31ce0220-0f53-4d63-aa7b-99357e9c6b04-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Mar 07 14:43:04 crc kubenswrapper[4943]: I0307 14:43:04.695243 4943 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-cnbpx container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.21:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 07 14:43:04 crc kubenswrapper[4943]: I0307 14:43:04.695353 4943 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" podUID="31ce0220-0f53-4d63-aa7b-99357e9c6b04" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.21:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 07 14:43:04 crc kubenswrapper[4943]: I0307 14:43:04.769113 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"414a81ff993efbcd2f42f2cba71b82119d9a13d9e88121d441f55c22794fa179"} Mar 07 14:43:04 crc kubenswrapper[4943]: I0307 14:43:04.769181 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"5be6d0a7cc1f353c53bd061a5fed67cf92c25b34d3b2746a66d711db9f053736"} Mar 07 14:43:04 crc kubenswrapper[4943]: I0307 14:43:04.769198 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"055179cead3449430f4c32b643c8dd81efb93c8ed706a9310f32a1ea6ad98984"} Mar 07 14:43:04 crc kubenswrapper[4943]: I0307 14:43:04.772563 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" event={"ID":"31ce0220-0f53-4d63-aa7b-99357e9c6b04","Type":"ContainerDied","Data":"6158936056aea525e30acbce078252b346960a76806c3229d50779a81c5820bc"} Mar 07 14:43:04 crc kubenswrapper[4943]: I0307 14:43:04.772606 4943 scope.go:117] "RemoveContainer" containerID="64735be97cb4b56fae79aae00df6781ca6b2bc8a37a772cbedb7b9239d45514d" Mar 07 14:43:04 crc kubenswrapper[4943]: I0307 14:43:04.772750 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-cnbpx" Mar 07 14:43:04 crc kubenswrapper[4943]: I0307 14:43:04.909192 4943 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Readiness probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Mar 07 14:43:04 crc kubenswrapper[4943]: I0307 14:43:04.909277 4943 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Mar 07 14:43:05 crc kubenswrapper[4943]: I0307 14:43:05.754269 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7494769d95-sxpjq" Mar 07 14:43:05 crc kubenswrapper[4943]: I0307 14:43:05.754699 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7494769d95-sxpjq" Mar 07 14:43:05 crc kubenswrapper[4943]: I0307 14:43:05.783776 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"33859ff9c180b21b39fdcb8d2fc617743b5158b153f3a91c57a475a70fa67114"} Mar 07 14:43:05 crc kubenswrapper[4943]: I0307 14:43:05.783821 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"9a0689b317766c996180effd51a1ddc946fe15e9d6c937a752c2fc05e17dac89"} Mar 07 14:43:05 crc kubenswrapper[4943]: I0307 14:43:05.784079 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 07 14:43:05 crc kubenswrapper[4943]: I0307 14:43:05.784120 4943 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="107a7da3-71d2-4990-815e-9906bb3ea525" Mar 07 14:43:05 crc kubenswrapper[4943]: I0307 14:43:05.784136 4943 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="107a7da3-71d2-4990-815e-9906bb3ea525" Mar 07 14:43:05 crc kubenswrapper[4943]: I0307 14:43:05.787031 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/1.log" Mar 07 14:43:05 crc kubenswrapper[4943]: I0307 14:43:05.788100 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 07 14:43:05 crc kubenswrapper[4943]: I0307 14:43:05.788158 4943 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="156482e9536d7a822362fe687438f9532e7eadbc2f914c05345730fc1cfd7fda" exitCode=1 Mar 07 14:43:05 crc kubenswrapper[4943]: I0307 14:43:05.788191 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"156482e9536d7a822362fe687438f9532e7eadbc2f914c05345730fc1cfd7fda"} Mar 07 14:43:05 crc kubenswrapper[4943]: I0307 14:43:05.788750 4943 scope.go:117] "RemoveContainer" containerID="156482e9536d7a822362fe687438f9532e7eadbc2f914c05345730fc1cfd7fda" Mar 07 14:43:06 crc kubenswrapper[4943]: I0307 14:43:06.074771 4943 patch_prober.go:28] interesting pod/machine-config-daemon-cpdmm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 07 14:43:06 crc kubenswrapper[4943]: I0307 14:43:06.074853 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 07 14:43:06 crc kubenswrapper[4943]: I0307 14:43:06.795940 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/1.log" Mar 07 14:43:06 crc kubenswrapper[4943]: I0307 14:43:06.798254 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 07 14:43:06 crc kubenswrapper[4943]: I0307 14:43:06.798340 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"da3ee0d7c1ae27d070af70d1f7c6b46c8df86f93bee0ca5550ac544c2a3044f2"} Mar 07 14:43:07 crc kubenswrapper[4943]: I0307 14:43:07.780105 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 07 14:43:07 crc kubenswrapper[4943]: I0307 14:43:07.784399 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 07 14:43:07 crc kubenswrapper[4943]: I0307 14:43:07.790350 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 07 14:43:08 crc kubenswrapper[4943]: I0307 14:43:08.755342 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-86cb58c977-j8f96" Mar 07 14:43:08 crc kubenswrapper[4943]: I0307 14:43:08.756379 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-86cb58c977-j8f96" Mar 07 14:43:09 crc kubenswrapper[4943]: I0307 14:43:09.822502 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-86cb58c977-j8f96" event={"ID":"39fa4113-e14f-48e0-9106-52cfeeeba068","Type":"ContainerStarted","Data":"37b4bfcb827e76e31238203cdd63d860d1b13585d4a1a79ca11fdb34413b1b88"} Mar 07 14:43:09 crc kubenswrapper[4943]: I0307 14:43:09.822956 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-86cb58c977-j8f96" event={"ID":"39fa4113-e14f-48e0-9106-52cfeeeba068","Type":"ContainerStarted","Data":"b04980e6d84d196c36e13262e1ec4ed9544e2a5bbb1ce06f2b71bcd58de0e4e5"} Mar 07 14:43:09 crc kubenswrapper[4943]: I0307 14:43:09.823344 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-86cb58c977-j8f96" Mar 07 14:43:09 crc kubenswrapper[4943]: I0307 14:43:09.829742 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-86cb58c977-j8f96" Mar 07 14:43:10 crc kubenswrapper[4943]: I0307 14:43:10.800466 4943 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 07 14:43:10 crc kubenswrapper[4943]: I0307 14:43:10.830522 4943 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="107a7da3-71d2-4990-815e-9906bb3ea525" Mar 07 14:43:10 crc kubenswrapper[4943]: I0307 14:43:10.830570 4943 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="107a7da3-71d2-4990-815e-9906bb3ea525" Mar 07 14:43:10 crc kubenswrapper[4943]: I0307 14:43:10.842215 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 07 14:43:10 crc kubenswrapper[4943]: I0307 14:43:10.846079 4943 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="ba71413e-b349-4bd8-bc9d-d0c558ad1a2f" Mar 07 14:43:11 crc kubenswrapper[4943]: W0307 14:43:11.211815 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6deeb990_db08_494d_af2b_5be91b4271f6.slice/crio-18c2d8a0474529dc38b1b41505a0bcd19698de1e777b60189c4703655d15f747 WatchSource:0}: Error finding container 18c2d8a0474529dc38b1b41505a0bcd19698de1e777b60189c4703655d15f747: Status 404 returned error can't find the container with id 18c2d8a0474529dc38b1b41505a0bcd19698de1e777b60189c4703655d15f747 Mar 07 14:43:11 crc kubenswrapper[4943]: I0307 14:43:11.843364 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7494769d95-sxpjq" event={"ID":"6deeb990-db08-494d-af2b-5be91b4271f6","Type":"ContainerStarted","Data":"3721e81143ff4171222da9bb333715b2a05401a62bf00cf3396f49dc3732f6de"} Mar 07 14:43:11 crc kubenswrapper[4943]: I0307 14:43:11.843460 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7494769d95-sxpjq" event={"ID":"6deeb990-db08-494d-af2b-5be91b4271f6","Type":"ContainerStarted","Data":"18c2d8a0474529dc38b1b41505a0bcd19698de1e777b60189c4703655d15f747"} Mar 07 14:43:11 crc kubenswrapper[4943]: I0307 14:43:11.844097 4943 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="107a7da3-71d2-4990-815e-9906bb3ea525" Mar 07 14:43:11 crc kubenswrapper[4943]: I0307 14:43:11.844145 4943 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="107a7da3-71d2-4990-815e-9906bb3ea525" Mar 07 14:43:14 crc kubenswrapper[4943]: I0307 14:43:14.102237 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 07 14:43:14 crc kubenswrapper[4943]: I0307 14:43:14.108110 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 07 14:43:14 crc kubenswrapper[4943]: I0307 14:43:14.866911 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 07 14:43:18 crc kubenswrapper[4943]: I0307 14:43:18.795273 4943 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="ba71413e-b349-4bd8-bc9d-d0c558ad1a2f" Mar 07 14:43:21 crc kubenswrapper[4943]: I0307 14:43:21.024187 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Mar 07 14:43:21 crc kubenswrapper[4943]: I0307 14:43:21.192479 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7494769d95-sxpjq" Mar 07 14:43:21 crc kubenswrapper[4943]: I0307 14:43:21.771452 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 07 14:43:21 crc kubenswrapper[4943]: I0307 14:43:21.942141 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Mar 07 14:43:22 crc kubenswrapper[4943]: I0307 14:43:22.103043 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 07 14:43:22 crc kubenswrapper[4943]: I0307 14:43:22.192674 4943 patch_prober.go:28] interesting pod/route-controller-manager-7494769d95-sxpjq container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.67:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 07 14:43:22 crc kubenswrapper[4943]: I0307 14:43:22.192783 4943 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-7494769d95-sxpjq" podUID="6deeb990-db08-494d-af2b-5be91b4271f6" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.67:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 07 14:43:22 crc kubenswrapper[4943]: I0307 14:43:22.268002 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 07 14:43:22 crc kubenswrapper[4943]: I0307 14:43:22.441198 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Mar 07 14:43:22 crc kubenswrapper[4943]: I0307 14:43:22.720427 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 07 14:43:22 crc kubenswrapper[4943]: I0307 14:43:22.883734 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 07 14:43:23 crc kubenswrapper[4943]: I0307 14:43:23.193052 4943 patch_prober.go:28] interesting pod/route-controller-manager-7494769d95-sxpjq container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.67:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 07 14:43:23 crc kubenswrapper[4943]: I0307 14:43:23.193159 4943 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-7494769d95-sxpjq" podUID="6deeb990-db08-494d-af2b-5be91b4271f6" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.67:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 07 14:43:23 crc kubenswrapper[4943]: I0307 14:43:23.273581 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 07 14:43:23 crc kubenswrapper[4943]: I0307 14:43:23.315230 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 07 14:43:23 crc kubenswrapper[4943]: I0307 14:43:23.786048 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Mar 07 14:43:23 crc kubenswrapper[4943]: I0307 14:43:23.869768 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Mar 07 14:43:23 crc kubenswrapper[4943]: I0307 14:43:23.927640 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Mar 07 14:43:24 crc kubenswrapper[4943]: I0307 14:43:24.023885 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Mar 07 14:43:24 crc kubenswrapper[4943]: I0307 14:43:24.107768 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Mar 07 14:43:24 crc kubenswrapper[4943]: I0307 14:43:24.240922 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Mar 07 14:43:24 crc kubenswrapper[4943]: I0307 14:43:24.261304 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 07 14:43:24 crc kubenswrapper[4943]: I0307 14:43:24.305620 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 07 14:43:24 crc kubenswrapper[4943]: I0307 14:43:24.400051 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Mar 07 14:43:24 crc kubenswrapper[4943]: I0307 14:43:24.553016 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 07 14:43:24 crc kubenswrapper[4943]: I0307 14:43:24.586290 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Mar 07 14:43:24 crc kubenswrapper[4943]: I0307 14:43:24.592850 4943 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Mar 07 14:43:24 crc kubenswrapper[4943]: I0307 14:43:24.889587 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Mar 07 14:43:24 crc kubenswrapper[4943]: I0307 14:43:24.915093 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 07 14:43:25 crc kubenswrapper[4943]: I0307 14:43:25.003742 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 07 14:43:25 crc kubenswrapper[4943]: I0307 14:43:25.239968 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Mar 07 14:43:25 crc kubenswrapper[4943]: I0307 14:43:25.249066 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Mar 07 14:43:25 crc kubenswrapper[4943]: I0307 14:43:25.249224 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Mar 07 14:43:25 crc kubenswrapper[4943]: I0307 14:43:25.259294 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Mar 07 14:43:25 crc kubenswrapper[4943]: I0307 14:43:25.402057 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Mar 07 14:43:25 crc kubenswrapper[4943]: I0307 14:43:25.502404 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 07 14:43:25 crc kubenswrapper[4943]: I0307 14:43:25.517566 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 07 14:43:25 crc kubenswrapper[4943]: I0307 14:43:25.553090 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 07 14:43:25 crc kubenswrapper[4943]: I0307 14:43:25.568235 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Mar 07 14:43:25 crc kubenswrapper[4943]: I0307 14:43:25.576599 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Mar 07 14:43:25 crc kubenswrapper[4943]: I0307 14:43:25.584035 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Mar 07 14:43:25 crc kubenswrapper[4943]: I0307 14:43:25.595117 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Mar 07 14:43:25 crc kubenswrapper[4943]: I0307 14:43:25.750043 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 07 14:43:25 crc kubenswrapper[4943]: I0307 14:43:25.814545 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 07 14:43:25 crc kubenswrapper[4943]: I0307 14:43:25.816899 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Mar 07 14:43:25 crc kubenswrapper[4943]: I0307 14:43:25.886703 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Mar 07 14:43:26 crc kubenswrapper[4943]: I0307 14:43:26.054916 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Mar 07 14:43:26 crc kubenswrapper[4943]: I0307 14:43:26.055452 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 07 14:43:26 crc kubenswrapper[4943]: I0307 14:43:26.071250 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 07 14:43:26 crc kubenswrapper[4943]: I0307 14:43:26.246695 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 07 14:43:26 crc kubenswrapper[4943]: I0307 14:43:26.308754 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 07 14:43:26 crc kubenswrapper[4943]: I0307 14:43:26.326667 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Mar 07 14:43:26 crc kubenswrapper[4943]: I0307 14:43:26.375356 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Mar 07 14:43:26 crc kubenswrapper[4943]: I0307 14:43:26.378986 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Mar 07 14:43:26 crc kubenswrapper[4943]: I0307 14:43:26.438336 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Mar 07 14:43:26 crc kubenswrapper[4943]: I0307 14:43:26.466520 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Mar 07 14:43:26 crc kubenswrapper[4943]: I0307 14:43:26.470320 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Mar 07 14:43:26 crc kubenswrapper[4943]: I0307 14:43:26.530293 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Mar 07 14:43:26 crc kubenswrapper[4943]: I0307 14:43:26.594285 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Mar 07 14:43:26 crc kubenswrapper[4943]: I0307 14:43:26.620169 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 07 14:43:26 crc kubenswrapper[4943]: I0307 14:43:26.783464 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Mar 07 14:43:26 crc kubenswrapper[4943]: I0307 14:43:26.797068 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Mar 07 14:43:26 crc kubenswrapper[4943]: I0307 14:43:26.804535 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Mar 07 14:43:26 crc kubenswrapper[4943]: I0307 14:43:26.884576 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Mar 07 14:43:26 crc kubenswrapper[4943]: I0307 14:43:26.895123 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 07 14:43:26 crc kubenswrapper[4943]: I0307 14:43:26.906435 4943 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Mar 07 14:43:26 crc kubenswrapper[4943]: I0307 14:43:26.906891 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-86cb58c977-j8f96" podStartSLOduration=37.906876512 podStartE2EDuration="37.906876512s" podCreationTimestamp="2026-03-07 14:42:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:43:10.690270382 +0000 UTC m=+232.642406950" watchObservedRunningTime="2026-03-07 14:43:26.906876512 +0000 UTC m=+248.859013010" Mar 07 14:43:26 crc kubenswrapper[4943]: I0307 14:43:26.909622 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7494769d95-sxpjq" podStartSLOduration=37.909563046 podStartE2EDuration="37.909563046s" podCreationTimestamp="2026-03-07 14:42:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:43:11.875533269 +0000 UTC m=+233.827669807" watchObservedRunningTime="2026-03-07 14:43:26.909563046 +0000 UTC m=+248.861699544" Mar 07 14:43:26 crc kubenswrapper[4943]: I0307 14:43:26.910599 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=35.910591561 podStartE2EDuration="35.910591561s" podCreationTimestamp="2026-03-07 14:42:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:43:10.636267435 +0000 UTC m=+232.588403983" watchObservedRunningTime="2026-03-07 14:43:26.910591561 +0000 UTC m=+248.862728059" Mar 07 14:43:26 crc kubenswrapper[4943]: I0307 14:43:26.910911 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-cnbpx","openshift-kube-apiserver/kube-apiserver-crc"] Mar 07 14:43:26 crc kubenswrapper[4943]: I0307 14:43:26.910989 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 07 14:43:26 crc kubenswrapper[4943]: I0307 14:43:26.911009 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-86cb58c977-j8f96","openshift-route-controller-manager/route-controller-manager-7494769d95-sxpjq"] Mar 07 14:43:26 crc kubenswrapper[4943]: I0307 14:43:26.941612 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=16.941580335 podStartE2EDuration="16.941580335s" podCreationTimestamp="2026-03-07 14:43:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:43:26.933624164 +0000 UTC m=+248.885760702" watchObservedRunningTime="2026-03-07 14:43:26.941580335 +0000 UTC m=+248.893716863" Mar 07 14:43:26 crc kubenswrapper[4943]: I0307 14:43:26.983829 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Mar 07 14:43:27 crc kubenswrapper[4943]: I0307 14:43:27.076296 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Mar 07 14:43:27 crc kubenswrapper[4943]: I0307 14:43:27.128235 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Mar 07 14:43:27 crc kubenswrapper[4943]: I0307 14:43:27.203075 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Mar 07 14:43:27 crc kubenswrapper[4943]: I0307 14:43:27.342528 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Mar 07 14:43:27 crc kubenswrapper[4943]: I0307 14:43:27.373103 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Mar 07 14:43:27 crc kubenswrapper[4943]: I0307 14:43:27.406972 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Mar 07 14:43:27 crc kubenswrapper[4943]: I0307 14:43:27.440906 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Mar 07 14:43:27 crc kubenswrapper[4943]: I0307 14:43:27.521099 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Mar 07 14:43:27 crc kubenswrapper[4943]: I0307 14:43:27.593154 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Mar 07 14:43:27 crc kubenswrapper[4943]: I0307 14:43:27.646975 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Mar 07 14:43:27 crc kubenswrapper[4943]: I0307 14:43:27.833427 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Mar 07 14:43:27 crc kubenswrapper[4943]: I0307 14:43:27.835451 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 07 14:43:27 crc kubenswrapper[4943]: I0307 14:43:27.838870 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Mar 07 14:43:27 crc kubenswrapper[4943]: I0307 14:43:27.911541 4943 patch_prober.go:28] interesting pod/route-controller-manager-7494769d95-sxpjq container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.67:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 07 14:43:27 crc kubenswrapper[4943]: I0307 14:43:27.911722 4943 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-7494769d95-sxpjq" podUID="6deeb990-db08-494d-af2b-5be91b4271f6" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.67:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 07 14:43:27 crc kubenswrapper[4943]: I0307 14:43:27.914454 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 07 14:43:27 crc kubenswrapper[4943]: I0307 14:43:27.956307 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Mar 07 14:43:28 crc kubenswrapper[4943]: I0307 14:43:28.134049 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Mar 07 14:43:28 crc kubenswrapper[4943]: I0307 14:43:28.138737 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 07 14:43:28 crc kubenswrapper[4943]: I0307 14:43:28.430618 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Mar 07 14:43:28 crc kubenswrapper[4943]: I0307 14:43:28.458532 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Mar 07 14:43:28 crc kubenswrapper[4943]: I0307 14:43:28.485498 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Mar 07 14:43:28 crc kubenswrapper[4943]: I0307 14:43:28.489069 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Mar 07 14:43:28 crc kubenswrapper[4943]: I0307 14:43:28.528916 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Mar 07 14:43:28 crc kubenswrapper[4943]: I0307 14:43:28.554824 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Mar 07 14:43:28 crc kubenswrapper[4943]: I0307 14:43:28.555373 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Mar 07 14:43:28 crc kubenswrapper[4943]: I0307 14:43:28.568422 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Mar 07 14:43:28 crc kubenswrapper[4943]: I0307 14:43:28.719004 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 07 14:43:28 crc kubenswrapper[4943]: I0307 14:43:28.728130 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 07 14:43:28 crc kubenswrapper[4943]: I0307 14:43:28.743544 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Mar 07 14:43:28 crc kubenswrapper[4943]: I0307 14:43:28.765678 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31ce0220-0f53-4d63-aa7b-99357e9c6b04" path="/var/lib/kubelet/pods/31ce0220-0f53-4d63-aa7b-99357e9c6b04/volumes" Mar 07 14:43:28 crc kubenswrapper[4943]: I0307 14:43:28.774535 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Mar 07 14:43:28 crc kubenswrapper[4943]: I0307 14:43:28.826554 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Mar 07 14:43:28 crc kubenswrapper[4943]: I0307 14:43:28.847339 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 07 14:43:28 crc kubenswrapper[4943]: I0307 14:43:28.916194 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 07 14:43:29 crc kubenswrapper[4943]: I0307 14:43:29.082114 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Mar 07 14:43:29 crc kubenswrapper[4943]: I0307 14:43:29.122436 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Mar 07 14:43:29 crc kubenswrapper[4943]: I0307 14:43:29.197310 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Mar 07 14:43:29 crc kubenswrapper[4943]: I0307 14:43:29.249595 4943 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Mar 07 14:43:29 crc kubenswrapper[4943]: I0307 14:43:29.260422 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Mar 07 14:43:29 crc kubenswrapper[4943]: I0307 14:43:29.316469 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 07 14:43:29 crc kubenswrapper[4943]: I0307 14:43:29.433701 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Mar 07 14:43:29 crc kubenswrapper[4943]: I0307 14:43:29.434077 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Mar 07 14:43:29 crc kubenswrapper[4943]: I0307 14:43:29.522170 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Mar 07 14:43:29 crc kubenswrapper[4943]: I0307 14:43:29.587471 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Mar 07 14:43:29 crc kubenswrapper[4943]: I0307 14:43:29.793785 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Mar 07 14:43:29 crc kubenswrapper[4943]: I0307 14:43:29.856669 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Mar 07 14:43:29 crc kubenswrapper[4943]: I0307 14:43:29.907998 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Mar 07 14:43:29 crc kubenswrapper[4943]: I0307 14:43:29.953843 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Mar 07 14:43:29 crc kubenswrapper[4943]: I0307 14:43:29.955900 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Mar 07 14:43:29 crc kubenswrapper[4943]: I0307 14:43:29.965708 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Mar 07 14:43:29 crc kubenswrapper[4943]: I0307 14:43:29.981027 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Mar 07 14:43:30 crc kubenswrapper[4943]: I0307 14:43:30.047979 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Mar 07 14:43:30 crc kubenswrapper[4943]: I0307 14:43:30.085730 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Mar 07 14:43:30 crc kubenswrapper[4943]: I0307 14:43:30.104467 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Mar 07 14:43:30 crc kubenswrapper[4943]: I0307 14:43:30.128021 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Mar 07 14:43:30 crc kubenswrapper[4943]: I0307 14:43:30.174030 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Mar 07 14:43:30 crc kubenswrapper[4943]: I0307 14:43:30.178090 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 07 14:43:30 crc kubenswrapper[4943]: I0307 14:43:30.196717 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Mar 07 14:43:30 crc kubenswrapper[4943]: I0307 14:43:30.205306 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Mar 07 14:43:30 crc kubenswrapper[4943]: I0307 14:43:30.295994 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Mar 07 14:43:30 crc kubenswrapper[4943]: I0307 14:43:30.370722 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 07 14:43:30 crc kubenswrapper[4943]: I0307 14:43:30.379251 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Mar 07 14:43:30 crc kubenswrapper[4943]: I0307 14:43:30.450094 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Mar 07 14:43:30 crc kubenswrapper[4943]: I0307 14:43:30.455704 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 07 14:43:30 crc kubenswrapper[4943]: I0307 14:43:30.541961 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Mar 07 14:43:30 crc kubenswrapper[4943]: I0307 14:43:30.559641 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Mar 07 14:43:30 crc kubenswrapper[4943]: I0307 14:43:30.700325 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Mar 07 14:43:30 crc kubenswrapper[4943]: I0307 14:43:30.721714 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Mar 07 14:43:30 crc kubenswrapper[4943]: I0307 14:43:30.856684 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 07 14:43:30 crc kubenswrapper[4943]: I0307 14:43:30.946830 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 07 14:43:30 crc kubenswrapper[4943]: I0307 14:43:30.966029 4943 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Mar 07 14:43:30 crc kubenswrapper[4943]: I0307 14:43:30.996124 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Mar 07 14:43:31 crc kubenswrapper[4943]: I0307 14:43:31.141916 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Mar 07 14:43:31 crc kubenswrapper[4943]: I0307 14:43:31.145707 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Mar 07 14:43:31 crc kubenswrapper[4943]: I0307 14:43:31.177434 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Mar 07 14:43:31 crc kubenswrapper[4943]: I0307 14:43:31.197011 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7494769d95-sxpjq" Mar 07 14:43:31 crc kubenswrapper[4943]: I0307 14:43:31.225981 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Mar 07 14:43:31 crc kubenswrapper[4943]: I0307 14:43:31.238699 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Mar 07 14:43:31 crc kubenswrapper[4943]: I0307 14:43:31.319484 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Mar 07 14:43:31 crc kubenswrapper[4943]: I0307 14:43:31.364833 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Mar 07 14:43:31 crc kubenswrapper[4943]: I0307 14:43:31.478872 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Mar 07 14:43:31 crc kubenswrapper[4943]: I0307 14:43:31.528781 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Mar 07 14:43:31 crc kubenswrapper[4943]: I0307 14:43:31.562489 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Mar 07 14:43:31 crc kubenswrapper[4943]: I0307 14:43:31.563227 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Mar 07 14:43:31 crc kubenswrapper[4943]: I0307 14:43:31.584097 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Mar 07 14:43:31 crc kubenswrapper[4943]: I0307 14:43:31.679807 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Mar 07 14:43:31 crc kubenswrapper[4943]: I0307 14:43:31.723689 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Mar 07 14:43:31 crc kubenswrapper[4943]: I0307 14:43:31.736690 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Mar 07 14:43:31 crc kubenswrapper[4943]: I0307 14:43:31.805224 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Mar 07 14:43:31 crc kubenswrapper[4943]: I0307 14:43:31.930500 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Mar 07 14:43:31 crc kubenswrapper[4943]: I0307 14:43:31.998643 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-74fd85d944-hhfzl"] Mar 07 14:43:31 crc kubenswrapper[4943]: E0307 14:43:31.999126 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31ce0220-0f53-4d63-aa7b-99357e9c6b04" containerName="oauth-openshift" Mar 07 14:43:31 crc kubenswrapper[4943]: I0307 14:43:31.999156 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="31ce0220-0f53-4d63-aa7b-99357e9c6b04" containerName="oauth-openshift" Mar 07 14:43:31 crc kubenswrapper[4943]: E0307 14:43:31.999209 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="979a9c5a-4f69-4b26-8127-f760a697ae91" containerName="installer" Mar 07 14:43:31 crc kubenswrapper[4943]: I0307 14:43:31.999229 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="979a9c5a-4f69-4b26-8127-f760a697ae91" containerName="installer" Mar 07 14:43:31 crc kubenswrapper[4943]: I0307 14:43:31.999466 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="979a9c5a-4f69-4b26-8127-f760a697ae91" containerName="installer" Mar 07 14:43:31 crc kubenswrapper[4943]: I0307 14:43:31.999503 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="31ce0220-0f53-4d63-aa7b-99357e9c6b04" containerName="oauth-openshift" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.000490 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-74fd85d944-hhfzl" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.004638 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.005898 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.008477 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.008559 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.009515 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.009514 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.009738 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.010018 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.010407 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.010544 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.010676 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.011913 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.019478 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-74fd85d944-hhfzl"] Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.028348 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.028743 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.038031 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.170500 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/9f926425-044f-4308-908d-5214d71732cd-v4-0-config-system-session\") pod \"oauth-openshift-74fd85d944-hhfzl\" (UID: \"9f926425-044f-4308-908d-5214d71732cd\") " pod="openshift-authentication/oauth-openshift-74fd85d944-hhfzl" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.170619 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/9f926425-044f-4308-908d-5214d71732cd-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-74fd85d944-hhfzl\" (UID: \"9f926425-044f-4308-908d-5214d71732cd\") " pod="openshift-authentication/oauth-openshift-74fd85d944-hhfzl" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.170695 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/9f926425-044f-4308-908d-5214d71732cd-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-74fd85d944-hhfzl\" (UID: \"9f926425-044f-4308-908d-5214d71732cd\") " pod="openshift-authentication/oauth-openshift-74fd85d944-hhfzl" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.171042 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/9f926425-044f-4308-908d-5214d71732cd-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-74fd85d944-hhfzl\" (UID: \"9f926425-044f-4308-908d-5214d71732cd\") " pod="openshift-authentication/oauth-openshift-74fd85d944-hhfzl" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.171129 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9f926425-044f-4308-908d-5214d71732cd-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-74fd85d944-hhfzl\" (UID: \"9f926425-044f-4308-908d-5214d71732cd\") " pod="openshift-authentication/oauth-openshift-74fd85d944-hhfzl" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.171163 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/9f926425-044f-4308-908d-5214d71732cd-v4-0-config-user-template-login\") pod \"oauth-openshift-74fd85d944-hhfzl\" (UID: \"9f926425-044f-4308-908d-5214d71732cd\") " pod="openshift-authentication/oauth-openshift-74fd85d944-hhfzl" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.171281 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/9f926425-044f-4308-908d-5214d71732cd-v4-0-config-system-serving-cert\") pod \"oauth-openshift-74fd85d944-hhfzl\" (UID: \"9f926425-044f-4308-908d-5214d71732cd\") " pod="openshift-authentication/oauth-openshift-74fd85d944-hhfzl" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.171329 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8h6ql\" (UniqueName: \"kubernetes.io/projected/9f926425-044f-4308-908d-5214d71732cd-kube-api-access-8h6ql\") pod \"oauth-openshift-74fd85d944-hhfzl\" (UID: \"9f926425-044f-4308-908d-5214d71732cd\") " pod="openshift-authentication/oauth-openshift-74fd85d944-hhfzl" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.171355 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/9f926425-044f-4308-908d-5214d71732cd-v4-0-config-system-service-ca\") pod \"oauth-openshift-74fd85d944-hhfzl\" (UID: \"9f926425-044f-4308-908d-5214d71732cd\") " pod="openshift-authentication/oauth-openshift-74fd85d944-hhfzl" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.171377 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9f926425-044f-4308-908d-5214d71732cd-audit-policies\") pod \"oauth-openshift-74fd85d944-hhfzl\" (UID: \"9f926425-044f-4308-908d-5214d71732cd\") " pod="openshift-authentication/oauth-openshift-74fd85d944-hhfzl" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.171397 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9f926425-044f-4308-908d-5214d71732cd-audit-dir\") pod \"oauth-openshift-74fd85d944-hhfzl\" (UID: \"9f926425-044f-4308-908d-5214d71732cd\") " pod="openshift-authentication/oauth-openshift-74fd85d944-hhfzl" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.171413 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/9f926425-044f-4308-908d-5214d71732cd-v4-0-config-system-cliconfig\") pod \"oauth-openshift-74fd85d944-hhfzl\" (UID: \"9f926425-044f-4308-908d-5214d71732cd\") " pod="openshift-authentication/oauth-openshift-74fd85d944-hhfzl" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.171445 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/9f926425-044f-4308-908d-5214d71732cd-v4-0-config-system-router-certs\") pod \"oauth-openshift-74fd85d944-hhfzl\" (UID: \"9f926425-044f-4308-908d-5214d71732cd\") " pod="openshift-authentication/oauth-openshift-74fd85d944-hhfzl" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.171470 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/9f926425-044f-4308-908d-5214d71732cd-v4-0-config-user-template-error\") pod \"oauth-openshift-74fd85d944-hhfzl\" (UID: \"9f926425-044f-4308-908d-5214d71732cd\") " pod="openshift-authentication/oauth-openshift-74fd85d944-hhfzl" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.249631 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.264089 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.272833 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/9f926425-044f-4308-908d-5214d71732cd-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-74fd85d944-hhfzl\" (UID: \"9f926425-044f-4308-908d-5214d71732cd\") " pod="openshift-authentication/oauth-openshift-74fd85d944-hhfzl" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.272963 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/9f926425-044f-4308-908d-5214d71732cd-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-74fd85d944-hhfzl\" (UID: \"9f926425-044f-4308-908d-5214d71732cd\") " pod="openshift-authentication/oauth-openshift-74fd85d944-hhfzl" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.273034 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9f926425-044f-4308-908d-5214d71732cd-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-74fd85d944-hhfzl\" (UID: \"9f926425-044f-4308-908d-5214d71732cd\") " pod="openshift-authentication/oauth-openshift-74fd85d944-hhfzl" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.273084 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/9f926425-044f-4308-908d-5214d71732cd-v4-0-config-user-template-login\") pod \"oauth-openshift-74fd85d944-hhfzl\" (UID: \"9f926425-044f-4308-908d-5214d71732cd\") " pod="openshift-authentication/oauth-openshift-74fd85d944-hhfzl" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.273141 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/9f926425-044f-4308-908d-5214d71732cd-v4-0-config-system-serving-cert\") pod \"oauth-openshift-74fd85d944-hhfzl\" (UID: \"9f926425-044f-4308-908d-5214d71732cd\") " pod="openshift-authentication/oauth-openshift-74fd85d944-hhfzl" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.273199 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8h6ql\" (UniqueName: \"kubernetes.io/projected/9f926425-044f-4308-908d-5214d71732cd-kube-api-access-8h6ql\") pod \"oauth-openshift-74fd85d944-hhfzl\" (UID: \"9f926425-044f-4308-908d-5214d71732cd\") " pod="openshift-authentication/oauth-openshift-74fd85d944-hhfzl" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.273296 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/9f926425-044f-4308-908d-5214d71732cd-v4-0-config-system-service-ca\") pod \"oauth-openshift-74fd85d944-hhfzl\" (UID: \"9f926425-044f-4308-908d-5214d71732cd\") " pod="openshift-authentication/oauth-openshift-74fd85d944-hhfzl" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.273348 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9f926425-044f-4308-908d-5214d71732cd-audit-policies\") pod \"oauth-openshift-74fd85d944-hhfzl\" (UID: \"9f926425-044f-4308-908d-5214d71732cd\") " pod="openshift-authentication/oauth-openshift-74fd85d944-hhfzl" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.273399 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/9f926425-044f-4308-908d-5214d71732cd-v4-0-config-system-cliconfig\") pod \"oauth-openshift-74fd85d944-hhfzl\" (UID: \"9f926425-044f-4308-908d-5214d71732cd\") " pod="openshift-authentication/oauth-openshift-74fd85d944-hhfzl" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.273450 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9f926425-044f-4308-908d-5214d71732cd-audit-dir\") pod \"oauth-openshift-74fd85d944-hhfzl\" (UID: \"9f926425-044f-4308-908d-5214d71732cd\") " pod="openshift-authentication/oauth-openshift-74fd85d944-hhfzl" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.273534 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/9f926425-044f-4308-908d-5214d71732cd-v4-0-config-system-router-certs\") pod \"oauth-openshift-74fd85d944-hhfzl\" (UID: \"9f926425-044f-4308-908d-5214d71732cd\") " pod="openshift-authentication/oauth-openshift-74fd85d944-hhfzl" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.273605 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/9f926425-044f-4308-908d-5214d71732cd-v4-0-config-user-template-error\") pod \"oauth-openshift-74fd85d944-hhfzl\" (UID: \"9f926425-044f-4308-908d-5214d71732cd\") " pod="openshift-authentication/oauth-openshift-74fd85d944-hhfzl" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.273679 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/9f926425-044f-4308-908d-5214d71732cd-v4-0-config-system-session\") pod \"oauth-openshift-74fd85d944-hhfzl\" (UID: \"9f926425-044f-4308-908d-5214d71732cd\") " pod="openshift-authentication/oauth-openshift-74fd85d944-hhfzl" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.273693 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9f926425-044f-4308-908d-5214d71732cd-audit-dir\") pod \"oauth-openshift-74fd85d944-hhfzl\" (UID: \"9f926425-044f-4308-908d-5214d71732cd\") " pod="openshift-authentication/oauth-openshift-74fd85d944-hhfzl" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.273752 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/9f926425-044f-4308-908d-5214d71732cd-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-74fd85d944-hhfzl\" (UID: \"9f926425-044f-4308-908d-5214d71732cd\") " pod="openshift-authentication/oauth-openshift-74fd85d944-hhfzl" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.275188 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9f926425-044f-4308-908d-5214d71732cd-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-74fd85d944-hhfzl\" (UID: \"9f926425-044f-4308-908d-5214d71732cd\") " pod="openshift-authentication/oauth-openshift-74fd85d944-hhfzl" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.275216 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/9f926425-044f-4308-908d-5214d71732cd-v4-0-config-system-cliconfig\") pod \"oauth-openshift-74fd85d944-hhfzl\" (UID: \"9f926425-044f-4308-908d-5214d71732cd\") " pod="openshift-authentication/oauth-openshift-74fd85d944-hhfzl" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.275472 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9f926425-044f-4308-908d-5214d71732cd-audit-policies\") pod \"oauth-openshift-74fd85d944-hhfzl\" (UID: \"9f926425-044f-4308-908d-5214d71732cd\") " pod="openshift-authentication/oauth-openshift-74fd85d944-hhfzl" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.277108 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/9f926425-044f-4308-908d-5214d71732cd-v4-0-config-system-service-ca\") pod \"oauth-openshift-74fd85d944-hhfzl\" (UID: \"9f926425-044f-4308-908d-5214d71732cd\") " pod="openshift-authentication/oauth-openshift-74fd85d944-hhfzl" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.284902 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/9f926425-044f-4308-908d-5214d71732cd-v4-0-config-user-template-error\") pod \"oauth-openshift-74fd85d944-hhfzl\" (UID: \"9f926425-044f-4308-908d-5214d71732cd\") " pod="openshift-authentication/oauth-openshift-74fd85d944-hhfzl" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.284906 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/9f926425-044f-4308-908d-5214d71732cd-v4-0-config-system-session\") pod \"oauth-openshift-74fd85d944-hhfzl\" (UID: \"9f926425-044f-4308-908d-5214d71732cd\") " pod="openshift-authentication/oauth-openshift-74fd85d944-hhfzl" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.285114 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/9f926425-044f-4308-908d-5214d71732cd-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-74fd85d944-hhfzl\" (UID: \"9f926425-044f-4308-908d-5214d71732cd\") " pod="openshift-authentication/oauth-openshift-74fd85d944-hhfzl" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.285704 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/9f926425-044f-4308-908d-5214d71732cd-v4-0-config-system-serving-cert\") pod \"oauth-openshift-74fd85d944-hhfzl\" (UID: \"9f926425-044f-4308-908d-5214d71732cd\") " pod="openshift-authentication/oauth-openshift-74fd85d944-hhfzl" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.286263 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/9f926425-044f-4308-908d-5214d71732cd-v4-0-config-user-template-login\") pod \"oauth-openshift-74fd85d944-hhfzl\" (UID: \"9f926425-044f-4308-908d-5214d71732cd\") " pod="openshift-authentication/oauth-openshift-74fd85d944-hhfzl" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.287587 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/9f926425-044f-4308-908d-5214d71732cd-v4-0-config-system-router-certs\") pod \"oauth-openshift-74fd85d944-hhfzl\" (UID: \"9f926425-044f-4308-908d-5214d71732cd\") " pod="openshift-authentication/oauth-openshift-74fd85d944-hhfzl" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.287670 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/9f926425-044f-4308-908d-5214d71732cd-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-74fd85d944-hhfzl\" (UID: \"9f926425-044f-4308-908d-5214d71732cd\") " pod="openshift-authentication/oauth-openshift-74fd85d944-hhfzl" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.287915 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/9f926425-044f-4308-908d-5214d71732cd-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-74fd85d944-hhfzl\" (UID: \"9f926425-044f-4308-908d-5214d71732cd\") " pod="openshift-authentication/oauth-openshift-74fd85d944-hhfzl" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.299716 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8h6ql\" (UniqueName: \"kubernetes.io/projected/9f926425-044f-4308-908d-5214d71732cd-kube-api-access-8h6ql\") pod \"oauth-openshift-74fd85d944-hhfzl\" (UID: \"9f926425-044f-4308-908d-5214d71732cd\") " pod="openshift-authentication/oauth-openshift-74fd85d944-hhfzl" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.328517 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-74fd85d944-hhfzl" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.379130 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.390624 4943 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.461516 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.520668 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.538838 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.590683 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.604650 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.612224 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-74fd85d944-hhfzl"] Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.762447 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.783810 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.866104 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.878848 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.905306 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.921376 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.998618 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-74fd85d944-hhfzl" event={"ID":"9f926425-044f-4308-908d-5214d71732cd","Type":"ContainerStarted","Data":"aeef44600f90b26188849f13eeb1c6bec2a5d0370f433466f75e2d2df40778e9"} Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.998677 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-74fd85d944-hhfzl" event={"ID":"9f926425-044f-4308-908d-5214d71732cd","Type":"ContainerStarted","Data":"fe953dcc60e7edbfaf8909177c3e3e9c06e4ae763adc5b7e2164b731ce2e3c9e"} Mar 07 14:43:32 crc kubenswrapper[4943]: I0307 14:43:32.998983 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-74fd85d944-hhfzl" Mar 07 14:43:33 crc kubenswrapper[4943]: I0307 14:43:33.022146 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-74fd85d944-hhfzl" podStartSLOduration=55.022125183 podStartE2EDuration="55.022125183s" podCreationTimestamp="2026-03-07 14:42:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:43:33.019562251 +0000 UTC m=+254.971698749" watchObservedRunningTime="2026-03-07 14:43:33.022125183 +0000 UTC m=+254.974261681" Mar 07 14:43:33 crc kubenswrapper[4943]: I0307 14:43:33.076790 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Mar 07 14:43:33 crc kubenswrapper[4943]: I0307 14:43:33.110908 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Mar 07 14:43:33 crc kubenswrapper[4943]: I0307 14:43:33.139679 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Mar 07 14:43:33 crc kubenswrapper[4943]: I0307 14:43:33.309638 4943 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 07 14:43:33 crc kubenswrapper[4943]: I0307 14:43:33.309946 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://f1df75a5adbef71c71f3333fa02b2226668f211c29e27ab17fe30a7a46310f93" gracePeriod=5 Mar 07 14:43:33 crc kubenswrapper[4943]: I0307 14:43:33.314037 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 07 14:43:33 crc kubenswrapper[4943]: I0307 14:43:33.326821 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Mar 07 14:43:33 crc kubenswrapper[4943]: I0307 14:43:33.361043 4943 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Mar 07 14:43:33 crc kubenswrapper[4943]: I0307 14:43:33.500600 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-74fd85d944-hhfzl" Mar 07 14:43:33 crc kubenswrapper[4943]: I0307 14:43:33.522374 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 07 14:43:33 crc kubenswrapper[4943]: I0307 14:43:33.528078 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Mar 07 14:43:33 crc kubenswrapper[4943]: I0307 14:43:33.568133 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Mar 07 14:43:33 crc kubenswrapper[4943]: I0307 14:43:33.649099 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Mar 07 14:43:33 crc kubenswrapper[4943]: I0307 14:43:33.698435 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 07 14:43:33 crc kubenswrapper[4943]: I0307 14:43:33.792305 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Mar 07 14:43:33 crc kubenswrapper[4943]: I0307 14:43:33.816716 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 07 14:43:33 crc kubenswrapper[4943]: I0307 14:43:33.824153 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Mar 07 14:43:33 crc kubenswrapper[4943]: I0307 14:43:33.853599 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Mar 07 14:43:33 crc kubenswrapper[4943]: I0307 14:43:33.904970 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Mar 07 14:43:33 crc kubenswrapper[4943]: I0307 14:43:33.905093 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 07 14:43:33 crc kubenswrapper[4943]: I0307 14:43:33.923189 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Mar 07 14:43:33 crc kubenswrapper[4943]: I0307 14:43:33.946876 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Mar 07 14:43:33 crc kubenswrapper[4943]: I0307 14:43:33.969031 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Mar 07 14:43:33 crc kubenswrapper[4943]: I0307 14:43:33.970123 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Mar 07 14:43:33 crc kubenswrapper[4943]: I0307 14:43:33.995223 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Mar 07 14:43:33 crc kubenswrapper[4943]: I0307 14:43:33.996472 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 07 14:43:33 crc kubenswrapper[4943]: I0307 14:43:33.998178 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Mar 07 14:43:34 crc kubenswrapper[4943]: I0307 14:43:34.010364 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Mar 07 14:43:34 crc kubenswrapper[4943]: I0307 14:43:34.242640 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Mar 07 14:43:34 crc kubenswrapper[4943]: I0307 14:43:34.497538 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 07 14:43:34 crc kubenswrapper[4943]: I0307 14:43:34.579905 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Mar 07 14:43:34 crc kubenswrapper[4943]: I0307 14:43:34.652580 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Mar 07 14:43:34 crc kubenswrapper[4943]: I0307 14:43:34.654797 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 07 14:43:34 crc kubenswrapper[4943]: I0307 14:43:34.710023 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Mar 07 14:43:34 crc kubenswrapper[4943]: I0307 14:43:34.733653 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Mar 07 14:43:34 crc kubenswrapper[4943]: I0307 14:43:34.777924 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Mar 07 14:43:34 crc kubenswrapper[4943]: I0307 14:43:34.788419 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Mar 07 14:43:34 crc kubenswrapper[4943]: I0307 14:43:34.861785 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Mar 07 14:43:35 crc kubenswrapper[4943]: I0307 14:43:35.030426 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Mar 07 14:43:35 crc kubenswrapper[4943]: I0307 14:43:35.045797 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Mar 07 14:43:35 crc kubenswrapper[4943]: I0307 14:43:35.085473 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 07 14:43:35 crc kubenswrapper[4943]: I0307 14:43:35.126221 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Mar 07 14:43:35 crc kubenswrapper[4943]: I0307 14:43:35.184721 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Mar 07 14:43:35 crc kubenswrapper[4943]: I0307 14:43:35.236834 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 07 14:43:35 crc kubenswrapper[4943]: I0307 14:43:35.301133 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Mar 07 14:43:35 crc kubenswrapper[4943]: I0307 14:43:35.360212 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Mar 07 14:43:35 crc kubenswrapper[4943]: I0307 14:43:35.376047 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Mar 07 14:43:35 crc kubenswrapper[4943]: I0307 14:43:35.487017 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 07 14:43:35 crc kubenswrapper[4943]: I0307 14:43:35.510301 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Mar 07 14:43:35 crc kubenswrapper[4943]: I0307 14:43:35.529991 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Mar 07 14:43:35 crc kubenswrapper[4943]: I0307 14:43:35.536759 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Mar 07 14:43:35 crc kubenswrapper[4943]: I0307 14:43:35.541108 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Mar 07 14:43:35 crc kubenswrapper[4943]: I0307 14:43:35.646745 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Mar 07 14:43:35 crc kubenswrapper[4943]: I0307 14:43:35.752874 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Mar 07 14:43:35 crc kubenswrapper[4943]: I0307 14:43:35.791296 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Mar 07 14:43:35 crc kubenswrapper[4943]: I0307 14:43:35.886466 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 07 14:43:35 crc kubenswrapper[4943]: I0307 14:43:35.956486 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Mar 07 14:43:36 crc kubenswrapper[4943]: I0307 14:43:36.074681 4943 patch_prober.go:28] interesting pod/machine-config-daemon-cpdmm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 07 14:43:36 crc kubenswrapper[4943]: I0307 14:43:36.074762 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 07 14:43:36 crc kubenswrapper[4943]: I0307 14:43:36.163542 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Mar 07 14:43:36 crc kubenswrapper[4943]: I0307 14:43:36.174141 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Mar 07 14:43:36 crc kubenswrapper[4943]: I0307 14:43:36.277709 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Mar 07 14:43:36 crc kubenswrapper[4943]: I0307 14:43:36.326892 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Mar 07 14:43:36 crc kubenswrapper[4943]: I0307 14:43:36.360804 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Mar 07 14:43:36 crc kubenswrapper[4943]: I0307 14:43:36.435568 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Mar 07 14:43:36 crc kubenswrapper[4943]: I0307 14:43:36.615352 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Mar 07 14:43:36 crc kubenswrapper[4943]: I0307 14:43:36.814082 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 07 14:43:37 crc kubenswrapper[4943]: I0307 14:43:37.004957 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Mar 07 14:43:37 crc kubenswrapper[4943]: I0307 14:43:37.068682 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Mar 07 14:43:37 crc kubenswrapper[4943]: I0307 14:43:37.631971 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Mar 07 14:43:37 crc kubenswrapper[4943]: I0307 14:43:37.684102 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Mar 07 14:43:37 crc kubenswrapper[4943]: I0307 14:43:37.688573 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Mar 07 14:43:37 crc kubenswrapper[4943]: I0307 14:43:37.940840 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Mar 07 14:43:37 crc kubenswrapper[4943]: I0307 14:43:37.966272 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Mar 07 14:43:37 crc kubenswrapper[4943]: I0307 14:43:37.998132 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 07 14:43:38 crc kubenswrapper[4943]: I0307 14:43:38.008848 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Mar 07 14:43:38 crc kubenswrapper[4943]: I0307 14:43:38.043061 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Mar 07 14:43:38 crc kubenswrapper[4943]: I0307 14:43:38.170171 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Mar 07 14:43:38 crc kubenswrapper[4943]: I0307 14:43:38.893908 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 07 14:43:38 crc kubenswrapper[4943]: I0307 14:43:38.894060 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 07 14:43:38 crc kubenswrapper[4943]: I0307 14:43:38.987671 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 07 14:43:38 crc kubenswrapper[4943]: I0307 14:43:38.988066 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 07 14:43:38 crc kubenswrapper[4943]: I0307 14:43:38.988164 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 07 14:43:38 crc kubenswrapper[4943]: I0307 14:43:38.988235 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 07 14:43:38 crc kubenswrapper[4943]: I0307 14:43:38.988256 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 07 14:43:38 crc kubenswrapper[4943]: I0307 14:43:38.988393 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 07 14:43:38 crc kubenswrapper[4943]: I0307 14:43:38.988397 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 07 14:43:38 crc kubenswrapper[4943]: I0307 14:43:38.988440 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 07 14:43:38 crc kubenswrapper[4943]: I0307 14:43:38.988428 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 07 14:43:38 crc kubenswrapper[4943]: I0307 14:43:38.989230 4943 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Mar 07 14:43:38 crc kubenswrapper[4943]: I0307 14:43:38.989260 4943 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Mar 07 14:43:38 crc kubenswrapper[4943]: I0307 14:43:38.989285 4943 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 07 14:43:38 crc kubenswrapper[4943]: I0307 14:43:38.989307 4943 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Mar 07 14:43:39 crc kubenswrapper[4943]: I0307 14:43:39.001428 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 07 14:43:39 crc kubenswrapper[4943]: I0307 14:43:39.043189 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 07 14:43:39 crc kubenswrapper[4943]: I0307 14:43:39.043274 4943 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="f1df75a5adbef71c71f3333fa02b2226668f211c29e27ab17fe30a7a46310f93" exitCode=137 Mar 07 14:43:39 crc kubenswrapper[4943]: I0307 14:43:39.043355 4943 scope.go:117] "RemoveContainer" containerID="f1df75a5adbef71c71f3333fa02b2226668f211c29e27ab17fe30a7a46310f93" Mar 07 14:43:39 crc kubenswrapper[4943]: I0307 14:43:39.043686 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 07 14:43:39 crc kubenswrapper[4943]: I0307 14:43:39.070069 4943 scope.go:117] "RemoveContainer" containerID="f1df75a5adbef71c71f3333fa02b2226668f211c29e27ab17fe30a7a46310f93" Mar 07 14:43:39 crc kubenswrapper[4943]: E0307 14:43:39.070783 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1df75a5adbef71c71f3333fa02b2226668f211c29e27ab17fe30a7a46310f93\": container with ID starting with f1df75a5adbef71c71f3333fa02b2226668f211c29e27ab17fe30a7a46310f93 not found: ID does not exist" containerID="f1df75a5adbef71c71f3333fa02b2226668f211c29e27ab17fe30a7a46310f93" Mar 07 14:43:39 crc kubenswrapper[4943]: I0307 14:43:39.070862 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1df75a5adbef71c71f3333fa02b2226668f211c29e27ab17fe30a7a46310f93"} err="failed to get container status \"f1df75a5adbef71c71f3333fa02b2226668f211c29e27ab17fe30a7a46310f93\": rpc error: code = NotFound desc = could not find container \"f1df75a5adbef71c71f3333fa02b2226668f211c29e27ab17fe30a7a46310f93\": container with ID starting with f1df75a5adbef71c71f3333fa02b2226668f211c29e27ab17fe30a7a46310f93 not found: ID does not exist" Mar 07 14:43:39 crc kubenswrapper[4943]: I0307 14:43:39.090704 4943 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 07 14:43:39 crc kubenswrapper[4943]: I0307 14:43:39.157119 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Mar 07 14:43:39 crc kubenswrapper[4943]: I0307 14:43:39.279507 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Mar 07 14:43:40 crc kubenswrapper[4943]: I0307 14:43:40.767334 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Mar 07 14:43:40 crc kubenswrapper[4943]: I0307 14:43:40.768274 4943 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Mar 07 14:43:40 crc kubenswrapper[4943]: I0307 14:43:40.781746 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 07 14:43:40 crc kubenswrapper[4943]: I0307 14:43:40.781808 4943 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="1bbf757f-98e5-483c-b878-b9873112f64e" Mar 07 14:43:40 crc kubenswrapper[4943]: I0307 14:43:40.786803 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 07 14:43:40 crc kubenswrapper[4943]: I0307 14:43:40.786889 4943 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="1bbf757f-98e5-483c-b878-b9873112f64e" Mar 07 14:43:56 crc kubenswrapper[4943]: I0307 14:43:56.161517 4943 generic.go:334] "Generic (PLEG): container finished" podID="712a2a31-4205-4346-9a32-858a77615eb6" containerID="51fa91e109cdf7e64fa128d5ebe2fa2b77ae12ee40c73925027012b869163b0c" exitCode=0 Mar 07 14:43:56 crc kubenswrapper[4943]: I0307 14:43:56.161616 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-sqsdb" event={"ID":"712a2a31-4205-4346-9a32-858a77615eb6","Type":"ContainerDied","Data":"51fa91e109cdf7e64fa128d5ebe2fa2b77ae12ee40c73925027012b869163b0c"} Mar 07 14:43:56 crc kubenswrapper[4943]: I0307 14:43:56.163312 4943 scope.go:117] "RemoveContainer" containerID="51fa91e109cdf7e64fa128d5ebe2fa2b77ae12ee40c73925027012b869163b0c" Mar 07 14:43:57 crc kubenswrapper[4943]: I0307 14:43:57.171372 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-sqsdb" event={"ID":"712a2a31-4205-4346-9a32-858a77615eb6","Type":"ContainerStarted","Data":"08e935ddfdfcb02fe9bdeb0fb05b2807df84fbf1297330bd13c1372e83260314"} Mar 07 14:43:57 crc kubenswrapper[4943]: I0307 14:43:57.172120 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-sqsdb" Mar 07 14:43:57 crc kubenswrapper[4943]: I0307 14:43:57.176284 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-sqsdb" Mar 07 14:44:00 crc kubenswrapper[4943]: I0307 14:44:00.202521 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29548244-2zdk7"] Mar 07 14:44:00 crc kubenswrapper[4943]: E0307 14:44:00.203223 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 07 14:44:00 crc kubenswrapper[4943]: I0307 14:44:00.203242 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 07 14:44:00 crc kubenswrapper[4943]: I0307 14:44:00.203377 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 07 14:44:00 crc kubenswrapper[4943]: I0307 14:44:00.203913 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548244-2zdk7" Mar 07 14:44:00 crc kubenswrapper[4943]: I0307 14:44:00.204809 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kd9n\" (UniqueName: \"kubernetes.io/projected/0c43c4c0-8965-43f9-9724-6970c0f15729-kube-api-access-2kd9n\") pod \"auto-csr-approver-29548244-2zdk7\" (UID: \"0c43c4c0-8965-43f9-9724-6970c0f15729\") " pod="openshift-infra/auto-csr-approver-29548244-2zdk7" Mar 07 14:44:00 crc kubenswrapper[4943]: I0307 14:44:00.207221 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 07 14:44:00 crc kubenswrapper[4943]: I0307 14:44:00.207334 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-qqlk5" Mar 07 14:44:00 crc kubenswrapper[4943]: I0307 14:44:00.207532 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 07 14:44:00 crc kubenswrapper[4943]: I0307 14:44:00.215591 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29548244-2zdk7"] Mar 07 14:44:00 crc kubenswrapper[4943]: I0307 14:44:00.306352 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kd9n\" (UniqueName: \"kubernetes.io/projected/0c43c4c0-8965-43f9-9724-6970c0f15729-kube-api-access-2kd9n\") pod \"auto-csr-approver-29548244-2zdk7\" (UID: \"0c43c4c0-8965-43f9-9724-6970c0f15729\") " pod="openshift-infra/auto-csr-approver-29548244-2zdk7" Mar 07 14:44:00 crc kubenswrapper[4943]: I0307 14:44:00.329419 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kd9n\" (UniqueName: \"kubernetes.io/projected/0c43c4c0-8965-43f9-9724-6970c0f15729-kube-api-access-2kd9n\") pod \"auto-csr-approver-29548244-2zdk7\" (UID: \"0c43c4c0-8965-43f9-9724-6970c0f15729\") " pod="openshift-infra/auto-csr-approver-29548244-2zdk7" Mar 07 14:44:00 crc kubenswrapper[4943]: I0307 14:44:00.532241 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548244-2zdk7" Mar 07 14:44:01 crc kubenswrapper[4943]: I0307 14:44:01.006671 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29548244-2zdk7"] Mar 07 14:44:01 crc kubenswrapper[4943]: I0307 14:44:01.200531 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548244-2zdk7" event={"ID":"0c43c4c0-8965-43f9-9724-6970c0f15729","Type":"ContainerStarted","Data":"a0625771c8ce07afd62861fb3b56928832a7bdb9cb169e11b25c9d51657a86a8"} Mar 07 14:44:03 crc kubenswrapper[4943]: I0307 14:44:03.219082 4943 generic.go:334] "Generic (PLEG): container finished" podID="0c43c4c0-8965-43f9-9724-6970c0f15729" containerID="a0539578f6fccd6a106c51804570f0018ba0682b32ca76a6ae19a1fcf8ccf643" exitCode=0 Mar 07 14:44:03 crc kubenswrapper[4943]: I0307 14:44:03.219214 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548244-2zdk7" event={"ID":"0c43c4c0-8965-43f9-9724-6970c0f15729","Type":"ContainerDied","Data":"a0539578f6fccd6a106c51804570f0018ba0682b32ca76a6ae19a1fcf8ccf643"} Mar 07 14:44:04 crc kubenswrapper[4943]: I0307 14:44:04.717133 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548244-2zdk7" Mar 07 14:44:04 crc kubenswrapper[4943]: I0307 14:44:04.870897 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2kd9n\" (UniqueName: \"kubernetes.io/projected/0c43c4c0-8965-43f9-9724-6970c0f15729-kube-api-access-2kd9n\") pod \"0c43c4c0-8965-43f9-9724-6970c0f15729\" (UID: \"0c43c4c0-8965-43f9-9724-6970c0f15729\") " Mar 07 14:44:04 crc kubenswrapper[4943]: I0307 14:44:04.879201 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c43c4c0-8965-43f9-9724-6970c0f15729-kube-api-access-2kd9n" (OuterVolumeSpecName: "kube-api-access-2kd9n") pod "0c43c4c0-8965-43f9-9724-6970c0f15729" (UID: "0c43c4c0-8965-43f9-9724-6970c0f15729"). InnerVolumeSpecName "kube-api-access-2kd9n". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:44:04 crc kubenswrapper[4943]: I0307 14:44:04.972737 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2kd9n\" (UniqueName: \"kubernetes.io/projected/0c43c4c0-8965-43f9-9724-6970c0f15729-kube-api-access-2kd9n\") on node \"crc\" DevicePath \"\"" Mar 07 14:44:05 crc kubenswrapper[4943]: I0307 14:44:05.238481 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548244-2zdk7" event={"ID":"0c43c4c0-8965-43f9-9724-6970c0f15729","Type":"ContainerDied","Data":"a0625771c8ce07afd62861fb3b56928832a7bdb9cb169e11b25c9d51657a86a8"} Mar 07 14:44:05 crc kubenswrapper[4943]: I0307 14:44:05.238979 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a0625771c8ce07afd62861fb3b56928832a7bdb9cb169e11b25c9d51657a86a8" Mar 07 14:44:05 crc kubenswrapper[4943]: I0307 14:44:05.238651 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548244-2zdk7" Mar 07 14:44:06 crc kubenswrapper[4943]: I0307 14:44:06.074209 4943 patch_prober.go:28] interesting pod/machine-config-daemon-cpdmm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 07 14:44:06 crc kubenswrapper[4943]: I0307 14:44:06.074297 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 07 14:44:06 crc kubenswrapper[4943]: I0307 14:44:06.074382 4943 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" Mar 07 14:44:06 crc kubenswrapper[4943]: I0307 14:44:06.075451 4943 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fca9947e0c4ec525f0c6b799660a1244b1d37a4616b55a03cd6f059674d81757"} pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 07 14:44:06 crc kubenswrapper[4943]: I0307 14:44:06.075559 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerName="machine-config-daemon" containerID="cri-o://fca9947e0c4ec525f0c6b799660a1244b1d37a4616b55a03cd6f059674d81757" gracePeriod=600 Mar 07 14:44:06 crc kubenswrapper[4943]: I0307 14:44:06.247248 4943 generic.go:334] "Generic (PLEG): container finished" podID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerID="fca9947e0c4ec525f0c6b799660a1244b1d37a4616b55a03cd6f059674d81757" exitCode=0 Mar 07 14:44:06 crc kubenswrapper[4943]: I0307 14:44:06.247317 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" event={"ID":"4c1f18ea-e557-41fc-95dd-ce4e1722a61e","Type":"ContainerDied","Data":"fca9947e0c4ec525f0c6b799660a1244b1d37a4616b55a03cd6f059674d81757"} Mar 07 14:44:07 crc kubenswrapper[4943]: I0307 14:44:07.254499 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" event={"ID":"4c1f18ea-e557-41fc-95dd-ce4e1722a61e","Type":"ContainerStarted","Data":"a9980c0c2b00970e1182ba487de7ac1c1a910169389e41db4b54d88fd83453b4"} Mar 07 14:44:49 crc kubenswrapper[4943]: I0307 14:44:49.828809 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-swsg8"] Mar 07 14:44:49 crc kubenswrapper[4943]: E0307 14:44:49.829770 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c43c4c0-8965-43f9-9724-6970c0f15729" containerName="oc" Mar 07 14:44:49 crc kubenswrapper[4943]: I0307 14:44:49.829790 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c43c4c0-8965-43f9-9724-6970c0f15729" containerName="oc" Mar 07 14:44:49 crc kubenswrapper[4943]: I0307 14:44:49.829999 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c43c4c0-8965-43f9-9724-6970c0f15729" containerName="oc" Mar 07 14:44:49 crc kubenswrapper[4943]: I0307 14:44:49.830660 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-swsg8" Mar 07 14:44:49 crc kubenswrapper[4943]: I0307 14:44:49.840139 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-swsg8"] Mar 07 14:44:50 crc kubenswrapper[4943]: I0307 14:44:50.020461 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a69c3ecb-a88f-4fac-91c7-28b2b12b40a3-registry-certificates\") pod \"image-registry-66df7c8f76-swsg8\" (UID: \"a69c3ecb-a88f-4fac-91c7-28b2b12b40a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-swsg8" Mar 07 14:44:50 crc kubenswrapper[4943]: I0307 14:44:50.020496 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmls7\" (UniqueName: \"kubernetes.io/projected/a69c3ecb-a88f-4fac-91c7-28b2b12b40a3-kube-api-access-fmls7\") pod \"image-registry-66df7c8f76-swsg8\" (UID: \"a69c3ecb-a88f-4fac-91c7-28b2b12b40a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-swsg8" Mar 07 14:44:50 crc kubenswrapper[4943]: I0307 14:44:50.020528 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-swsg8\" (UID: \"a69c3ecb-a88f-4fac-91c7-28b2b12b40a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-swsg8" Mar 07 14:44:50 crc kubenswrapper[4943]: I0307 14:44:50.020653 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a69c3ecb-a88f-4fac-91c7-28b2b12b40a3-registry-tls\") pod \"image-registry-66df7c8f76-swsg8\" (UID: \"a69c3ecb-a88f-4fac-91c7-28b2b12b40a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-swsg8" Mar 07 14:44:50 crc kubenswrapper[4943]: I0307 14:44:50.020734 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a69c3ecb-a88f-4fac-91c7-28b2b12b40a3-installation-pull-secrets\") pod \"image-registry-66df7c8f76-swsg8\" (UID: \"a69c3ecb-a88f-4fac-91c7-28b2b12b40a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-swsg8" Mar 07 14:44:50 crc kubenswrapper[4943]: I0307 14:44:50.020884 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a69c3ecb-a88f-4fac-91c7-28b2b12b40a3-trusted-ca\") pod \"image-registry-66df7c8f76-swsg8\" (UID: \"a69c3ecb-a88f-4fac-91c7-28b2b12b40a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-swsg8" Mar 07 14:44:50 crc kubenswrapper[4943]: I0307 14:44:50.020922 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a69c3ecb-a88f-4fac-91c7-28b2b12b40a3-bound-sa-token\") pod \"image-registry-66df7c8f76-swsg8\" (UID: \"a69c3ecb-a88f-4fac-91c7-28b2b12b40a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-swsg8" Mar 07 14:44:50 crc kubenswrapper[4943]: I0307 14:44:50.021109 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a69c3ecb-a88f-4fac-91c7-28b2b12b40a3-ca-trust-extracted\") pod \"image-registry-66df7c8f76-swsg8\" (UID: \"a69c3ecb-a88f-4fac-91c7-28b2b12b40a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-swsg8" Mar 07 14:44:50 crc kubenswrapper[4943]: I0307 14:44:50.047715 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-swsg8\" (UID: \"a69c3ecb-a88f-4fac-91c7-28b2b12b40a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-swsg8" Mar 07 14:44:50 crc kubenswrapper[4943]: I0307 14:44:50.123148 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a69c3ecb-a88f-4fac-91c7-28b2b12b40a3-ca-trust-extracted\") pod \"image-registry-66df7c8f76-swsg8\" (UID: \"a69c3ecb-a88f-4fac-91c7-28b2b12b40a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-swsg8" Mar 07 14:44:50 crc kubenswrapper[4943]: I0307 14:44:50.123583 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a69c3ecb-a88f-4fac-91c7-28b2b12b40a3-registry-certificates\") pod \"image-registry-66df7c8f76-swsg8\" (UID: \"a69c3ecb-a88f-4fac-91c7-28b2b12b40a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-swsg8" Mar 07 14:44:50 crc kubenswrapper[4943]: I0307 14:44:50.123657 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a69c3ecb-a88f-4fac-91c7-28b2b12b40a3-ca-trust-extracted\") pod \"image-registry-66df7c8f76-swsg8\" (UID: \"a69c3ecb-a88f-4fac-91c7-28b2b12b40a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-swsg8" Mar 07 14:44:50 crc kubenswrapper[4943]: I0307 14:44:50.123818 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmls7\" (UniqueName: \"kubernetes.io/projected/a69c3ecb-a88f-4fac-91c7-28b2b12b40a3-kube-api-access-fmls7\") pod \"image-registry-66df7c8f76-swsg8\" (UID: \"a69c3ecb-a88f-4fac-91c7-28b2b12b40a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-swsg8" Mar 07 14:44:50 crc kubenswrapper[4943]: I0307 14:44:50.124188 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a69c3ecb-a88f-4fac-91c7-28b2b12b40a3-registry-tls\") pod \"image-registry-66df7c8f76-swsg8\" (UID: \"a69c3ecb-a88f-4fac-91c7-28b2b12b40a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-swsg8" Mar 07 14:44:50 crc kubenswrapper[4943]: I0307 14:44:50.124264 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a69c3ecb-a88f-4fac-91c7-28b2b12b40a3-installation-pull-secrets\") pod \"image-registry-66df7c8f76-swsg8\" (UID: \"a69c3ecb-a88f-4fac-91c7-28b2b12b40a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-swsg8" Mar 07 14:44:50 crc kubenswrapper[4943]: I0307 14:44:50.124409 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a69c3ecb-a88f-4fac-91c7-28b2b12b40a3-trusted-ca\") pod \"image-registry-66df7c8f76-swsg8\" (UID: \"a69c3ecb-a88f-4fac-91c7-28b2b12b40a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-swsg8" Mar 07 14:44:50 crc kubenswrapper[4943]: I0307 14:44:50.124458 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a69c3ecb-a88f-4fac-91c7-28b2b12b40a3-bound-sa-token\") pod \"image-registry-66df7c8f76-swsg8\" (UID: \"a69c3ecb-a88f-4fac-91c7-28b2b12b40a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-swsg8" Mar 07 14:44:50 crc kubenswrapper[4943]: I0307 14:44:50.125421 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a69c3ecb-a88f-4fac-91c7-28b2b12b40a3-registry-certificates\") pod \"image-registry-66df7c8f76-swsg8\" (UID: \"a69c3ecb-a88f-4fac-91c7-28b2b12b40a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-swsg8" Mar 07 14:44:50 crc kubenswrapper[4943]: I0307 14:44:50.125650 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a69c3ecb-a88f-4fac-91c7-28b2b12b40a3-trusted-ca\") pod \"image-registry-66df7c8f76-swsg8\" (UID: \"a69c3ecb-a88f-4fac-91c7-28b2b12b40a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-swsg8" Mar 07 14:44:50 crc kubenswrapper[4943]: I0307 14:44:50.134633 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a69c3ecb-a88f-4fac-91c7-28b2b12b40a3-registry-tls\") pod \"image-registry-66df7c8f76-swsg8\" (UID: \"a69c3ecb-a88f-4fac-91c7-28b2b12b40a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-swsg8" Mar 07 14:44:50 crc kubenswrapper[4943]: I0307 14:44:50.134671 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a69c3ecb-a88f-4fac-91c7-28b2b12b40a3-installation-pull-secrets\") pod \"image-registry-66df7c8f76-swsg8\" (UID: \"a69c3ecb-a88f-4fac-91c7-28b2b12b40a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-swsg8" Mar 07 14:44:50 crc kubenswrapper[4943]: I0307 14:44:50.153498 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a69c3ecb-a88f-4fac-91c7-28b2b12b40a3-bound-sa-token\") pod \"image-registry-66df7c8f76-swsg8\" (UID: \"a69c3ecb-a88f-4fac-91c7-28b2b12b40a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-swsg8" Mar 07 14:44:50 crc kubenswrapper[4943]: I0307 14:44:50.157601 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmls7\" (UniqueName: \"kubernetes.io/projected/a69c3ecb-a88f-4fac-91c7-28b2b12b40a3-kube-api-access-fmls7\") pod \"image-registry-66df7c8f76-swsg8\" (UID: \"a69c3ecb-a88f-4fac-91c7-28b2b12b40a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-swsg8" Mar 07 14:44:50 crc kubenswrapper[4943]: I0307 14:44:50.448725 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-swsg8" Mar 07 14:44:50 crc kubenswrapper[4943]: I0307 14:44:50.728132 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-swsg8"] Mar 07 14:44:50 crc kubenswrapper[4943]: W0307 14:44:50.741827 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda69c3ecb_a88f_4fac_91c7_28b2b12b40a3.slice/crio-c27d2551b5b2ff5551097223152acd9ae692585da7f76bd875f161e3f443ff58 WatchSource:0}: Error finding container c27d2551b5b2ff5551097223152acd9ae692585da7f76bd875f161e3f443ff58: Status 404 returned error can't find the container with id c27d2551b5b2ff5551097223152acd9ae692585da7f76bd875f161e3f443ff58 Mar 07 14:44:51 crc kubenswrapper[4943]: I0307 14:44:51.558454 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-swsg8" event={"ID":"a69c3ecb-a88f-4fac-91c7-28b2b12b40a3","Type":"ContainerStarted","Data":"98253441612a255490c682898fcf72008da1080fc4ee9fdc328f0a207db45495"} Mar 07 14:44:51 crc kubenswrapper[4943]: I0307 14:44:51.558859 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-swsg8" event={"ID":"a69c3ecb-a88f-4fac-91c7-28b2b12b40a3","Type":"ContainerStarted","Data":"c27d2551b5b2ff5551097223152acd9ae692585da7f76bd875f161e3f443ff58"} Mar 07 14:44:51 crc kubenswrapper[4943]: I0307 14:44:51.558889 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-swsg8" Mar 07 14:44:51 crc kubenswrapper[4943]: I0307 14:44:51.589054 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-swsg8" podStartSLOduration=2.589028079 podStartE2EDuration="2.589028079s" podCreationTimestamp="2026-03-07 14:44:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:44:51.583617674 +0000 UTC m=+333.535754192" watchObservedRunningTime="2026-03-07 14:44:51.589028079 +0000 UTC m=+333.541164587" Mar 07 14:45:00 crc kubenswrapper[4943]: I0307 14:45:00.159875 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29548245-6qzc4"] Mar 07 14:45:00 crc kubenswrapper[4943]: I0307 14:45:00.161910 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29548245-6qzc4" Mar 07 14:45:00 crc kubenswrapper[4943]: I0307 14:45:00.166253 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 07 14:45:00 crc kubenswrapper[4943]: I0307 14:45:00.168429 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 07 14:45:00 crc kubenswrapper[4943]: I0307 14:45:00.184046 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29548245-6qzc4"] Mar 07 14:45:00 crc kubenswrapper[4943]: I0307 14:45:00.302047 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1af9c9ab-147e-48d7-87ff-8e9ec1ae7f25-secret-volume\") pod \"collect-profiles-29548245-6qzc4\" (UID: \"1af9c9ab-147e-48d7-87ff-8e9ec1ae7f25\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29548245-6qzc4" Mar 07 14:45:00 crc kubenswrapper[4943]: I0307 14:45:00.302114 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwpjp\" (UniqueName: \"kubernetes.io/projected/1af9c9ab-147e-48d7-87ff-8e9ec1ae7f25-kube-api-access-zwpjp\") pod \"collect-profiles-29548245-6qzc4\" (UID: \"1af9c9ab-147e-48d7-87ff-8e9ec1ae7f25\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29548245-6qzc4" Mar 07 14:45:00 crc kubenswrapper[4943]: I0307 14:45:00.302138 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1af9c9ab-147e-48d7-87ff-8e9ec1ae7f25-config-volume\") pod \"collect-profiles-29548245-6qzc4\" (UID: \"1af9c9ab-147e-48d7-87ff-8e9ec1ae7f25\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29548245-6qzc4" Mar 07 14:45:00 crc kubenswrapper[4943]: I0307 14:45:00.404388 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1af9c9ab-147e-48d7-87ff-8e9ec1ae7f25-config-volume\") pod \"collect-profiles-29548245-6qzc4\" (UID: \"1af9c9ab-147e-48d7-87ff-8e9ec1ae7f25\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29548245-6qzc4" Mar 07 14:45:00 crc kubenswrapper[4943]: I0307 14:45:00.404622 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1af9c9ab-147e-48d7-87ff-8e9ec1ae7f25-secret-volume\") pod \"collect-profiles-29548245-6qzc4\" (UID: \"1af9c9ab-147e-48d7-87ff-8e9ec1ae7f25\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29548245-6qzc4" Mar 07 14:45:00 crc kubenswrapper[4943]: I0307 14:45:00.404873 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwpjp\" (UniqueName: \"kubernetes.io/projected/1af9c9ab-147e-48d7-87ff-8e9ec1ae7f25-kube-api-access-zwpjp\") pod \"collect-profiles-29548245-6qzc4\" (UID: \"1af9c9ab-147e-48d7-87ff-8e9ec1ae7f25\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29548245-6qzc4" Mar 07 14:45:00 crc kubenswrapper[4943]: I0307 14:45:00.406230 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1af9c9ab-147e-48d7-87ff-8e9ec1ae7f25-config-volume\") pod \"collect-profiles-29548245-6qzc4\" (UID: \"1af9c9ab-147e-48d7-87ff-8e9ec1ae7f25\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29548245-6qzc4" Mar 07 14:45:00 crc kubenswrapper[4943]: I0307 14:45:00.415464 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1af9c9ab-147e-48d7-87ff-8e9ec1ae7f25-secret-volume\") pod \"collect-profiles-29548245-6qzc4\" (UID: \"1af9c9ab-147e-48d7-87ff-8e9ec1ae7f25\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29548245-6qzc4" Mar 07 14:45:00 crc kubenswrapper[4943]: I0307 14:45:00.436135 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwpjp\" (UniqueName: \"kubernetes.io/projected/1af9c9ab-147e-48d7-87ff-8e9ec1ae7f25-kube-api-access-zwpjp\") pod \"collect-profiles-29548245-6qzc4\" (UID: \"1af9c9ab-147e-48d7-87ff-8e9ec1ae7f25\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29548245-6qzc4" Mar 07 14:45:00 crc kubenswrapper[4943]: I0307 14:45:00.498784 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29548245-6qzc4" Mar 07 14:45:01 crc kubenswrapper[4943]: I0307 14:45:01.015715 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29548245-6qzc4"] Mar 07 14:45:01 crc kubenswrapper[4943]: W0307 14:45:01.020874 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1af9c9ab_147e_48d7_87ff_8e9ec1ae7f25.slice/crio-e85980dcaec19df995b9e2771d2f4b6ad9449aad3d865d908043737730921622 WatchSource:0}: Error finding container e85980dcaec19df995b9e2771d2f4b6ad9449aad3d865d908043737730921622: Status 404 returned error can't find the container with id e85980dcaec19df995b9e2771d2f4b6ad9449aad3d865d908043737730921622 Mar 07 14:45:01 crc kubenswrapper[4943]: I0307 14:45:01.648591 4943 generic.go:334] "Generic (PLEG): container finished" podID="1af9c9ab-147e-48d7-87ff-8e9ec1ae7f25" containerID="e5f4a0f85cfc5b2f4d4283ab108405768677a21de107c56b50634e42c6d8abd9" exitCode=0 Mar 07 14:45:01 crc kubenswrapper[4943]: I0307 14:45:01.648708 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29548245-6qzc4" event={"ID":"1af9c9ab-147e-48d7-87ff-8e9ec1ae7f25","Type":"ContainerDied","Data":"e5f4a0f85cfc5b2f4d4283ab108405768677a21de107c56b50634e42c6d8abd9"} Mar 07 14:45:01 crc kubenswrapper[4943]: I0307 14:45:01.649142 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29548245-6qzc4" event={"ID":"1af9c9ab-147e-48d7-87ff-8e9ec1ae7f25","Type":"ContainerStarted","Data":"e85980dcaec19df995b9e2771d2f4b6ad9449aad3d865d908043737730921622"} Mar 07 14:45:03 crc kubenswrapper[4943]: I0307 14:45:03.011582 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29548245-6qzc4" Mar 07 14:45:03 crc kubenswrapper[4943]: I0307 14:45:03.156568 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zwpjp\" (UniqueName: \"kubernetes.io/projected/1af9c9ab-147e-48d7-87ff-8e9ec1ae7f25-kube-api-access-zwpjp\") pod \"1af9c9ab-147e-48d7-87ff-8e9ec1ae7f25\" (UID: \"1af9c9ab-147e-48d7-87ff-8e9ec1ae7f25\") " Mar 07 14:45:03 crc kubenswrapper[4943]: I0307 14:45:03.156637 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1af9c9ab-147e-48d7-87ff-8e9ec1ae7f25-secret-volume\") pod \"1af9c9ab-147e-48d7-87ff-8e9ec1ae7f25\" (UID: \"1af9c9ab-147e-48d7-87ff-8e9ec1ae7f25\") " Mar 07 14:45:03 crc kubenswrapper[4943]: I0307 14:45:03.156710 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1af9c9ab-147e-48d7-87ff-8e9ec1ae7f25-config-volume\") pod \"1af9c9ab-147e-48d7-87ff-8e9ec1ae7f25\" (UID: \"1af9c9ab-147e-48d7-87ff-8e9ec1ae7f25\") " Mar 07 14:45:03 crc kubenswrapper[4943]: I0307 14:45:03.157452 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1af9c9ab-147e-48d7-87ff-8e9ec1ae7f25-config-volume" (OuterVolumeSpecName: "config-volume") pod "1af9c9ab-147e-48d7-87ff-8e9ec1ae7f25" (UID: "1af9c9ab-147e-48d7-87ff-8e9ec1ae7f25"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:45:03 crc kubenswrapper[4943]: I0307 14:45:03.162962 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1af9c9ab-147e-48d7-87ff-8e9ec1ae7f25-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1af9c9ab-147e-48d7-87ff-8e9ec1ae7f25" (UID: "1af9c9ab-147e-48d7-87ff-8e9ec1ae7f25"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:45:03 crc kubenswrapper[4943]: I0307 14:45:03.168025 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1af9c9ab-147e-48d7-87ff-8e9ec1ae7f25-kube-api-access-zwpjp" (OuterVolumeSpecName: "kube-api-access-zwpjp") pod "1af9c9ab-147e-48d7-87ff-8e9ec1ae7f25" (UID: "1af9c9ab-147e-48d7-87ff-8e9ec1ae7f25"). InnerVolumeSpecName "kube-api-access-zwpjp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:45:03 crc kubenswrapper[4943]: I0307 14:45:03.258684 4943 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1af9c9ab-147e-48d7-87ff-8e9ec1ae7f25-config-volume\") on node \"crc\" DevicePath \"\"" Mar 07 14:45:03 crc kubenswrapper[4943]: I0307 14:45:03.258747 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zwpjp\" (UniqueName: \"kubernetes.io/projected/1af9c9ab-147e-48d7-87ff-8e9ec1ae7f25-kube-api-access-zwpjp\") on node \"crc\" DevicePath \"\"" Mar 07 14:45:03 crc kubenswrapper[4943]: I0307 14:45:03.258769 4943 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1af9c9ab-147e-48d7-87ff-8e9ec1ae7f25-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 07 14:45:03 crc kubenswrapper[4943]: I0307 14:45:03.666762 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29548245-6qzc4" event={"ID":"1af9c9ab-147e-48d7-87ff-8e9ec1ae7f25","Type":"ContainerDied","Data":"e85980dcaec19df995b9e2771d2f4b6ad9449aad3d865d908043737730921622"} Mar 07 14:45:03 crc kubenswrapper[4943]: I0307 14:45:03.666826 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e85980dcaec19df995b9e2771d2f4b6ad9449aad3d865d908043737730921622" Mar 07 14:45:03 crc kubenswrapper[4943]: I0307 14:45:03.667309 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29548245-6qzc4" Mar 07 14:45:10 crc kubenswrapper[4943]: I0307 14:45:10.456891 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-swsg8" Mar 07 14:45:10 crc kubenswrapper[4943]: I0307 14:45:10.532076 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-r82fw"] Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.238774 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dfqzr"] Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.239572 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dfqzr" podUID="0d1657d7-a787-4648-91f6-e0ff987cede2" containerName="registry-server" containerID="cri-o://2a6e8e90c36b929a9f6b254045d721389004a538bf9b2b26618515396b8cecda" gracePeriod=30 Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.246996 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jrc6h"] Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.247715 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jrc6h" podUID="f10fa8f5-f504-40c9-81a5-d1658c1ef268" containerName="registry-server" containerID="cri-o://2589f77538011f2ab019b72637e957313f033e18e37f87b0825a4312ec83bbd5" gracePeriod=30 Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.262066 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sqsdb"] Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.262453 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-sqsdb" podUID="712a2a31-4205-4346-9a32-858a77615eb6" containerName="marketplace-operator" containerID="cri-o://08e935ddfdfcb02fe9bdeb0fb05b2807df84fbf1297330bd13c1372e83260314" gracePeriod=30 Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.282317 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xwdpr"] Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.282845 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xwdpr" podUID="254f0c0f-ee54-434e-bf7b-bf2e5274fdc9" containerName="registry-server" containerID="cri-o://786d27c9685fe23593de377ea7485e2d68bf44446fa5662aad6df02c8625e99b" gracePeriod=30 Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.296693 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cpfgk"] Mar 07 14:45:12 crc kubenswrapper[4943]: E0307 14:45:12.297074 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1af9c9ab-147e-48d7-87ff-8e9ec1ae7f25" containerName="collect-profiles" Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.297099 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="1af9c9ab-147e-48d7-87ff-8e9ec1ae7f25" containerName="collect-profiles" Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.297222 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="1af9c9ab-147e-48d7-87ff-8e9ec1ae7f25" containerName="collect-profiles" Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.297747 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-cpfgk" Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.314101 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-f95bw"] Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.314500 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-f95bw" podUID="a41f89c1-5f6d-46bc-801b-a29ebebf4468" containerName="registry-server" containerID="cri-o://1b1573f313b949b3ea11b97a155d031e94aeebb27dc909c06ba03f22ba4e6cfa" gracePeriod=30 Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.376029 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cpfgk"] Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.402461 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9ba51437-0e4b-4aca-87c7-1152f9c2a461-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-cpfgk\" (UID: \"9ba51437-0e4b-4aca-87c7-1152f9c2a461\") " pod="openshift-marketplace/marketplace-operator-79b997595-cpfgk" Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.402679 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxblj\" (UniqueName: \"kubernetes.io/projected/9ba51437-0e4b-4aca-87c7-1152f9c2a461-kube-api-access-fxblj\") pod \"marketplace-operator-79b997595-cpfgk\" (UID: \"9ba51437-0e4b-4aca-87c7-1152f9c2a461\") " pod="openshift-marketplace/marketplace-operator-79b997595-cpfgk" Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.402735 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9ba51437-0e4b-4aca-87c7-1152f9c2a461-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-cpfgk\" (UID: \"9ba51437-0e4b-4aca-87c7-1152f9c2a461\") " pod="openshift-marketplace/marketplace-operator-79b997595-cpfgk" Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.503847 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxblj\" (UniqueName: \"kubernetes.io/projected/9ba51437-0e4b-4aca-87c7-1152f9c2a461-kube-api-access-fxblj\") pod \"marketplace-operator-79b997595-cpfgk\" (UID: \"9ba51437-0e4b-4aca-87c7-1152f9c2a461\") " pod="openshift-marketplace/marketplace-operator-79b997595-cpfgk" Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.503897 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9ba51437-0e4b-4aca-87c7-1152f9c2a461-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-cpfgk\" (UID: \"9ba51437-0e4b-4aca-87c7-1152f9c2a461\") " pod="openshift-marketplace/marketplace-operator-79b997595-cpfgk" Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.503946 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9ba51437-0e4b-4aca-87c7-1152f9c2a461-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-cpfgk\" (UID: \"9ba51437-0e4b-4aca-87c7-1152f9c2a461\") " pod="openshift-marketplace/marketplace-operator-79b997595-cpfgk" Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.505414 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9ba51437-0e4b-4aca-87c7-1152f9c2a461-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-cpfgk\" (UID: \"9ba51437-0e4b-4aca-87c7-1152f9c2a461\") " pod="openshift-marketplace/marketplace-operator-79b997595-cpfgk" Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.511824 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9ba51437-0e4b-4aca-87c7-1152f9c2a461-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-cpfgk\" (UID: \"9ba51437-0e4b-4aca-87c7-1152f9c2a461\") " pod="openshift-marketplace/marketplace-operator-79b997595-cpfgk" Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.518887 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxblj\" (UniqueName: \"kubernetes.io/projected/9ba51437-0e4b-4aca-87c7-1152f9c2a461-kube-api-access-fxblj\") pod \"marketplace-operator-79b997595-cpfgk\" (UID: \"9ba51437-0e4b-4aca-87c7-1152f9c2a461\") " pod="openshift-marketplace/marketplace-operator-79b997595-cpfgk" Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.735183 4943 generic.go:334] "Generic (PLEG): container finished" podID="712a2a31-4205-4346-9a32-858a77615eb6" containerID="08e935ddfdfcb02fe9bdeb0fb05b2807df84fbf1297330bd13c1372e83260314" exitCode=0 Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.735275 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-sqsdb" event={"ID":"712a2a31-4205-4346-9a32-858a77615eb6","Type":"ContainerDied","Data":"08e935ddfdfcb02fe9bdeb0fb05b2807df84fbf1297330bd13c1372e83260314"} Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.735348 4943 scope.go:117] "RemoveContainer" containerID="51fa91e109cdf7e64fa128d5ebe2fa2b77ae12ee40c73925027012b869163b0c" Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.736288 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-cpfgk" Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.739290 4943 generic.go:334] "Generic (PLEG): container finished" podID="f10fa8f5-f504-40c9-81a5-d1658c1ef268" containerID="2589f77538011f2ab019b72637e957313f033e18e37f87b0825a4312ec83bbd5" exitCode=0 Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.739351 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jrc6h" event={"ID":"f10fa8f5-f504-40c9-81a5-d1658c1ef268","Type":"ContainerDied","Data":"2589f77538011f2ab019b72637e957313f033e18e37f87b0825a4312ec83bbd5"} Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.740340 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dfqzr" Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.744358 4943 generic.go:334] "Generic (PLEG): container finished" podID="0d1657d7-a787-4648-91f6-e0ff987cede2" containerID="2a6e8e90c36b929a9f6b254045d721389004a538bf9b2b26618515396b8cecda" exitCode=0 Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.744415 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dfqzr" event={"ID":"0d1657d7-a787-4648-91f6-e0ff987cede2","Type":"ContainerDied","Data":"2a6e8e90c36b929a9f6b254045d721389004a538bf9b2b26618515396b8cecda"} Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.744443 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dfqzr" event={"ID":"0d1657d7-a787-4648-91f6-e0ff987cede2","Type":"ContainerDied","Data":"170f3f2e2722a619cc0ea7dcfe8df679d89d9665f25432341c3067e7f0ebb1c0"} Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.746031 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xwdpr" Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.746430 4943 generic.go:334] "Generic (PLEG): container finished" podID="a41f89c1-5f6d-46bc-801b-a29ebebf4468" containerID="1b1573f313b949b3ea11b97a155d031e94aeebb27dc909c06ba03f22ba4e6cfa" exitCode=0 Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.746516 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f95bw" event={"ID":"a41f89c1-5f6d-46bc-801b-a29ebebf4468","Type":"ContainerDied","Data":"1b1573f313b949b3ea11b97a155d031e94aeebb27dc909c06ba03f22ba4e6cfa"} Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.761211 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xwdpr" Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.761230 4943 generic.go:334] "Generic (PLEG): container finished" podID="254f0c0f-ee54-434e-bf7b-bf2e5274fdc9" containerID="786d27c9685fe23593de377ea7485e2d68bf44446fa5662aad6df02c8625e99b" exitCode=0 Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.781704 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jrc6h" Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.790173 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xwdpr" event={"ID":"254f0c0f-ee54-434e-bf7b-bf2e5274fdc9","Type":"ContainerDied","Data":"786d27c9685fe23593de377ea7485e2d68bf44446fa5662aad6df02c8625e99b"} Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.790246 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xwdpr" event={"ID":"254f0c0f-ee54-434e-bf7b-bf2e5274fdc9","Type":"ContainerDied","Data":"001de696f44d832458359bbd56358a8936643cfa6074a217dee572b906259c5c"} Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.807863 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/254f0c0f-ee54-434e-bf7b-bf2e5274fdc9-utilities\") pod \"254f0c0f-ee54-434e-bf7b-bf2e5274fdc9\" (UID: \"254f0c0f-ee54-434e-bf7b-bf2e5274fdc9\") " Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.807940 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d1657d7-a787-4648-91f6-e0ff987cede2-utilities\") pod \"0d1657d7-a787-4648-91f6-e0ff987cede2\" (UID: \"0d1657d7-a787-4648-91f6-e0ff987cede2\") " Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.808029 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d1657d7-a787-4648-91f6-e0ff987cede2-catalog-content\") pod \"0d1657d7-a787-4648-91f6-e0ff987cede2\" (UID: \"0d1657d7-a787-4648-91f6-e0ff987cede2\") " Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.810964 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d1657d7-a787-4648-91f6-e0ff987cede2-utilities" (OuterVolumeSpecName: "utilities") pod "0d1657d7-a787-4648-91f6-e0ff987cede2" (UID: "0d1657d7-a787-4648-91f6-e0ff987cede2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.811417 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-582l8\" (UniqueName: \"kubernetes.io/projected/0d1657d7-a787-4648-91f6-e0ff987cede2-kube-api-access-582l8\") pod \"0d1657d7-a787-4648-91f6-e0ff987cede2\" (UID: \"0d1657d7-a787-4648-91f6-e0ff987cede2\") " Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.812057 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kmd9h\" (UniqueName: \"kubernetes.io/projected/254f0c0f-ee54-434e-bf7b-bf2e5274fdc9-kube-api-access-kmd9h\") pod \"254f0c0f-ee54-434e-bf7b-bf2e5274fdc9\" (UID: \"254f0c0f-ee54-434e-bf7b-bf2e5274fdc9\") " Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.812090 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/254f0c0f-ee54-434e-bf7b-bf2e5274fdc9-catalog-content\") pod \"254f0c0f-ee54-434e-bf7b-bf2e5274fdc9\" (UID: \"254f0c0f-ee54-434e-bf7b-bf2e5274fdc9\") " Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.817672 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/254f0c0f-ee54-434e-bf7b-bf2e5274fdc9-utilities" (OuterVolumeSpecName: "utilities") pod "254f0c0f-ee54-434e-bf7b-bf2e5274fdc9" (UID: "254f0c0f-ee54-434e-bf7b-bf2e5274fdc9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.818604 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d1657d7-a787-4648-91f6-e0ff987cede2-kube-api-access-582l8" (OuterVolumeSpecName: "kube-api-access-582l8") pod "0d1657d7-a787-4648-91f6-e0ff987cede2" (UID: "0d1657d7-a787-4648-91f6-e0ff987cede2"). InnerVolumeSpecName "kube-api-access-582l8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.818722 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/254f0c0f-ee54-434e-bf7b-bf2e5274fdc9-kube-api-access-kmd9h" (OuterVolumeSpecName: "kube-api-access-kmd9h") pod "254f0c0f-ee54-434e-bf7b-bf2e5274fdc9" (UID: "254f0c0f-ee54-434e-bf7b-bf2e5274fdc9"). InnerVolumeSpecName "kube-api-access-kmd9h". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.819500 4943 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/254f0c0f-ee54-434e-bf7b-bf2e5274fdc9-utilities\") on node \"crc\" DevicePath \"\"" Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.819522 4943 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d1657d7-a787-4648-91f6-e0ff987cede2-utilities\") on node \"crc\" DevicePath \"\"" Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.819544 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-582l8\" (UniqueName: \"kubernetes.io/projected/0d1657d7-a787-4648-91f6-e0ff987cede2-kube-api-access-582l8\") on node \"crc\" DevicePath \"\"" Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.819553 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kmd9h\" (UniqueName: \"kubernetes.io/projected/254f0c0f-ee54-434e-bf7b-bf2e5274fdc9-kube-api-access-kmd9h\") on node \"crc\" DevicePath \"\"" Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.837046 4943 scope.go:117] "RemoveContainer" containerID="2a6e8e90c36b929a9f6b254045d721389004a538bf9b2b26618515396b8cecda" Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.848866 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f95bw" Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.856290 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/254f0c0f-ee54-434e-bf7b-bf2e5274fdc9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "254f0c0f-ee54-434e-bf7b-bf2e5274fdc9" (UID: "254f0c0f-ee54-434e-bf7b-bf2e5274fdc9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.881564 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d1657d7-a787-4648-91f6-e0ff987cede2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0d1657d7-a787-4648-91f6-e0ff987cede2" (UID: "0d1657d7-a787-4648-91f6-e0ff987cede2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.909878 4943 scope.go:117] "RemoveContainer" containerID="fac900a8e0187f803bf8032d0085fb713054b6666413f52381eaf48989c9cb80" Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.920493 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7mt9r\" (UniqueName: \"kubernetes.io/projected/a41f89c1-5f6d-46bc-801b-a29ebebf4468-kube-api-access-7mt9r\") pod \"a41f89c1-5f6d-46bc-801b-a29ebebf4468\" (UID: \"a41f89c1-5f6d-46bc-801b-a29ebebf4468\") " Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.920537 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a41f89c1-5f6d-46bc-801b-a29ebebf4468-utilities\") pod \"a41f89c1-5f6d-46bc-801b-a29ebebf4468\" (UID: \"a41f89c1-5f6d-46bc-801b-a29ebebf4468\") " Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.920559 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f10fa8f5-f504-40c9-81a5-d1658c1ef268-utilities\") pod \"f10fa8f5-f504-40c9-81a5-d1658c1ef268\" (UID: \"f10fa8f5-f504-40c9-81a5-d1658c1ef268\") " Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.920592 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a41f89c1-5f6d-46bc-801b-a29ebebf4468-catalog-content\") pod \"a41f89c1-5f6d-46bc-801b-a29ebebf4468\" (UID: \"a41f89c1-5f6d-46bc-801b-a29ebebf4468\") " Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.921389 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f10fa8f5-f504-40c9-81a5-d1658c1ef268-catalog-content\") pod \"f10fa8f5-f504-40c9-81a5-d1658c1ef268\" (UID: \"f10fa8f5-f504-40c9-81a5-d1658c1ef268\") " Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.921421 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wthgk\" (UniqueName: \"kubernetes.io/projected/f10fa8f5-f504-40c9-81a5-d1658c1ef268-kube-api-access-wthgk\") pod \"f10fa8f5-f504-40c9-81a5-d1658c1ef268\" (UID: \"f10fa8f5-f504-40c9-81a5-d1658c1ef268\") " Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.921317 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f10fa8f5-f504-40c9-81a5-d1658c1ef268-utilities" (OuterVolumeSpecName: "utilities") pod "f10fa8f5-f504-40c9-81a5-d1658c1ef268" (UID: "f10fa8f5-f504-40c9-81a5-d1658c1ef268"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.921443 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a41f89c1-5f6d-46bc-801b-a29ebebf4468-utilities" (OuterVolumeSpecName: "utilities") pod "a41f89c1-5f6d-46bc-801b-a29ebebf4468" (UID: "a41f89c1-5f6d-46bc-801b-a29ebebf4468"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.925361 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a41f89c1-5f6d-46bc-801b-a29ebebf4468-kube-api-access-7mt9r" (OuterVolumeSpecName: "kube-api-access-7mt9r") pod "a41f89c1-5f6d-46bc-801b-a29ebebf4468" (UID: "a41f89c1-5f6d-46bc-801b-a29ebebf4468"). InnerVolumeSpecName "kube-api-access-7mt9r". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.925803 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f10fa8f5-f504-40c9-81a5-d1658c1ef268-kube-api-access-wthgk" (OuterVolumeSpecName: "kube-api-access-wthgk") pod "f10fa8f5-f504-40c9-81a5-d1658c1ef268" (UID: "f10fa8f5-f504-40c9-81a5-d1658c1ef268"). InnerVolumeSpecName "kube-api-access-wthgk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.928627 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7mt9r\" (UniqueName: \"kubernetes.io/projected/a41f89c1-5f6d-46bc-801b-a29ebebf4468-kube-api-access-7mt9r\") on node \"crc\" DevicePath \"\"" Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.928657 4943 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/254f0c0f-ee54-434e-bf7b-bf2e5274fdc9-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.928669 4943 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a41f89c1-5f6d-46bc-801b-a29ebebf4468-utilities\") on node \"crc\" DevicePath \"\"" Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.928679 4943 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f10fa8f5-f504-40c9-81a5-d1658c1ef268-utilities\") on node \"crc\" DevicePath \"\"" Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.928690 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wthgk\" (UniqueName: \"kubernetes.io/projected/f10fa8f5-f504-40c9-81a5-d1658c1ef268-kube-api-access-wthgk\") on node \"crc\" DevicePath \"\"" Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.928700 4943 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d1657d7-a787-4648-91f6-e0ff987cede2-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.941786 4943 scope.go:117] "RemoveContainer" containerID="a07aa093e1d3c936f3f368f109c8afbcd14d4091f79d0b410b6d4258aa5fe893" Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.975871 4943 scope.go:117] "RemoveContainer" containerID="2a6e8e90c36b929a9f6b254045d721389004a538bf9b2b26618515396b8cecda" Mar 07 14:45:12 crc kubenswrapper[4943]: E0307 14:45:12.979392 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a6e8e90c36b929a9f6b254045d721389004a538bf9b2b26618515396b8cecda\": container with ID starting with 2a6e8e90c36b929a9f6b254045d721389004a538bf9b2b26618515396b8cecda not found: ID does not exist" containerID="2a6e8e90c36b929a9f6b254045d721389004a538bf9b2b26618515396b8cecda" Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.979426 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a6e8e90c36b929a9f6b254045d721389004a538bf9b2b26618515396b8cecda"} err="failed to get container status \"2a6e8e90c36b929a9f6b254045d721389004a538bf9b2b26618515396b8cecda\": rpc error: code = NotFound desc = could not find container \"2a6e8e90c36b929a9f6b254045d721389004a538bf9b2b26618515396b8cecda\": container with ID starting with 2a6e8e90c36b929a9f6b254045d721389004a538bf9b2b26618515396b8cecda not found: ID does not exist" Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.979453 4943 scope.go:117] "RemoveContainer" containerID="fac900a8e0187f803bf8032d0085fb713054b6666413f52381eaf48989c9cb80" Mar 07 14:45:12 crc kubenswrapper[4943]: E0307 14:45:12.979622 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fac900a8e0187f803bf8032d0085fb713054b6666413f52381eaf48989c9cb80\": container with ID starting with fac900a8e0187f803bf8032d0085fb713054b6666413f52381eaf48989c9cb80 not found: ID does not exist" containerID="fac900a8e0187f803bf8032d0085fb713054b6666413f52381eaf48989c9cb80" Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.979643 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fac900a8e0187f803bf8032d0085fb713054b6666413f52381eaf48989c9cb80"} err="failed to get container status \"fac900a8e0187f803bf8032d0085fb713054b6666413f52381eaf48989c9cb80\": rpc error: code = NotFound desc = could not find container \"fac900a8e0187f803bf8032d0085fb713054b6666413f52381eaf48989c9cb80\": container with ID starting with fac900a8e0187f803bf8032d0085fb713054b6666413f52381eaf48989c9cb80 not found: ID does not exist" Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.979659 4943 scope.go:117] "RemoveContainer" containerID="a07aa093e1d3c936f3f368f109c8afbcd14d4091f79d0b410b6d4258aa5fe893" Mar 07 14:45:12 crc kubenswrapper[4943]: E0307 14:45:12.979896 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a07aa093e1d3c936f3f368f109c8afbcd14d4091f79d0b410b6d4258aa5fe893\": container with ID starting with a07aa093e1d3c936f3f368f109c8afbcd14d4091f79d0b410b6d4258aa5fe893 not found: ID does not exist" containerID="a07aa093e1d3c936f3f368f109c8afbcd14d4091f79d0b410b6d4258aa5fe893" Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.979917 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a07aa093e1d3c936f3f368f109c8afbcd14d4091f79d0b410b6d4258aa5fe893"} err="failed to get container status \"a07aa093e1d3c936f3f368f109c8afbcd14d4091f79d0b410b6d4258aa5fe893\": rpc error: code = NotFound desc = could not find container \"a07aa093e1d3c936f3f368f109c8afbcd14d4091f79d0b410b6d4258aa5fe893\": container with ID starting with a07aa093e1d3c936f3f368f109c8afbcd14d4091f79d0b410b6d4258aa5fe893 not found: ID does not exist" Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.979946 4943 scope.go:117] "RemoveContainer" containerID="786d27c9685fe23593de377ea7485e2d68bf44446fa5662aad6df02c8625e99b" Mar 07 14:45:12 crc kubenswrapper[4943]: I0307 14:45:12.995773 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f10fa8f5-f504-40c9-81a5-d1658c1ef268-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f10fa8f5-f504-40c9-81a5-d1658c1ef268" (UID: "f10fa8f5-f504-40c9-81a5-d1658c1ef268"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:45:13 crc kubenswrapper[4943]: I0307 14:45:13.028800 4943 scope.go:117] "RemoveContainer" containerID="0163d96eb2ce8fb71907d1a0fc94f1275f85e77a54c20cb56ce46c9882453aec" Mar 07 14:45:13 crc kubenswrapper[4943]: I0307 14:45:13.030524 4943 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f10fa8f5-f504-40c9-81a5-d1658c1ef268-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 07 14:45:13 crc kubenswrapper[4943]: I0307 14:45:13.039294 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cpfgk"] Mar 07 14:45:13 crc kubenswrapper[4943]: W0307 14:45:13.039762 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9ba51437_0e4b_4aca_87c7_1152f9c2a461.slice/crio-9b382521ffa7fc6348f78203838f142e9bcfd68ca7f2c23aa32ddd113ed3b70e WatchSource:0}: Error finding container 9b382521ffa7fc6348f78203838f142e9bcfd68ca7f2c23aa32ddd113ed3b70e: Status 404 returned error can't find the container with id 9b382521ffa7fc6348f78203838f142e9bcfd68ca7f2c23aa32ddd113ed3b70e Mar 07 14:45:13 crc kubenswrapper[4943]: I0307 14:45:13.047811 4943 scope.go:117] "RemoveContainer" containerID="85734361061e123d9cb4c89ee7923bcd0395bfaef11c025edf86d71024ca1137" Mar 07 14:45:13 crc kubenswrapper[4943]: I0307 14:45:13.070148 4943 scope.go:117] "RemoveContainer" containerID="786d27c9685fe23593de377ea7485e2d68bf44446fa5662aad6df02c8625e99b" Mar 07 14:45:13 crc kubenswrapper[4943]: E0307 14:45:13.071008 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"786d27c9685fe23593de377ea7485e2d68bf44446fa5662aad6df02c8625e99b\": container with ID starting with 786d27c9685fe23593de377ea7485e2d68bf44446fa5662aad6df02c8625e99b not found: ID does not exist" containerID="786d27c9685fe23593de377ea7485e2d68bf44446fa5662aad6df02c8625e99b" Mar 07 14:45:13 crc kubenswrapper[4943]: I0307 14:45:13.071055 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"786d27c9685fe23593de377ea7485e2d68bf44446fa5662aad6df02c8625e99b"} err="failed to get container status \"786d27c9685fe23593de377ea7485e2d68bf44446fa5662aad6df02c8625e99b\": rpc error: code = NotFound desc = could not find container \"786d27c9685fe23593de377ea7485e2d68bf44446fa5662aad6df02c8625e99b\": container with ID starting with 786d27c9685fe23593de377ea7485e2d68bf44446fa5662aad6df02c8625e99b not found: ID does not exist" Mar 07 14:45:13 crc kubenswrapper[4943]: I0307 14:45:13.071088 4943 scope.go:117] "RemoveContainer" containerID="0163d96eb2ce8fb71907d1a0fc94f1275f85e77a54c20cb56ce46c9882453aec" Mar 07 14:45:13 crc kubenswrapper[4943]: E0307 14:45:13.071422 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0163d96eb2ce8fb71907d1a0fc94f1275f85e77a54c20cb56ce46c9882453aec\": container with ID starting with 0163d96eb2ce8fb71907d1a0fc94f1275f85e77a54c20cb56ce46c9882453aec not found: ID does not exist" containerID="0163d96eb2ce8fb71907d1a0fc94f1275f85e77a54c20cb56ce46c9882453aec" Mar 07 14:45:13 crc kubenswrapper[4943]: I0307 14:45:13.071454 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0163d96eb2ce8fb71907d1a0fc94f1275f85e77a54c20cb56ce46c9882453aec"} err="failed to get container status \"0163d96eb2ce8fb71907d1a0fc94f1275f85e77a54c20cb56ce46c9882453aec\": rpc error: code = NotFound desc = could not find container \"0163d96eb2ce8fb71907d1a0fc94f1275f85e77a54c20cb56ce46c9882453aec\": container with ID starting with 0163d96eb2ce8fb71907d1a0fc94f1275f85e77a54c20cb56ce46c9882453aec not found: ID does not exist" Mar 07 14:45:13 crc kubenswrapper[4943]: I0307 14:45:13.071472 4943 scope.go:117] "RemoveContainer" containerID="85734361061e123d9cb4c89ee7923bcd0395bfaef11c025edf86d71024ca1137" Mar 07 14:45:13 crc kubenswrapper[4943]: E0307 14:45:13.071769 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85734361061e123d9cb4c89ee7923bcd0395bfaef11c025edf86d71024ca1137\": container with ID starting with 85734361061e123d9cb4c89ee7923bcd0395bfaef11c025edf86d71024ca1137 not found: ID does not exist" containerID="85734361061e123d9cb4c89ee7923bcd0395bfaef11c025edf86d71024ca1137" Mar 07 14:45:13 crc kubenswrapper[4943]: I0307 14:45:13.071826 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85734361061e123d9cb4c89ee7923bcd0395bfaef11c025edf86d71024ca1137"} err="failed to get container status \"85734361061e123d9cb4c89ee7923bcd0395bfaef11c025edf86d71024ca1137\": rpc error: code = NotFound desc = could not find container \"85734361061e123d9cb4c89ee7923bcd0395bfaef11c025edf86d71024ca1137\": container with ID starting with 85734361061e123d9cb4c89ee7923bcd0395bfaef11c025edf86d71024ca1137 not found: ID does not exist" Mar 07 14:45:13 crc kubenswrapper[4943]: I0307 14:45:13.108356 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xwdpr"] Mar 07 14:45:13 crc kubenswrapper[4943]: I0307 14:45:13.111123 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xwdpr"] Mar 07 14:45:13 crc kubenswrapper[4943]: I0307 14:45:13.109400 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a41f89c1-5f6d-46bc-801b-a29ebebf4468-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a41f89c1-5f6d-46bc-801b-a29ebebf4468" (UID: "a41f89c1-5f6d-46bc-801b-a29ebebf4468"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:45:13 crc kubenswrapper[4943]: I0307 14:45:13.131700 4943 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a41f89c1-5f6d-46bc-801b-a29ebebf4468-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 07 14:45:13 crc kubenswrapper[4943]: I0307 14:45:13.150713 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-sqsdb" Mar 07 14:45:13 crc kubenswrapper[4943]: I0307 14:45:13.233189 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/712a2a31-4205-4346-9a32-858a77615eb6-marketplace-trusted-ca\") pod \"712a2a31-4205-4346-9a32-858a77615eb6\" (UID: \"712a2a31-4205-4346-9a32-858a77615eb6\") " Mar 07 14:45:13 crc kubenswrapper[4943]: I0307 14:45:13.233279 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-92fpw\" (UniqueName: \"kubernetes.io/projected/712a2a31-4205-4346-9a32-858a77615eb6-kube-api-access-92fpw\") pod \"712a2a31-4205-4346-9a32-858a77615eb6\" (UID: \"712a2a31-4205-4346-9a32-858a77615eb6\") " Mar 07 14:45:13 crc kubenswrapper[4943]: I0307 14:45:13.233358 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/712a2a31-4205-4346-9a32-858a77615eb6-marketplace-operator-metrics\") pod \"712a2a31-4205-4346-9a32-858a77615eb6\" (UID: \"712a2a31-4205-4346-9a32-858a77615eb6\") " Mar 07 14:45:13 crc kubenswrapper[4943]: I0307 14:45:13.234001 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/712a2a31-4205-4346-9a32-858a77615eb6-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "712a2a31-4205-4346-9a32-858a77615eb6" (UID: "712a2a31-4205-4346-9a32-858a77615eb6"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:45:13 crc kubenswrapper[4943]: I0307 14:45:13.239972 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/712a2a31-4205-4346-9a32-858a77615eb6-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "712a2a31-4205-4346-9a32-858a77615eb6" (UID: "712a2a31-4205-4346-9a32-858a77615eb6"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:45:13 crc kubenswrapper[4943]: I0307 14:45:13.240682 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/712a2a31-4205-4346-9a32-858a77615eb6-kube-api-access-92fpw" (OuterVolumeSpecName: "kube-api-access-92fpw") pod "712a2a31-4205-4346-9a32-858a77615eb6" (UID: "712a2a31-4205-4346-9a32-858a77615eb6"). InnerVolumeSpecName "kube-api-access-92fpw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:45:13 crc kubenswrapper[4943]: I0307 14:45:13.334999 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-92fpw\" (UniqueName: \"kubernetes.io/projected/712a2a31-4205-4346-9a32-858a77615eb6-kube-api-access-92fpw\") on node \"crc\" DevicePath \"\"" Mar 07 14:45:13 crc kubenswrapper[4943]: I0307 14:45:13.335065 4943 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/712a2a31-4205-4346-9a32-858a77615eb6-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Mar 07 14:45:13 crc kubenswrapper[4943]: I0307 14:45:13.335086 4943 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/712a2a31-4205-4346-9a32-858a77615eb6-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 07 14:45:13 crc kubenswrapper[4943]: I0307 14:45:13.771067 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f95bw" event={"ID":"a41f89c1-5f6d-46bc-801b-a29ebebf4468","Type":"ContainerDied","Data":"7652bbe2473079e8df9950f4a114b06f40bffb90917d1251c67fd891bb0374a0"} Mar 07 14:45:13 crc kubenswrapper[4943]: I0307 14:45:13.771122 4943 scope.go:117] "RemoveContainer" containerID="1b1573f313b949b3ea11b97a155d031e94aeebb27dc909c06ba03f22ba4e6cfa" Mar 07 14:45:13 crc kubenswrapper[4943]: I0307 14:45:13.771156 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f95bw" Mar 07 14:45:13 crc kubenswrapper[4943]: I0307 14:45:13.775287 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cpfgk" event={"ID":"9ba51437-0e4b-4aca-87c7-1152f9c2a461","Type":"ContainerStarted","Data":"2ec7ab4b6cc1b6678dc848111a6d0f143ad7cbf1392a031815fbdb398f2381ac"} Mar 07 14:45:13 crc kubenswrapper[4943]: I0307 14:45:13.775324 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cpfgk" event={"ID":"9ba51437-0e4b-4aca-87c7-1152f9c2a461","Type":"ContainerStarted","Data":"9b382521ffa7fc6348f78203838f142e9bcfd68ca7f2c23aa32ddd113ed3b70e"} Mar 07 14:45:13 crc kubenswrapper[4943]: I0307 14:45:13.775639 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-cpfgk" Mar 07 14:45:13 crc kubenswrapper[4943]: I0307 14:45:13.777648 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-sqsdb" event={"ID":"712a2a31-4205-4346-9a32-858a77615eb6","Type":"ContainerDied","Data":"07df969ccc164094099b25423ccbaaff05f842977aea3d1c5560e14d8169fd99"} Mar 07 14:45:13 crc kubenswrapper[4943]: I0307 14:45:13.777720 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-sqsdb" Mar 07 14:45:13 crc kubenswrapper[4943]: I0307 14:45:13.780821 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jrc6h" event={"ID":"f10fa8f5-f504-40c9-81a5-d1658c1ef268","Type":"ContainerDied","Data":"3c06a48028f715519a06fd42092c90004ec700a2701f57713cf67b76f542b7a9"} Mar 07 14:45:13 crc kubenswrapper[4943]: I0307 14:45:13.781019 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jrc6h" Mar 07 14:45:13 crc kubenswrapper[4943]: I0307 14:45:13.782322 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-cpfgk" Mar 07 14:45:13 crc kubenswrapper[4943]: I0307 14:45:13.783678 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dfqzr" Mar 07 14:45:13 crc kubenswrapper[4943]: I0307 14:45:13.806889 4943 scope.go:117] "RemoveContainer" containerID="8be7d1144cb9e9546f622f3ad7227b4339e050c7ecf5514fcf23425048fc8ff5" Mar 07 14:45:13 crc kubenswrapper[4943]: I0307 14:45:13.825356 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-cpfgk" podStartSLOduration=1.8253131329999999 podStartE2EDuration="1.825313133s" podCreationTimestamp="2026-03-07 14:45:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:45:13.804682139 +0000 UTC m=+355.756818637" watchObservedRunningTime="2026-03-07 14:45:13.825313133 +0000 UTC m=+355.777449651" Mar 07 14:45:13 crc kubenswrapper[4943]: I0307 14:45:13.837218 4943 scope.go:117] "RemoveContainer" containerID="7e1875f1478912123de7be2c99d34eca0ef570d508fb004db0cd5ab210429a91" Mar 07 14:45:13 crc kubenswrapper[4943]: I0307 14:45:13.890948 4943 scope.go:117] "RemoveContainer" containerID="08e935ddfdfcb02fe9bdeb0fb05b2807df84fbf1297330bd13c1372e83260314" Mar 07 14:45:13 crc kubenswrapper[4943]: I0307 14:45:13.910546 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jrc6h"] Mar 07 14:45:13 crc kubenswrapper[4943]: I0307 14:45:13.921972 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jrc6h"] Mar 07 14:45:13 crc kubenswrapper[4943]: I0307 14:45:13.926420 4943 scope.go:117] "RemoveContainer" containerID="2589f77538011f2ab019b72637e957313f033e18e37f87b0825a4312ec83bbd5" Mar 07 14:45:13 crc kubenswrapper[4943]: I0307 14:45:13.931683 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sqsdb"] Mar 07 14:45:13 crc kubenswrapper[4943]: I0307 14:45:13.939112 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sqsdb"] Mar 07 14:45:13 crc kubenswrapper[4943]: I0307 14:45:13.947813 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-f95bw"] Mar 07 14:45:13 crc kubenswrapper[4943]: I0307 14:45:13.951355 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-f95bw"] Mar 07 14:45:13 crc kubenswrapper[4943]: I0307 14:45:13.956203 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dfqzr"] Mar 07 14:45:13 crc kubenswrapper[4943]: I0307 14:45:13.961136 4943 scope.go:117] "RemoveContainer" containerID="20bdfe5cb0c8c7149d03c99da6b8dec4af6ec25d0b1fa3174e4cced7f6a7790b" Mar 07 14:45:13 crc kubenswrapper[4943]: I0307 14:45:13.962586 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dfqzr"] Mar 07 14:45:13 crc kubenswrapper[4943]: I0307 14:45:13.975461 4943 scope.go:117] "RemoveContainer" containerID="eb4adba6696048589e1e42c2302ca9cfc67db345f210720f61ecec85240427e2" Mar 07 14:45:14 crc kubenswrapper[4943]: I0307 14:45:14.459870 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4pbcq"] Mar 07 14:45:14 crc kubenswrapper[4943]: E0307 14:45:14.460541 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="712a2a31-4205-4346-9a32-858a77615eb6" containerName="marketplace-operator" Mar 07 14:45:14 crc kubenswrapper[4943]: I0307 14:45:14.460560 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="712a2a31-4205-4346-9a32-858a77615eb6" containerName="marketplace-operator" Mar 07 14:45:14 crc kubenswrapper[4943]: E0307 14:45:14.460575 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="254f0c0f-ee54-434e-bf7b-bf2e5274fdc9" containerName="extract-utilities" Mar 07 14:45:14 crc kubenswrapper[4943]: I0307 14:45:14.460586 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="254f0c0f-ee54-434e-bf7b-bf2e5274fdc9" containerName="extract-utilities" Mar 07 14:45:14 crc kubenswrapper[4943]: E0307 14:45:14.460597 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d1657d7-a787-4648-91f6-e0ff987cede2" containerName="extract-content" Mar 07 14:45:14 crc kubenswrapper[4943]: I0307 14:45:14.460605 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d1657d7-a787-4648-91f6-e0ff987cede2" containerName="extract-content" Mar 07 14:45:14 crc kubenswrapper[4943]: E0307 14:45:14.460619 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="712a2a31-4205-4346-9a32-858a77615eb6" containerName="marketplace-operator" Mar 07 14:45:14 crc kubenswrapper[4943]: I0307 14:45:14.460629 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="712a2a31-4205-4346-9a32-858a77615eb6" containerName="marketplace-operator" Mar 07 14:45:14 crc kubenswrapper[4943]: E0307 14:45:14.460638 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a41f89c1-5f6d-46bc-801b-a29ebebf4468" containerName="extract-content" Mar 07 14:45:14 crc kubenswrapper[4943]: I0307 14:45:14.460647 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="a41f89c1-5f6d-46bc-801b-a29ebebf4468" containerName="extract-content" Mar 07 14:45:14 crc kubenswrapper[4943]: E0307 14:45:14.460663 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a41f89c1-5f6d-46bc-801b-a29ebebf4468" containerName="extract-utilities" Mar 07 14:45:14 crc kubenswrapper[4943]: I0307 14:45:14.460671 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="a41f89c1-5f6d-46bc-801b-a29ebebf4468" containerName="extract-utilities" Mar 07 14:45:14 crc kubenswrapper[4943]: E0307 14:45:14.460683 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d1657d7-a787-4648-91f6-e0ff987cede2" containerName="registry-server" Mar 07 14:45:14 crc kubenswrapper[4943]: I0307 14:45:14.460691 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d1657d7-a787-4648-91f6-e0ff987cede2" containerName="registry-server" Mar 07 14:45:14 crc kubenswrapper[4943]: E0307 14:45:14.460703 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f10fa8f5-f504-40c9-81a5-d1658c1ef268" containerName="extract-content" Mar 07 14:45:14 crc kubenswrapper[4943]: I0307 14:45:14.460713 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="f10fa8f5-f504-40c9-81a5-d1658c1ef268" containerName="extract-content" Mar 07 14:45:14 crc kubenswrapper[4943]: E0307 14:45:14.460725 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f10fa8f5-f504-40c9-81a5-d1658c1ef268" containerName="registry-server" Mar 07 14:45:14 crc kubenswrapper[4943]: I0307 14:45:14.460733 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="f10fa8f5-f504-40c9-81a5-d1658c1ef268" containerName="registry-server" Mar 07 14:45:14 crc kubenswrapper[4943]: E0307 14:45:14.460743 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="254f0c0f-ee54-434e-bf7b-bf2e5274fdc9" containerName="extract-content" Mar 07 14:45:14 crc kubenswrapper[4943]: I0307 14:45:14.460750 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="254f0c0f-ee54-434e-bf7b-bf2e5274fdc9" containerName="extract-content" Mar 07 14:45:14 crc kubenswrapper[4943]: E0307 14:45:14.460762 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d1657d7-a787-4648-91f6-e0ff987cede2" containerName="extract-utilities" Mar 07 14:45:14 crc kubenswrapper[4943]: I0307 14:45:14.460771 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d1657d7-a787-4648-91f6-e0ff987cede2" containerName="extract-utilities" Mar 07 14:45:14 crc kubenswrapper[4943]: E0307 14:45:14.460783 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a41f89c1-5f6d-46bc-801b-a29ebebf4468" containerName="registry-server" Mar 07 14:45:14 crc kubenswrapper[4943]: I0307 14:45:14.460791 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="a41f89c1-5f6d-46bc-801b-a29ebebf4468" containerName="registry-server" Mar 07 14:45:14 crc kubenswrapper[4943]: E0307 14:45:14.460807 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f10fa8f5-f504-40c9-81a5-d1658c1ef268" containerName="extract-utilities" Mar 07 14:45:14 crc kubenswrapper[4943]: I0307 14:45:14.460816 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="f10fa8f5-f504-40c9-81a5-d1658c1ef268" containerName="extract-utilities" Mar 07 14:45:14 crc kubenswrapper[4943]: E0307 14:45:14.460825 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="254f0c0f-ee54-434e-bf7b-bf2e5274fdc9" containerName="registry-server" Mar 07 14:45:14 crc kubenswrapper[4943]: I0307 14:45:14.460834 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="254f0c0f-ee54-434e-bf7b-bf2e5274fdc9" containerName="registry-server" Mar 07 14:45:14 crc kubenswrapper[4943]: I0307 14:45:14.460987 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="a41f89c1-5f6d-46bc-801b-a29ebebf4468" containerName="registry-server" Mar 07 14:45:14 crc kubenswrapper[4943]: I0307 14:45:14.460998 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="712a2a31-4205-4346-9a32-858a77615eb6" containerName="marketplace-operator" Mar 07 14:45:14 crc kubenswrapper[4943]: I0307 14:45:14.461010 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="f10fa8f5-f504-40c9-81a5-d1658c1ef268" containerName="registry-server" Mar 07 14:45:14 crc kubenswrapper[4943]: I0307 14:45:14.461025 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="254f0c0f-ee54-434e-bf7b-bf2e5274fdc9" containerName="registry-server" Mar 07 14:45:14 crc kubenswrapper[4943]: I0307 14:45:14.461037 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="712a2a31-4205-4346-9a32-858a77615eb6" containerName="marketplace-operator" Mar 07 14:45:14 crc kubenswrapper[4943]: I0307 14:45:14.461049 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d1657d7-a787-4648-91f6-e0ff987cede2" containerName="registry-server" Mar 07 14:45:14 crc kubenswrapper[4943]: I0307 14:45:14.462124 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4pbcq" Mar 07 14:45:14 crc kubenswrapper[4943]: I0307 14:45:14.466502 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 07 14:45:14 crc kubenswrapper[4943]: I0307 14:45:14.485093 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4pbcq"] Mar 07 14:45:14 crc kubenswrapper[4943]: I0307 14:45:14.556337 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wl48c\" (UniqueName: \"kubernetes.io/projected/0e0c280e-3497-4f16-bd4d-b26b9f196700-kube-api-access-wl48c\") pod \"certified-operators-4pbcq\" (UID: \"0e0c280e-3497-4f16-bd4d-b26b9f196700\") " pod="openshift-marketplace/certified-operators-4pbcq" Mar 07 14:45:14 crc kubenswrapper[4943]: I0307 14:45:14.556423 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e0c280e-3497-4f16-bd4d-b26b9f196700-utilities\") pod \"certified-operators-4pbcq\" (UID: \"0e0c280e-3497-4f16-bd4d-b26b9f196700\") " pod="openshift-marketplace/certified-operators-4pbcq" Mar 07 14:45:14 crc kubenswrapper[4943]: I0307 14:45:14.556473 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e0c280e-3497-4f16-bd4d-b26b9f196700-catalog-content\") pod \"certified-operators-4pbcq\" (UID: \"0e0c280e-3497-4f16-bd4d-b26b9f196700\") " pod="openshift-marketplace/certified-operators-4pbcq" Mar 07 14:45:14 crc kubenswrapper[4943]: I0307 14:45:14.651830 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rhp5s"] Mar 07 14:45:14 crc kubenswrapper[4943]: I0307 14:45:14.652779 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rhp5s" Mar 07 14:45:14 crc kubenswrapper[4943]: I0307 14:45:14.655902 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 07 14:45:14 crc kubenswrapper[4943]: I0307 14:45:14.658086 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e0c280e-3497-4f16-bd4d-b26b9f196700-utilities\") pod \"certified-operators-4pbcq\" (UID: \"0e0c280e-3497-4f16-bd4d-b26b9f196700\") " pod="openshift-marketplace/certified-operators-4pbcq" Mar 07 14:45:14 crc kubenswrapper[4943]: I0307 14:45:14.658131 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e0c280e-3497-4f16-bd4d-b26b9f196700-catalog-content\") pod \"certified-operators-4pbcq\" (UID: \"0e0c280e-3497-4f16-bd4d-b26b9f196700\") " pod="openshift-marketplace/certified-operators-4pbcq" Mar 07 14:45:14 crc kubenswrapper[4943]: I0307 14:45:14.658181 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wl48c\" (UniqueName: \"kubernetes.io/projected/0e0c280e-3497-4f16-bd4d-b26b9f196700-kube-api-access-wl48c\") pod \"certified-operators-4pbcq\" (UID: \"0e0c280e-3497-4f16-bd4d-b26b9f196700\") " pod="openshift-marketplace/certified-operators-4pbcq" Mar 07 14:45:14 crc kubenswrapper[4943]: I0307 14:45:14.658876 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e0c280e-3497-4f16-bd4d-b26b9f196700-utilities\") pod \"certified-operators-4pbcq\" (UID: \"0e0c280e-3497-4f16-bd4d-b26b9f196700\") " pod="openshift-marketplace/certified-operators-4pbcq" Mar 07 14:45:14 crc kubenswrapper[4943]: I0307 14:45:14.659092 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e0c280e-3497-4f16-bd4d-b26b9f196700-catalog-content\") pod \"certified-operators-4pbcq\" (UID: \"0e0c280e-3497-4f16-bd4d-b26b9f196700\") " pod="openshift-marketplace/certified-operators-4pbcq" Mar 07 14:45:14 crc kubenswrapper[4943]: I0307 14:45:14.665244 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rhp5s"] Mar 07 14:45:14 crc kubenswrapper[4943]: I0307 14:45:14.709994 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wl48c\" (UniqueName: \"kubernetes.io/projected/0e0c280e-3497-4f16-bd4d-b26b9f196700-kube-api-access-wl48c\") pod \"certified-operators-4pbcq\" (UID: \"0e0c280e-3497-4f16-bd4d-b26b9f196700\") " pod="openshift-marketplace/certified-operators-4pbcq" Mar 07 14:45:14 crc kubenswrapper[4943]: I0307 14:45:14.758870 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86m2z\" (UniqueName: \"kubernetes.io/projected/d52577a0-c952-477f-bcd3-9eaa06cb2575-kube-api-access-86m2z\") pod \"community-operators-rhp5s\" (UID: \"d52577a0-c952-477f-bcd3-9eaa06cb2575\") " pod="openshift-marketplace/community-operators-rhp5s" Mar 07 14:45:14 crc kubenswrapper[4943]: I0307 14:45:14.759112 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d52577a0-c952-477f-bcd3-9eaa06cb2575-utilities\") pod \"community-operators-rhp5s\" (UID: \"d52577a0-c952-477f-bcd3-9eaa06cb2575\") " pod="openshift-marketplace/community-operators-rhp5s" Mar 07 14:45:14 crc kubenswrapper[4943]: I0307 14:45:14.759274 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d52577a0-c952-477f-bcd3-9eaa06cb2575-catalog-content\") pod \"community-operators-rhp5s\" (UID: \"d52577a0-c952-477f-bcd3-9eaa06cb2575\") " pod="openshift-marketplace/community-operators-rhp5s" Mar 07 14:45:14 crc kubenswrapper[4943]: I0307 14:45:14.763845 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d1657d7-a787-4648-91f6-e0ff987cede2" path="/var/lib/kubelet/pods/0d1657d7-a787-4648-91f6-e0ff987cede2/volumes" Mar 07 14:45:14 crc kubenswrapper[4943]: I0307 14:45:14.765588 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="254f0c0f-ee54-434e-bf7b-bf2e5274fdc9" path="/var/lib/kubelet/pods/254f0c0f-ee54-434e-bf7b-bf2e5274fdc9/volumes" Mar 07 14:45:14 crc kubenswrapper[4943]: I0307 14:45:14.766971 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="712a2a31-4205-4346-9a32-858a77615eb6" path="/var/lib/kubelet/pods/712a2a31-4205-4346-9a32-858a77615eb6/volumes" Mar 07 14:45:14 crc kubenswrapper[4943]: I0307 14:45:14.768822 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a41f89c1-5f6d-46bc-801b-a29ebebf4468" path="/var/lib/kubelet/pods/a41f89c1-5f6d-46bc-801b-a29ebebf4468/volumes" Mar 07 14:45:14 crc kubenswrapper[4943]: I0307 14:45:14.770044 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f10fa8f5-f504-40c9-81a5-d1658c1ef268" path="/var/lib/kubelet/pods/f10fa8f5-f504-40c9-81a5-d1658c1ef268/volumes" Mar 07 14:45:14 crc kubenswrapper[4943]: I0307 14:45:14.801659 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4pbcq" Mar 07 14:45:14 crc kubenswrapper[4943]: I0307 14:45:14.860465 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d52577a0-c952-477f-bcd3-9eaa06cb2575-utilities\") pod \"community-operators-rhp5s\" (UID: \"d52577a0-c952-477f-bcd3-9eaa06cb2575\") " pod="openshift-marketplace/community-operators-rhp5s" Mar 07 14:45:14 crc kubenswrapper[4943]: I0307 14:45:14.860530 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d52577a0-c952-477f-bcd3-9eaa06cb2575-catalog-content\") pod \"community-operators-rhp5s\" (UID: \"d52577a0-c952-477f-bcd3-9eaa06cb2575\") " pod="openshift-marketplace/community-operators-rhp5s" Mar 07 14:45:14 crc kubenswrapper[4943]: I0307 14:45:14.860567 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86m2z\" (UniqueName: \"kubernetes.io/projected/d52577a0-c952-477f-bcd3-9eaa06cb2575-kube-api-access-86m2z\") pod \"community-operators-rhp5s\" (UID: \"d52577a0-c952-477f-bcd3-9eaa06cb2575\") " pod="openshift-marketplace/community-operators-rhp5s" Mar 07 14:45:14 crc kubenswrapper[4943]: I0307 14:45:14.861514 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d52577a0-c952-477f-bcd3-9eaa06cb2575-utilities\") pod \"community-operators-rhp5s\" (UID: \"d52577a0-c952-477f-bcd3-9eaa06cb2575\") " pod="openshift-marketplace/community-operators-rhp5s" Mar 07 14:45:14 crc kubenswrapper[4943]: I0307 14:45:14.862243 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d52577a0-c952-477f-bcd3-9eaa06cb2575-catalog-content\") pod \"community-operators-rhp5s\" (UID: \"d52577a0-c952-477f-bcd3-9eaa06cb2575\") " pod="openshift-marketplace/community-operators-rhp5s" Mar 07 14:45:14 crc kubenswrapper[4943]: I0307 14:45:14.877504 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86m2z\" (UniqueName: \"kubernetes.io/projected/d52577a0-c952-477f-bcd3-9eaa06cb2575-kube-api-access-86m2z\") pod \"community-operators-rhp5s\" (UID: \"d52577a0-c952-477f-bcd3-9eaa06cb2575\") " pod="openshift-marketplace/community-operators-rhp5s" Mar 07 14:45:15 crc kubenswrapper[4943]: I0307 14:45:15.028803 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rhp5s" Mar 07 14:45:15 crc kubenswrapper[4943]: I0307 14:45:15.242890 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4pbcq"] Mar 07 14:45:15 crc kubenswrapper[4943]: W0307 14:45:15.248305 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0e0c280e_3497_4f16_bd4d_b26b9f196700.slice/crio-d283f97c70dfcf730b627c9dedb89a7ea81f413803d0ab9914e1ff4e22aecb18 WatchSource:0}: Error finding container d283f97c70dfcf730b627c9dedb89a7ea81f413803d0ab9914e1ff4e22aecb18: Status 404 returned error can't find the container with id d283f97c70dfcf730b627c9dedb89a7ea81f413803d0ab9914e1ff4e22aecb18 Mar 07 14:45:15 crc kubenswrapper[4943]: I0307 14:45:15.413670 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rhp5s"] Mar 07 14:45:15 crc kubenswrapper[4943]: W0307 14:45:15.452672 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd52577a0_c952_477f_bcd3_9eaa06cb2575.slice/crio-90e3791a1b2cdbce674cd74dcced0a072198ae3b328f7722bb9a4d40a816cf27 WatchSource:0}: Error finding container 90e3791a1b2cdbce674cd74dcced0a072198ae3b328f7722bb9a4d40a816cf27: Status 404 returned error can't find the container with id 90e3791a1b2cdbce674cd74dcced0a072198ae3b328f7722bb9a4d40a816cf27 Mar 07 14:45:15 crc kubenswrapper[4943]: I0307 14:45:15.803029 4943 generic.go:334] "Generic (PLEG): container finished" podID="0e0c280e-3497-4f16-bd4d-b26b9f196700" containerID="e67d237afd8b7f96cfa54daebf4eaeb8e663c844c2ae297b207ec30a5d44a260" exitCode=0 Mar 07 14:45:15 crc kubenswrapper[4943]: I0307 14:45:15.803211 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4pbcq" event={"ID":"0e0c280e-3497-4f16-bd4d-b26b9f196700","Type":"ContainerDied","Data":"e67d237afd8b7f96cfa54daebf4eaeb8e663c844c2ae297b207ec30a5d44a260"} Mar 07 14:45:15 crc kubenswrapper[4943]: I0307 14:45:15.803649 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4pbcq" event={"ID":"0e0c280e-3497-4f16-bd4d-b26b9f196700","Type":"ContainerStarted","Data":"d283f97c70dfcf730b627c9dedb89a7ea81f413803d0ab9914e1ff4e22aecb18"} Mar 07 14:45:15 crc kubenswrapper[4943]: I0307 14:45:15.806525 4943 generic.go:334] "Generic (PLEG): container finished" podID="d52577a0-c952-477f-bcd3-9eaa06cb2575" containerID="dcc0ce27da8f1b206cff314c0d75ac4f563aedeec91a0ad43d2d3ec7eaafa24f" exitCode=0 Mar 07 14:45:15 crc kubenswrapper[4943]: I0307 14:45:15.806610 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rhp5s" event={"ID":"d52577a0-c952-477f-bcd3-9eaa06cb2575","Type":"ContainerDied","Data":"dcc0ce27da8f1b206cff314c0d75ac4f563aedeec91a0ad43d2d3ec7eaafa24f"} Mar 07 14:45:15 crc kubenswrapper[4943]: I0307 14:45:15.806683 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rhp5s" event={"ID":"d52577a0-c952-477f-bcd3-9eaa06cb2575","Type":"ContainerStarted","Data":"90e3791a1b2cdbce674cd74dcced0a072198ae3b328f7722bb9a4d40a816cf27"} Mar 07 14:45:16 crc kubenswrapper[4943]: I0307 14:45:16.817043 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rhp5s" event={"ID":"d52577a0-c952-477f-bcd3-9eaa06cb2575","Type":"ContainerStarted","Data":"4b9ae14f6535dce626438f1072298976613ae185cf2d2e1922f942de51daea80"} Mar 07 14:45:16 crc kubenswrapper[4943]: I0307 14:45:16.820699 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4pbcq" event={"ID":"0e0c280e-3497-4f16-bd4d-b26b9f196700","Type":"ContainerStarted","Data":"470bd737da09cd09efa43bd93f98e929d14d3ccd83acc1813a066620e95f9ca9"} Mar 07 14:45:16 crc kubenswrapper[4943]: I0307 14:45:16.865812 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-b2fvs"] Mar 07 14:45:16 crc kubenswrapper[4943]: I0307 14:45:16.867255 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b2fvs" Mar 07 14:45:16 crc kubenswrapper[4943]: I0307 14:45:16.871980 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 07 14:45:16 crc kubenswrapper[4943]: I0307 14:45:16.876196 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b2fvs"] Mar 07 14:45:16 crc kubenswrapper[4943]: I0307 14:45:16.989823 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xk999\" (UniqueName: \"kubernetes.io/projected/94d5996e-5894-4f62-961f-ce9f41375047-kube-api-access-xk999\") pod \"redhat-marketplace-b2fvs\" (UID: \"94d5996e-5894-4f62-961f-ce9f41375047\") " pod="openshift-marketplace/redhat-marketplace-b2fvs" Mar 07 14:45:16 crc kubenswrapper[4943]: I0307 14:45:16.989891 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94d5996e-5894-4f62-961f-ce9f41375047-utilities\") pod \"redhat-marketplace-b2fvs\" (UID: \"94d5996e-5894-4f62-961f-ce9f41375047\") " pod="openshift-marketplace/redhat-marketplace-b2fvs" Mar 07 14:45:16 crc kubenswrapper[4943]: I0307 14:45:16.990273 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94d5996e-5894-4f62-961f-ce9f41375047-catalog-content\") pod \"redhat-marketplace-b2fvs\" (UID: \"94d5996e-5894-4f62-961f-ce9f41375047\") " pod="openshift-marketplace/redhat-marketplace-b2fvs" Mar 07 14:45:17 crc kubenswrapper[4943]: I0307 14:45:17.051736 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-k6478"] Mar 07 14:45:17 crc kubenswrapper[4943]: I0307 14:45:17.052670 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k6478" Mar 07 14:45:17 crc kubenswrapper[4943]: I0307 14:45:17.055414 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 07 14:45:17 crc kubenswrapper[4943]: I0307 14:45:17.071212 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-k6478"] Mar 07 14:45:17 crc kubenswrapper[4943]: I0307 14:45:17.091882 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xk999\" (UniqueName: \"kubernetes.io/projected/94d5996e-5894-4f62-961f-ce9f41375047-kube-api-access-xk999\") pod \"redhat-marketplace-b2fvs\" (UID: \"94d5996e-5894-4f62-961f-ce9f41375047\") " pod="openshift-marketplace/redhat-marketplace-b2fvs" Mar 07 14:45:17 crc kubenswrapper[4943]: I0307 14:45:17.091973 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94d5996e-5894-4f62-961f-ce9f41375047-utilities\") pod \"redhat-marketplace-b2fvs\" (UID: \"94d5996e-5894-4f62-961f-ce9f41375047\") " pod="openshift-marketplace/redhat-marketplace-b2fvs" Mar 07 14:45:17 crc kubenswrapper[4943]: I0307 14:45:17.092089 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94d5996e-5894-4f62-961f-ce9f41375047-catalog-content\") pod \"redhat-marketplace-b2fvs\" (UID: \"94d5996e-5894-4f62-961f-ce9f41375047\") " pod="openshift-marketplace/redhat-marketplace-b2fvs" Mar 07 14:45:17 crc kubenswrapper[4943]: I0307 14:45:17.092815 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94d5996e-5894-4f62-961f-ce9f41375047-catalog-content\") pod \"redhat-marketplace-b2fvs\" (UID: \"94d5996e-5894-4f62-961f-ce9f41375047\") " pod="openshift-marketplace/redhat-marketplace-b2fvs" Mar 07 14:45:17 crc kubenswrapper[4943]: I0307 14:45:17.092956 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94d5996e-5894-4f62-961f-ce9f41375047-utilities\") pod \"redhat-marketplace-b2fvs\" (UID: \"94d5996e-5894-4f62-961f-ce9f41375047\") " pod="openshift-marketplace/redhat-marketplace-b2fvs" Mar 07 14:45:17 crc kubenswrapper[4943]: I0307 14:45:17.110017 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xk999\" (UniqueName: \"kubernetes.io/projected/94d5996e-5894-4f62-961f-ce9f41375047-kube-api-access-xk999\") pod \"redhat-marketplace-b2fvs\" (UID: \"94d5996e-5894-4f62-961f-ce9f41375047\") " pod="openshift-marketplace/redhat-marketplace-b2fvs" Mar 07 14:45:17 crc kubenswrapper[4943]: I0307 14:45:17.193782 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db1d6561-f6c0-44fe-80f5-e5c6f9179c0a-catalog-content\") pod \"redhat-operators-k6478\" (UID: \"db1d6561-f6c0-44fe-80f5-e5c6f9179c0a\") " pod="openshift-marketplace/redhat-operators-k6478" Mar 07 14:45:17 crc kubenswrapper[4943]: I0307 14:45:17.193892 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwrkg\" (UniqueName: \"kubernetes.io/projected/db1d6561-f6c0-44fe-80f5-e5c6f9179c0a-kube-api-access-nwrkg\") pod \"redhat-operators-k6478\" (UID: \"db1d6561-f6c0-44fe-80f5-e5c6f9179c0a\") " pod="openshift-marketplace/redhat-operators-k6478" Mar 07 14:45:17 crc kubenswrapper[4943]: I0307 14:45:17.193949 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db1d6561-f6c0-44fe-80f5-e5c6f9179c0a-utilities\") pod \"redhat-operators-k6478\" (UID: \"db1d6561-f6c0-44fe-80f5-e5c6f9179c0a\") " pod="openshift-marketplace/redhat-operators-k6478" Mar 07 14:45:17 crc kubenswrapper[4943]: I0307 14:45:17.294615 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db1d6561-f6c0-44fe-80f5-e5c6f9179c0a-catalog-content\") pod \"redhat-operators-k6478\" (UID: \"db1d6561-f6c0-44fe-80f5-e5c6f9179c0a\") " pod="openshift-marketplace/redhat-operators-k6478" Mar 07 14:45:17 crc kubenswrapper[4943]: I0307 14:45:17.295054 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwrkg\" (UniqueName: \"kubernetes.io/projected/db1d6561-f6c0-44fe-80f5-e5c6f9179c0a-kube-api-access-nwrkg\") pod \"redhat-operators-k6478\" (UID: \"db1d6561-f6c0-44fe-80f5-e5c6f9179c0a\") " pod="openshift-marketplace/redhat-operators-k6478" Mar 07 14:45:17 crc kubenswrapper[4943]: I0307 14:45:17.295143 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db1d6561-f6c0-44fe-80f5-e5c6f9179c0a-utilities\") pod \"redhat-operators-k6478\" (UID: \"db1d6561-f6c0-44fe-80f5-e5c6f9179c0a\") " pod="openshift-marketplace/redhat-operators-k6478" Mar 07 14:45:17 crc kubenswrapper[4943]: I0307 14:45:17.295196 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db1d6561-f6c0-44fe-80f5-e5c6f9179c0a-catalog-content\") pod \"redhat-operators-k6478\" (UID: \"db1d6561-f6c0-44fe-80f5-e5c6f9179c0a\") " pod="openshift-marketplace/redhat-operators-k6478" Mar 07 14:45:17 crc kubenswrapper[4943]: I0307 14:45:17.295494 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db1d6561-f6c0-44fe-80f5-e5c6f9179c0a-utilities\") pod \"redhat-operators-k6478\" (UID: \"db1d6561-f6c0-44fe-80f5-e5c6f9179c0a\") " pod="openshift-marketplace/redhat-operators-k6478" Mar 07 14:45:17 crc kubenswrapper[4943]: I0307 14:45:17.310489 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwrkg\" (UniqueName: \"kubernetes.io/projected/db1d6561-f6c0-44fe-80f5-e5c6f9179c0a-kube-api-access-nwrkg\") pod \"redhat-operators-k6478\" (UID: \"db1d6561-f6c0-44fe-80f5-e5c6f9179c0a\") " pod="openshift-marketplace/redhat-operators-k6478" Mar 07 14:45:17 crc kubenswrapper[4943]: I0307 14:45:17.345781 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b2fvs" Mar 07 14:45:17 crc kubenswrapper[4943]: I0307 14:45:17.369246 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k6478" Mar 07 14:45:17 crc kubenswrapper[4943]: I0307 14:45:17.639585 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-k6478"] Mar 07 14:45:17 crc kubenswrapper[4943]: I0307 14:45:17.798009 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b2fvs"] Mar 07 14:45:17 crc kubenswrapper[4943]: I0307 14:45:17.829871 4943 generic.go:334] "Generic (PLEG): container finished" podID="0e0c280e-3497-4f16-bd4d-b26b9f196700" containerID="470bd737da09cd09efa43bd93f98e929d14d3ccd83acc1813a066620e95f9ca9" exitCode=0 Mar 07 14:45:17 crc kubenswrapper[4943]: I0307 14:45:17.830043 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4pbcq" event={"ID":"0e0c280e-3497-4f16-bd4d-b26b9f196700","Type":"ContainerDied","Data":"470bd737da09cd09efa43bd93f98e929d14d3ccd83acc1813a066620e95f9ca9"} Mar 07 14:45:17 crc kubenswrapper[4943]: I0307 14:45:17.831844 4943 generic.go:334] "Generic (PLEG): container finished" podID="db1d6561-f6c0-44fe-80f5-e5c6f9179c0a" containerID="65077efd1a898bb694bd6eb2a2f2123f48d82b418c170b5e36116c8097debad1" exitCode=0 Mar 07 14:45:17 crc kubenswrapper[4943]: I0307 14:45:17.831912 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k6478" event={"ID":"db1d6561-f6c0-44fe-80f5-e5c6f9179c0a","Type":"ContainerDied","Data":"65077efd1a898bb694bd6eb2a2f2123f48d82b418c170b5e36116c8097debad1"} Mar 07 14:45:17 crc kubenswrapper[4943]: I0307 14:45:17.831956 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k6478" event={"ID":"db1d6561-f6c0-44fe-80f5-e5c6f9179c0a","Type":"ContainerStarted","Data":"9850b4d11197f0261a87a22f8bab6b96dccd1a146c1a2614688f06f71b877e7d"} Mar 07 14:45:17 crc kubenswrapper[4943]: I0307 14:45:17.838763 4943 generic.go:334] "Generic (PLEG): container finished" podID="d52577a0-c952-477f-bcd3-9eaa06cb2575" containerID="4b9ae14f6535dce626438f1072298976613ae185cf2d2e1922f942de51daea80" exitCode=0 Mar 07 14:45:17 crc kubenswrapper[4943]: I0307 14:45:17.838811 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rhp5s" event={"ID":"d52577a0-c952-477f-bcd3-9eaa06cb2575","Type":"ContainerDied","Data":"4b9ae14f6535dce626438f1072298976613ae185cf2d2e1922f942de51daea80"} Mar 07 14:45:17 crc kubenswrapper[4943]: W0307 14:45:17.857065 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod94d5996e_5894_4f62_961f_ce9f41375047.slice/crio-d9ecfec7c83c4ee8e9414a632c8a15f8e4488a0c04a85b32ab2408e97397eea4 WatchSource:0}: Error finding container d9ecfec7c83c4ee8e9414a632c8a15f8e4488a0c04a85b32ab2408e97397eea4: Status 404 returned error can't find the container with id d9ecfec7c83c4ee8e9414a632c8a15f8e4488a0c04a85b32ab2408e97397eea4 Mar 07 14:45:18 crc kubenswrapper[4943]: I0307 14:45:18.848315 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rhp5s" event={"ID":"d52577a0-c952-477f-bcd3-9eaa06cb2575","Type":"ContainerStarted","Data":"a68f05ff15687491a5a2b2398f3a1d3c474ae02f3f15995eb0c105aa898d7678"} Mar 07 14:45:18 crc kubenswrapper[4943]: I0307 14:45:18.851303 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4pbcq" event={"ID":"0e0c280e-3497-4f16-bd4d-b26b9f196700","Type":"ContainerStarted","Data":"a7d0b7d7ff9f360243d524e596601ecc736be659540534aa7ad500f08ce6e520"} Mar 07 14:45:18 crc kubenswrapper[4943]: I0307 14:45:18.854052 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k6478" event={"ID":"db1d6561-f6c0-44fe-80f5-e5c6f9179c0a","Type":"ContainerStarted","Data":"30b16eb9df40065a6afaad8c74611f3ace518571047fc8bb82480e99d85bd45a"} Mar 07 14:45:18 crc kubenswrapper[4943]: I0307 14:45:18.855950 4943 generic.go:334] "Generic (PLEG): container finished" podID="94d5996e-5894-4f62-961f-ce9f41375047" containerID="d9b95429ce60498b11a7a5cf9edd2f5358f4277abc6fad4b39b76b2c1fe22615" exitCode=0 Mar 07 14:45:18 crc kubenswrapper[4943]: I0307 14:45:18.855991 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b2fvs" event={"ID":"94d5996e-5894-4f62-961f-ce9f41375047","Type":"ContainerDied","Data":"d9b95429ce60498b11a7a5cf9edd2f5358f4277abc6fad4b39b76b2c1fe22615"} Mar 07 14:45:18 crc kubenswrapper[4943]: I0307 14:45:18.856011 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b2fvs" event={"ID":"94d5996e-5894-4f62-961f-ce9f41375047","Type":"ContainerStarted","Data":"d9ecfec7c83c4ee8e9414a632c8a15f8e4488a0c04a85b32ab2408e97397eea4"} Mar 07 14:45:18 crc kubenswrapper[4943]: I0307 14:45:18.872386 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rhp5s" podStartSLOduration=2.464235142 podStartE2EDuration="4.872373331s" podCreationTimestamp="2026-03-07 14:45:14 +0000 UTC" firstStartedPulling="2026-03-07 14:45:15.810279706 +0000 UTC m=+357.762416204" lastFinishedPulling="2026-03-07 14:45:18.218417905 +0000 UTC m=+360.170554393" observedRunningTime="2026-03-07 14:45:18.865803208 +0000 UTC m=+360.817939766" watchObservedRunningTime="2026-03-07 14:45:18.872373331 +0000 UTC m=+360.824509829" Mar 07 14:45:18 crc kubenswrapper[4943]: I0307 14:45:18.908269 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4pbcq" podStartSLOduration=2.456351326 podStartE2EDuration="4.908244855s" podCreationTimestamp="2026-03-07 14:45:14 +0000 UTC" firstStartedPulling="2026-03-07 14:45:15.804690197 +0000 UTC m=+357.756826695" lastFinishedPulling="2026-03-07 14:45:18.256583726 +0000 UTC m=+360.208720224" observedRunningTime="2026-03-07 14:45:18.90522007 +0000 UTC m=+360.857356608" watchObservedRunningTime="2026-03-07 14:45:18.908244855 +0000 UTC m=+360.860381353" Mar 07 14:45:19 crc kubenswrapper[4943]: I0307 14:45:19.868913 4943 generic.go:334] "Generic (PLEG): container finished" podID="db1d6561-f6c0-44fe-80f5-e5c6f9179c0a" containerID="30b16eb9df40065a6afaad8c74611f3ace518571047fc8bb82480e99d85bd45a" exitCode=0 Mar 07 14:45:19 crc kubenswrapper[4943]: I0307 14:45:19.869051 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k6478" event={"ID":"db1d6561-f6c0-44fe-80f5-e5c6f9179c0a","Type":"ContainerDied","Data":"30b16eb9df40065a6afaad8c74611f3ace518571047fc8bb82480e99d85bd45a"} Mar 07 14:45:19 crc kubenswrapper[4943]: I0307 14:45:19.872642 4943 generic.go:334] "Generic (PLEG): container finished" podID="94d5996e-5894-4f62-961f-ce9f41375047" containerID="44500800205b618883f50b8ae98c42b3348d156f655f9974acb70652d0289ca0" exitCode=0 Mar 07 14:45:19 crc kubenswrapper[4943]: I0307 14:45:19.873618 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b2fvs" event={"ID":"94d5996e-5894-4f62-961f-ce9f41375047","Type":"ContainerDied","Data":"44500800205b618883f50b8ae98c42b3348d156f655f9974acb70652d0289ca0"} Mar 07 14:45:20 crc kubenswrapper[4943]: I0307 14:45:20.883375 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k6478" event={"ID":"db1d6561-f6c0-44fe-80f5-e5c6f9179c0a","Type":"ContainerStarted","Data":"68d2da44871d51448a3419e758c96173247f7556d80bd96b1abb0c1b8a62e5be"} Mar 07 14:45:20 crc kubenswrapper[4943]: I0307 14:45:20.888365 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b2fvs" event={"ID":"94d5996e-5894-4f62-961f-ce9f41375047","Type":"ContainerStarted","Data":"01d21dd2843c015c2d0e9c0a9d8328608a7abdd7cbc1d85afacf9ddcb2fa5537"} Mar 07 14:45:20 crc kubenswrapper[4943]: I0307 14:45:20.913710 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-k6478" podStartSLOduration=1.460740744 podStartE2EDuration="3.913683118s" podCreationTimestamp="2026-03-07 14:45:17 +0000 UTC" firstStartedPulling="2026-03-07 14:45:17.833970805 +0000 UTC m=+359.786107303" lastFinishedPulling="2026-03-07 14:45:20.286913169 +0000 UTC m=+362.239049677" observedRunningTime="2026-03-07 14:45:20.908640542 +0000 UTC m=+362.860777060" watchObservedRunningTime="2026-03-07 14:45:20.913683118 +0000 UTC m=+362.865819626" Mar 07 14:45:20 crc kubenswrapper[4943]: I0307 14:45:20.930611 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-b2fvs" podStartSLOduration=3.437024873 podStartE2EDuration="4.930565539s" podCreationTimestamp="2026-03-07 14:45:16 +0000 UTC" firstStartedPulling="2026-03-07 14:45:18.857823659 +0000 UTC m=+360.809960157" lastFinishedPulling="2026-03-07 14:45:20.351364305 +0000 UTC m=+362.303500823" observedRunningTime="2026-03-07 14:45:20.927447701 +0000 UTC m=+362.879584219" watchObservedRunningTime="2026-03-07 14:45:20.930565539 +0000 UTC m=+362.882702037" Mar 07 14:45:24 crc kubenswrapper[4943]: I0307 14:45:24.802559 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4pbcq" Mar 07 14:45:24 crc kubenswrapper[4943]: I0307 14:45:24.803070 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4pbcq" Mar 07 14:45:24 crc kubenswrapper[4943]: I0307 14:45:24.884063 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4pbcq" Mar 07 14:45:24 crc kubenswrapper[4943]: I0307 14:45:24.965112 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4pbcq" Mar 07 14:45:25 crc kubenswrapper[4943]: I0307 14:45:25.029960 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rhp5s" Mar 07 14:45:25 crc kubenswrapper[4943]: I0307 14:45:25.030069 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rhp5s" Mar 07 14:45:25 crc kubenswrapper[4943]: I0307 14:45:25.094545 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rhp5s" Mar 07 14:45:25 crc kubenswrapper[4943]: I0307 14:45:25.985284 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rhp5s" Mar 07 14:45:27 crc kubenswrapper[4943]: I0307 14:45:27.346762 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-b2fvs" Mar 07 14:45:27 crc kubenswrapper[4943]: I0307 14:45:27.347348 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-b2fvs" Mar 07 14:45:27 crc kubenswrapper[4943]: I0307 14:45:27.370148 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-k6478" Mar 07 14:45:27 crc kubenswrapper[4943]: I0307 14:45:27.370430 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-k6478" Mar 07 14:45:27 crc kubenswrapper[4943]: I0307 14:45:27.417328 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-b2fvs" Mar 07 14:45:27 crc kubenswrapper[4943]: I0307 14:45:27.977807 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-b2fvs" Mar 07 14:45:28 crc kubenswrapper[4943]: I0307 14:45:28.447545 4943 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-k6478" podUID="db1d6561-f6c0-44fe-80f5-e5c6f9179c0a" containerName="registry-server" probeResult="failure" output=< Mar 07 14:45:28 crc kubenswrapper[4943]: timeout: failed to connect service ":50051" within 1s Mar 07 14:45:28 crc kubenswrapper[4943]: > Mar 07 14:45:35 crc kubenswrapper[4943]: I0307 14:45:35.577539 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" podUID="a9fcbefd-dfa7-4afd-b198-d872137a9f51" containerName="registry" containerID="cri-o://a179c3c12d12b89207c401aa3d7da0a6ff2852cb58c1e775d073ebdda8006c95" gracePeriod=30 Mar 07 14:45:36 crc kubenswrapper[4943]: I0307 14:45:36.000570 4943 generic.go:334] "Generic (PLEG): container finished" podID="a9fcbefd-dfa7-4afd-b198-d872137a9f51" containerID="a179c3c12d12b89207c401aa3d7da0a6ff2852cb58c1e775d073ebdda8006c95" exitCode=0 Mar 07 14:45:36 crc kubenswrapper[4943]: I0307 14:45:36.000635 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" event={"ID":"a9fcbefd-dfa7-4afd-b198-d872137a9f51","Type":"ContainerDied","Data":"a179c3c12d12b89207c401aa3d7da0a6ff2852cb58c1e775d073ebdda8006c95"} Mar 07 14:45:36 crc kubenswrapper[4943]: I0307 14:45:36.063898 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:45:36 crc kubenswrapper[4943]: I0307 14:45:36.107683 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " Mar 07 14:45:36 crc kubenswrapper[4943]: I0307 14:45:36.107749 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a9fcbefd-dfa7-4afd-b198-d872137a9f51-registry-certificates\") pod \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " Mar 07 14:45:36 crc kubenswrapper[4943]: I0307 14:45:36.107792 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a9fcbefd-dfa7-4afd-b198-d872137a9f51-bound-sa-token\") pod \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " Mar 07 14:45:36 crc kubenswrapper[4943]: I0307 14:45:36.108055 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a9fcbefd-dfa7-4afd-b198-d872137a9f51-registry-tls\") pod \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " Mar 07 14:45:36 crc kubenswrapper[4943]: I0307 14:45:36.108111 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lh8b9\" (UniqueName: \"kubernetes.io/projected/a9fcbefd-dfa7-4afd-b198-d872137a9f51-kube-api-access-lh8b9\") pod \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " Mar 07 14:45:36 crc kubenswrapper[4943]: I0307 14:45:36.108139 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a9fcbefd-dfa7-4afd-b198-d872137a9f51-ca-trust-extracted\") pod \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " Mar 07 14:45:36 crc kubenswrapper[4943]: I0307 14:45:36.108186 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a9fcbefd-dfa7-4afd-b198-d872137a9f51-trusted-ca\") pod \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " Mar 07 14:45:36 crc kubenswrapper[4943]: I0307 14:45:36.108230 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a9fcbefd-dfa7-4afd-b198-d872137a9f51-installation-pull-secrets\") pod \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\" (UID: \"a9fcbefd-dfa7-4afd-b198-d872137a9f51\") " Mar 07 14:45:36 crc kubenswrapper[4943]: I0307 14:45:36.109428 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9fcbefd-dfa7-4afd-b198-d872137a9f51-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "a9fcbefd-dfa7-4afd-b198-d872137a9f51" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:45:36 crc kubenswrapper[4943]: I0307 14:45:36.110030 4943 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a9fcbefd-dfa7-4afd-b198-d872137a9f51-registry-certificates\") on node \"crc\" DevicePath \"\"" Mar 07 14:45:36 crc kubenswrapper[4943]: I0307 14:45:36.110527 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9fcbefd-dfa7-4afd-b198-d872137a9f51-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a9fcbefd-dfa7-4afd-b198-d872137a9f51" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:45:36 crc kubenswrapper[4943]: I0307 14:45:36.123368 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9fcbefd-dfa7-4afd-b198-d872137a9f51-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a9fcbefd-dfa7-4afd-b198-d872137a9f51" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:45:36 crc kubenswrapper[4943]: I0307 14:45:36.125573 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9fcbefd-dfa7-4afd-b198-d872137a9f51-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "a9fcbefd-dfa7-4afd-b198-d872137a9f51" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:45:36 crc kubenswrapper[4943]: I0307 14:45:36.126623 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9fcbefd-dfa7-4afd-b198-d872137a9f51-kube-api-access-lh8b9" (OuterVolumeSpecName: "kube-api-access-lh8b9") pod "a9fcbefd-dfa7-4afd-b198-d872137a9f51" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51"). InnerVolumeSpecName "kube-api-access-lh8b9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:45:36 crc kubenswrapper[4943]: I0307 14:45:36.127392 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9fcbefd-dfa7-4afd-b198-d872137a9f51-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "a9fcbefd-dfa7-4afd-b198-d872137a9f51" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:45:36 crc kubenswrapper[4943]: I0307 14:45:36.131427 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "a9fcbefd-dfa7-4afd-b198-d872137a9f51" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 07 14:45:36 crc kubenswrapper[4943]: I0307 14:45:36.156429 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9fcbefd-dfa7-4afd-b198-d872137a9f51-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "a9fcbefd-dfa7-4afd-b198-d872137a9f51" (UID: "a9fcbefd-dfa7-4afd-b198-d872137a9f51"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:45:36 crc kubenswrapper[4943]: I0307 14:45:36.211271 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lh8b9\" (UniqueName: \"kubernetes.io/projected/a9fcbefd-dfa7-4afd-b198-d872137a9f51-kube-api-access-lh8b9\") on node \"crc\" DevicePath \"\"" Mar 07 14:45:36 crc kubenswrapper[4943]: I0307 14:45:36.211701 4943 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a9fcbefd-dfa7-4afd-b198-d872137a9f51-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Mar 07 14:45:36 crc kubenswrapper[4943]: I0307 14:45:36.211718 4943 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a9fcbefd-dfa7-4afd-b198-d872137a9f51-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 07 14:45:36 crc kubenswrapper[4943]: I0307 14:45:36.211734 4943 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a9fcbefd-dfa7-4afd-b198-d872137a9f51-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Mar 07 14:45:36 crc kubenswrapper[4943]: I0307 14:45:36.211749 4943 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a9fcbefd-dfa7-4afd-b198-d872137a9f51-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 07 14:45:36 crc kubenswrapper[4943]: I0307 14:45:36.211761 4943 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a9fcbefd-dfa7-4afd-b198-d872137a9f51-registry-tls\") on node \"crc\" DevicePath \"\"" Mar 07 14:45:37 crc kubenswrapper[4943]: I0307 14:45:37.010410 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" event={"ID":"a9fcbefd-dfa7-4afd-b198-d872137a9f51","Type":"ContainerDied","Data":"72ebd4d6c80a14ea3189fd90adacea2d525b1ffc057ba0a83e9c5aaf8b1eeb98"} Mar 07 14:45:37 crc kubenswrapper[4943]: I0307 14:45:37.010469 4943 scope.go:117] "RemoveContainer" containerID="a179c3c12d12b89207c401aa3d7da0a6ff2852cb58c1e775d073ebdda8006c95" Mar 07 14:45:37 crc kubenswrapper[4943]: I0307 14:45:37.010604 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-r82fw" Mar 07 14:45:37 crc kubenswrapper[4943]: I0307 14:45:37.041117 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-r82fw"] Mar 07 14:45:37 crc kubenswrapper[4943]: I0307 14:45:37.047787 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-r82fw"] Mar 07 14:45:37 crc kubenswrapper[4943]: I0307 14:45:37.426913 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-k6478" Mar 07 14:45:37 crc kubenswrapper[4943]: I0307 14:45:37.483251 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-k6478" Mar 07 14:45:38 crc kubenswrapper[4943]: I0307 14:45:38.767381 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9fcbefd-dfa7-4afd-b198-d872137a9f51" path="/var/lib/kubelet/pods/a9fcbefd-dfa7-4afd-b198-d872137a9f51/volumes" Mar 07 14:46:00 crc kubenswrapper[4943]: I0307 14:46:00.188051 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29548246-r7df9"] Mar 07 14:46:00 crc kubenswrapper[4943]: E0307 14:46:00.188870 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9fcbefd-dfa7-4afd-b198-d872137a9f51" containerName="registry" Mar 07 14:46:00 crc kubenswrapper[4943]: I0307 14:46:00.188884 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9fcbefd-dfa7-4afd-b198-d872137a9f51" containerName="registry" Mar 07 14:46:00 crc kubenswrapper[4943]: I0307 14:46:00.189049 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9fcbefd-dfa7-4afd-b198-d872137a9f51" containerName="registry" Mar 07 14:46:00 crc kubenswrapper[4943]: I0307 14:46:00.189535 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548246-r7df9" Mar 07 14:46:00 crc kubenswrapper[4943]: I0307 14:46:00.200233 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-qqlk5" Mar 07 14:46:00 crc kubenswrapper[4943]: I0307 14:46:00.200587 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 07 14:46:00 crc kubenswrapper[4943]: I0307 14:46:00.200596 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 07 14:46:00 crc kubenswrapper[4943]: I0307 14:46:00.212531 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29548246-r7df9"] Mar 07 14:46:00 crc kubenswrapper[4943]: I0307 14:46:00.303253 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4p8s5\" (UniqueName: \"kubernetes.io/projected/5dceb930-7f1e-49cd-a53d-db7045ab6984-kube-api-access-4p8s5\") pod \"auto-csr-approver-29548246-r7df9\" (UID: \"5dceb930-7f1e-49cd-a53d-db7045ab6984\") " pod="openshift-infra/auto-csr-approver-29548246-r7df9" Mar 07 14:46:00 crc kubenswrapper[4943]: I0307 14:46:00.405225 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4p8s5\" (UniqueName: \"kubernetes.io/projected/5dceb930-7f1e-49cd-a53d-db7045ab6984-kube-api-access-4p8s5\") pod \"auto-csr-approver-29548246-r7df9\" (UID: \"5dceb930-7f1e-49cd-a53d-db7045ab6984\") " pod="openshift-infra/auto-csr-approver-29548246-r7df9" Mar 07 14:46:00 crc kubenswrapper[4943]: I0307 14:46:00.438652 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4p8s5\" (UniqueName: \"kubernetes.io/projected/5dceb930-7f1e-49cd-a53d-db7045ab6984-kube-api-access-4p8s5\") pod \"auto-csr-approver-29548246-r7df9\" (UID: \"5dceb930-7f1e-49cd-a53d-db7045ab6984\") " pod="openshift-infra/auto-csr-approver-29548246-r7df9" Mar 07 14:46:00 crc kubenswrapper[4943]: I0307 14:46:00.533825 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548246-r7df9" Mar 07 14:46:00 crc kubenswrapper[4943]: I0307 14:46:00.795320 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29548246-r7df9"] Mar 07 14:46:01 crc kubenswrapper[4943]: I0307 14:46:01.183665 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548246-r7df9" event={"ID":"5dceb930-7f1e-49cd-a53d-db7045ab6984","Type":"ContainerStarted","Data":"dc17fed17d15f2c5136aa190b523fe2536798b57a18b39841c276448b8ebc434"} Mar 07 14:46:02 crc kubenswrapper[4943]: I0307 14:46:02.195570 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548246-r7df9" event={"ID":"5dceb930-7f1e-49cd-a53d-db7045ab6984","Type":"ContainerStarted","Data":"0571265e58bbf9b45930119ab783ac8030ba47a1e31fd5d62b1f1d9f596b4d55"} Mar 07 14:46:02 crc kubenswrapper[4943]: I0307 14:46:02.222845 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29548246-r7df9" podStartSLOduration=1.3197576 podStartE2EDuration="2.222815077s" podCreationTimestamp="2026-03-07 14:46:00 +0000 UTC" firstStartedPulling="2026-03-07 14:46:00.804341908 +0000 UTC m=+402.756478406" lastFinishedPulling="2026-03-07 14:46:01.707399345 +0000 UTC m=+403.659535883" observedRunningTime="2026-03-07 14:46:02.22134185 +0000 UTC m=+404.173478378" watchObservedRunningTime="2026-03-07 14:46:02.222815077 +0000 UTC m=+404.174951615" Mar 07 14:46:03 crc kubenswrapper[4943]: I0307 14:46:03.207887 4943 generic.go:334] "Generic (PLEG): container finished" podID="5dceb930-7f1e-49cd-a53d-db7045ab6984" containerID="0571265e58bbf9b45930119ab783ac8030ba47a1e31fd5d62b1f1d9f596b4d55" exitCode=0 Mar 07 14:46:03 crc kubenswrapper[4943]: I0307 14:46:03.208013 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548246-r7df9" event={"ID":"5dceb930-7f1e-49cd-a53d-db7045ab6984","Type":"ContainerDied","Data":"0571265e58bbf9b45930119ab783ac8030ba47a1e31fd5d62b1f1d9f596b4d55"} Mar 07 14:46:04 crc kubenswrapper[4943]: I0307 14:46:04.517010 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548246-r7df9" Mar 07 14:46:04 crc kubenswrapper[4943]: I0307 14:46:04.575960 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4p8s5\" (UniqueName: \"kubernetes.io/projected/5dceb930-7f1e-49cd-a53d-db7045ab6984-kube-api-access-4p8s5\") pod \"5dceb930-7f1e-49cd-a53d-db7045ab6984\" (UID: \"5dceb930-7f1e-49cd-a53d-db7045ab6984\") " Mar 07 14:46:04 crc kubenswrapper[4943]: I0307 14:46:04.581094 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5dceb930-7f1e-49cd-a53d-db7045ab6984-kube-api-access-4p8s5" (OuterVolumeSpecName: "kube-api-access-4p8s5") pod "5dceb930-7f1e-49cd-a53d-db7045ab6984" (UID: "5dceb930-7f1e-49cd-a53d-db7045ab6984"). InnerVolumeSpecName "kube-api-access-4p8s5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:46:04 crc kubenswrapper[4943]: I0307 14:46:04.678211 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4p8s5\" (UniqueName: \"kubernetes.io/projected/5dceb930-7f1e-49cd-a53d-db7045ab6984-kube-api-access-4p8s5\") on node \"crc\" DevicePath \"\"" Mar 07 14:46:05 crc kubenswrapper[4943]: I0307 14:46:05.225140 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548246-r7df9" event={"ID":"5dceb930-7f1e-49cd-a53d-db7045ab6984","Type":"ContainerDied","Data":"dc17fed17d15f2c5136aa190b523fe2536798b57a18b39841c276448b8ebc434"} Mar 07 14:46:05 crc kubenswrapper[4943]: I0307 14:46:05.225207 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548246-r7df9" Mar 07 14:46:05 crc kubenswrapper[4943]: I0307 14:46:05.225216 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dc17fed17d15f2c5136aa190b523fe2536798b57a18b39841c276448b8ebc434" Mar 07 14:46:06 crc kubenswrapper[4943]: I0307 14:46:06.074845 4943 patch_prober.go:28] interesting pod/machine-config-daemon-cpdmm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 07 14:46:06 crc kubenswrapper[4943]: I0307 14:46:06.075338 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 07 14:46:36 crc kubenswrapper[4943]: I0307 14:46:36.074428 4943 patch_prober.go:28] interesting pod/machine-config-daemon-cpdmm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 07 14:46:36 crc kubenswrapper[4943]: I0307 14:46:36.075488 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 07 14:47:06 crc kubenswrapper[4943]: I0307 14:47:06.075048 4943 patch_prober.go:28] interesting pod/machine-config-daemon-cpdmm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 07 14:47:06 crc kubenswrapper[4943]: I0307 14:47:06.075892 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 07 14:47:06 crc kubenswrapper[4943]: I0307 14:47:06.075998 4943 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" Mar 07 14:47:06 crc kubenswrapper[4943]: I0307 14:47:06.077074 4943 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a9980c0c2b00970e1182ba487de7ac1c1a910169389e41db4b54d88fd83453b4"} pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 07 14:47:06 crc kubenswrapper[4943]: I0307 14:47:06.077274 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerName="machine-config-daemon" containerID="cri-o://a9980c0c2b00970e1182ba487de7ac1c1a910169389e41db4b54d88fd83453b4" gracePeriod=600 Mar 07 14:47:06 crc kubenswrapper[4943]: I0307 14:47:06.749195 4943 generic.go:334] "Generic (PLEG): container finished" podID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerID="a9980c0c2b00970e1182ba487de7ac1c1a910169389e41db4b54d88fd83453b4" exitCode=0 Mar 07 14:47:06 crc kubenswrapper[4943]: I0307 14:47:06.749248 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" event={"ID":"4c1f18ea-e557-41fc-95dd-ce4e1722a61e","Type":"ContainerDied","Data":"a9980c0c2b00970e1182ba487de7ac1c1a910169389e41db4b54d88fd83453b4"} Mar 07 14:47:06 crc kubenswrapper[4943]: I0307 14:47:06.749895 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" event={"ID":"4c1f18ea-e557-41fc-95dd-ce4e1722a61e","Type":"ContainerStarted","Data":"b2616e0e9d2516da6111a60698ff218442b059990c7d49e009916d0f941b033b"} Mar 07 14:47:06 crc kubenswrapper[4943]: I0307 14:47:06.749995 4943 scope.go:117] "RemoveContainer" containerID="fca9947e0c4ec525f0c6b799660a1244b1d37a4616b55a03cd6f059674d81757" Mar 07 14:48:00 crc kubenswrapper[4943]: I0307 14:48:00.155739 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29548248-2459n"] Mar 07 14:48:00 crc kubenswrapper[4943]: E0307 14:48:00.156789 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5dceb930-7f1e-49cd-a53d-db7045ab6984" containerName="oc" Mar 07 14:48:00 crc kubenswrapper[4943]: I0307 14:48:00.156812 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="5dceb930-7f1e-49cd-a53d-db7045ab6984" containerName="oc" Mar 07 14:48:00 crc kubenswrapper[4943]: I0307 14:48:00.157037 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="5dceb930-7f1e-49cd-a53d-db7045ab6984" containerName="oc" Mar 07 14:48:00 crc kubenswrapper[4943]: I0307 14:48:00.157693 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548248-2459n" Mar 07 14:48:00 crc kubenswrapper[4943]: I0307 14:48:00.160843 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 07 14:48:00 crc kubenswrapper[4943]: I0307 14:48:00.161471 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 07 14:48:00 crc kubenswrapper[4943]: I0307 14:48:00.163412 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29548248-2459n"] Mar 07 14:48:00 crc kubenswrapper[4943]: I0307 14:48:00.164084 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-qqlk5" Mar 07 14:48:00 crc kubenswrapper[4943]: I0307 14:48:00.236217 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-km8dr\" (UniqueName: \"kubernetes.io/projected/1349e3eb-9576-43fe-bd36-6e2832255ae4-kube-api-access-km8dr\") pod \"auto-csr-approver-29548248-2459n\" (UID: \"1349e3eb-9576-43fe-bd36-6e2832255ae4\") " pod="openshift-infra/auto-csr-approver-29548248-2459n" Mar 07 14:48:00 crc kubenswrapper[4943]: I0307 14:48:00.338312 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-km8dr\" (UniqueName: \"kubernetes.io/projected/1349e3eb-9576-43fe-bd36-6e2832255ae4-kube-api-access-km8dr\") pod \"auto-csr-approver-29548248-2459n\" (UID: \"1349e3eb-9576-43fe-bd36-6e2832255ae4\") " pod="openshift-infra/auto-csr-approver-29548248-2459n" Mar 07 14:48:00 crc kubenswrapper[4943]: I0307 14:48:00.372645 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-km8dr\" (UniqueName: \"kubernetes.io/projected/1349e3eb-9576-43fe-bd36-6e2832255ae4-kube-api-access-km8dr\") pod \"auto-csr-approver-29548248-2459n\" (UID: \"1349e3eb-9576-43fe-bd36-6e2832255ae4\") " pod="openshift-infra/auto-csr-approver-29548248-2459n" Mar 07 14:48:00 crc kubenswrapper[4943]: I0307 14:48:00.492123 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548248-2459n" Mar 07 14:48:00 crc kubenswrapper[4943]: I0307 14:48:00.769306 4943 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 07 14:48:00 crc kubenswrapper[4943]: I0307 14:48:00.775122 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29548248-2459n"] Mar 07 14:48:01 crc kubenswrapper[4943]: I0307 14:48:01.161472 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548248-2459n" event={"ID":"1349e3eb-9576-43fe-bd36-6e2832255ae4","Type":"ContainerStarted","Data":"141b4b58362b87e2c6bb613388f71af5c4d2e9f681c1cb685e9ff37eedbf4cdb"} Mar 07 14:48:02 crc kubenswrapper[4943]: I0307 14:48:02.172762 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548248-2459n" event={"ID":"1349e3eb-9576-43fe-bd36-6e2832255ae4","Type":"ContainerStarted","Data":"c3c1ce363e4f0f7a4f265024b98311e29e9ea1ea4b506f5fff498cb14b3ee2fd"} Mar 07 14:48:02 crc kubenswrapper[4943]: I0307 14:48:02.193977 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29548248-2459n" podStartSLOduration=1.266275376 podStartE2EDuration="2.193923453s" podCreationTimestamp="2026-03-07 14:48:00 +0000 UTC" firstStartedPulling="2026-03-07 14:48:00.769017767 +0000 UTC m=+522.721154275" lastFinishedPulling="2026-03-07 14:48:01.696665824 +0000 UTC m=+523.648802352" observedRunningTime="2026-03-07 14:48:02.191320199 +0000 UTC m=+524.143456727" watchObservedRunningTime="2026-03-07 14:48:02.193923453 +0000 UTC m=+524.146059981" Mar 07 14:48:03 crc kubenswrapper[4943]: I0307 14:48:03.181160 4943 generic.go:334] "Generic (PLEG): container finished" podID="1349e3eb-9576-43fe-bd36-6e2832255ae4" containerID="c3c1ce363e4f0f7a4f265024b98311e29e9ea1ea4b506f5fff498cb14b3ee2fd" exitCode=0 Mar 07 14:48:03 crc kubenswrapper[4943]: I0307 14:48:03.181225 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548248-2459n" event={"ID":"1349e3eb-9576-43fe-bd36-6e2832255ae4","Type":"ContainerDied","Data":"c3c1ce363e4f0f7a4f265024b98311e29e9ea1ea4b506f5fff498cb14b3ee2fd"} Mar 07 14:48:04 crc kubenswrapper[4943]: I0307 14:48:04.469850 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548248-2459n" Mar 07 14:48:04 crc kubenswrapper[4943]: I0307 14:48:04.510469 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-km8dr\" (UniqueName: \"kubernetes.io/projected/1349e3eb-9576-43fe-bd36-6e2832255ae4-kube-api-access-km8dr\") pod \"1349e3eb-9576-43fe-bd36-6e2832255ae4\" (UID: \"1349e3eb-9576-43fe-bd36-6e2832255ae4\") " Mar 07 14:48:04 crc kubenswrapper[4943]: I0307 14:48:04.518362 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1349e3eb-9576-43fe-bd36-6e2832255ae4-kube-api-access-km8dr" (OuterVolumeSpecName: "kube-api-access-km8dr") pod "1349e3eb-9576-43fe-bd36-6e2832255ae4" (UID: "1349e3eb-9576-43fe-bd36-6e2832255ae4"). InnerVolumeSpecName "kube-api-access-km8dr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:48:04 crc kubenswrapper[4943]: I0307 14:48:04.611954 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-km8dr\" (UniqueName: \"kubernetes.io/projected/1349e3eb-9576-43fe-bd36-6e2832255ae4-kube-api-access-km8dr\") on node \"crc\" DevicePath \"\"" Mar 07 14:48:05 crc kubenswrapper[4943]: I0307 14:48:05.207393 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548248-2459n" event={"ID":"1349e3eb-9576-43fe-bd36-6e2832255ae4","Type":"ContainerDied","Data":"141b4b58362b87e2c6bb613388f71af5c4d2e9f681c1cb685e9ff37eedbf4cdb"} Mar 07 14:48:05 crc kubenswrapper[4943]: I0307 14:48:05.207474 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="141b4b58362b87e2c6bb613388f71af5c4d2e9f681c1cb685e9ff37eedbf4cdb" Mar 07 14:48:05 crc kubenswrapper[4943]: I0307 14:48:05.207568 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548248-2459n" Mar 07 14:48:05 crc kubenswrapper[4943]: I0307 14:48:05.261075 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29548242-scpk7"] Mar 07 14:48:05 crc kubenswrapper[4943]: I0307 14:48:05.268601 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29548242-scpk7"] Mar 07 14:48:06 crc kubenswrapper[4943]: I0307 14:48:06.767287 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14d07994-bd34-4802-8dd5-8fd8afe5e878" path="/var/lib/kubelet/pods/14d07994-bd34-4802-8dd5-8fd8afe5e878/volumes" Mar 07 14:48:19 crc kubenswrapper[4943]: I0307 14:48:19.181714 4943 scope.go:117] "RemoveContainer" containerID="ac2a8956e015025d29002a8478f290cd8b4087387667c0a6c52658f31f2d8176" Mar 07 14:49:06 crc kubenswrapper[4943]: I0307 14:49:06.074799 4943 patch_prober.go:28] interesting pod/machine-config-daemon-cpdmm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 07 14:49:06 crc kubenswrapper[4943]: I0307 14:49:06.075543 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 07 14:49:19 crc kubenswrapper[4943]: I0307 14:49:19.262672 4943 scope.go:117] "RemoveContainer" containerID="51e6cc808f78e2b17b8de230ede94a6602929050d1899c36e29302d6685c6f18" Mar 07 14:49:36 crc kubenswrapper[4943]: I0307 14:49:36.073859 4943 patch_prober.go:28] interesting pod/machine-config-daemon-cpdmm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 07 14:49:36 crc kubenswrapper[4943]: I0307 14:49:36.074561 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 07 14:50:00 crc kubenswrapper[4943]: I0307 14:50:00.146272 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29548250-cgbqg"] Mar 07 14:50:00 crc kubenswrapper[4943]: E0307 14:50:00.147245 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1349e3eb-9576-43fe-bd36-6e2832255ae4" containerName="oc" Mar 07 14:50:00 crc kubenswrapper[4943]: I0307 14:50:00.147267 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="1349e3eb-9576-43fe-bd36-6e2832255ae4" containerName="oc" Mar 07 14:50:00 crc kubenswrapper[4943]: I0307 14:50:00.147466 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="1349e3eb-9576-43fe-bd36-6e2832255ae4" containerName="oc" Mar 07 14:50:00 crc kubenswrapper[4943]: I0307 14:50:00.148057 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548250-cgbqg" Mar 07 14:50:00 crc kubenswrapper[4943]: I0307 14:50:00.154012 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 07 14:50:00 crc kubenswrapper[4943]: I0307 14:50:00.155050 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-qqlk5" Mar 07 14:50:00 crc kubenswrapper[4943]: I0307 14:50:00.156621 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 07 14:50:00 crc kubenswrapper[4943]: I0307 14:50:00.161598 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29548250-cgbqg"] Mar 07 14:50:00 crc kubenswrapper[4943]: I0307 14:50:00.265394 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-976kn\" (UniqueName: \"kubernetes.io/projected/ce0f9eb4-53fa-4a58-b68f-e4765e56739d-kube-api-access-976kn\") pod \"auto-csr-approver-29548250-cgbqg\" (UID: \"ce0f9eb4-53fa-4a58-b68f-e4765e56739d\") " pod="openshift-infra/auto-csr-approver-29548250-cgbqg" Mar 07 14:50:00 crc kubenswrapper[4943]: I0307 14:50:00.367637 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-976kn\" (UniqueName: \"kubernetes.io/projected/ce0f9eb4-53fa-4a58-b68f-e4765e56739d-kube-api-access-976kn\") pod \"auto-csr-approver-29548250-cgbqg\" (UID: \"ce0f9eb4-53fa-4a58-b68f-e4765e56739d\") " pod="openshift-infra/auto-csr-approver-29548250-cgbqg" Mar 07 14:50:00 crc kubenswrapper[4943]: I0307 14:50:00.400052 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-976kn\" (UniqueName: \"kubernetes.io/projected/ce0f9eb4-53fa-4a58-b68f-e4765e56739d-kube-api-access-976kn\") pod \"auto-csr-approver-29548250-cgbqg\" (UID: \"ce0f9eb4-53fa-4a58-b68f-e4765e56739d\") " pod="openshift-infra/auto-csr-approver-29548250-cgbqg" Mar 07 14:50:00 crc kubenswrapper[4943]: I0307 14:50:00.525559 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548250-cgbqg" Mar 07 14:50:00 crc kubenswrapper[4943]: I0307 14:50:00.819980 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29548250-cgbqg"] Mar 07 14:50:01 crc kubenswrapper[4943]: I0307 14:50:01.049355 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548250-cgbqg" event={"ID":"ce0f9eb4-53fa-4a58-b68f-e4765e56739d","Type":"ContainerStarted","Data":"68dc4073e1a23af8247d300f2f48fd676fa78eaeabf92f48eba75ca41609c22a"} Mar 07 14:50:03 crc kubenswrapper[4943]: I0307 14:50:03.077208 4943 generic.go:334] "Generic (PLEG): container finished" podID="ce0f9eb4-53fa-4a58-b68f-e4765e56739d" containerID="47a10ef328ff258dcc2bbdee9f96df4df3cb5742ca81b9d42a9a55dcb2c42fb4" exitCode=0 Mar 07 14:50:03 crc kubenswrapper[4943]: I0307 14:50:03.077385 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548250-cgbqg" event={"ID":"ce0f9eb4-53fa-4a58-b68f-e4765e56739d","Type":"ContainerDied","Data":"47a10ef328ff258dcc2bbdee9f96df4df3cb5742ca81b9d42a9a55dcb2c42fb4"} Mar 07 14:50:04 crc kubenswrapper[4943]: I0307 14:50:04.441700 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548250-cgbqg" Mar 07 14:50:04 crc kubenswrapper[4943]: I0307 14:50:04.536568 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-976kn\" (UniqueName: \"kubernetes.io/projected/ce0f9eb4-53fa-4a58-b68f-e4765e56739d-kube-api-access-976kn\") pod \"ce0f9eb4-53fa-4a58-b68f-e4765e56739d\" (UID: \"ce0f9eb4-53fa-4a58-b68f-e4765e56739d\") " Mar 07 14:50:04 crc kubenswrapper[4943]: I0307 14:50:04.548307 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce0f9eb4-53fa-4a58-b68f-e4765e56739d-kube-api-access-976kn" (OuterVolumeSpecName: "kube-api-access-976kn") pod "ce0f9eb4-53fa-4a58-b68f-e4765e56739d" (UID: "ce0f9eb4-53fa-4a58-b68f-e4765e56739d"). InnerVolumeSpecName "kube-api-access-976kn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:50:04 crc kubenswrapper[4943]: I0307 14:50:04.638654 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-976kn\" (UniqueName: \"kubernetes.io/projected/ce0f9eb4-53fa-4a58-b68f-e4765e56739d-kube-api-access-976kn\") on node \"crc\" DevicePath \"\"" Mar 07 14:50:05 crc kubenswrapper[4943]: I0307 14:50:05.096265 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548250-cgbqg" event={"ID":"ce0f9eb4-53fa-4a58-b68f-e4765e56739d","Type":"ContainerDied","Data":"68dc4073e1a23af8247d300f2f48fd676fa78eaeabf92f48eba75ca41609c22a"} Mar 07 14:50:05 crc kubenswrapper[4943]: I0307 14:50:05.096327 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="68dc4073e1a23af8247d300f2f48fd676fa78eaeabf92f48eba75ca41609c22a" Mar 07 14:50:05 crc kubenswrapper[4943]: I0307 14:50:05.096351 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548250-cgbqg" Mar 07 14:50:05 crc kubenswrapper[4943]: I0307 14:50:05.527401 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29548244-2zdk7"] Mar 07 14:50:05 crc kubenswrapper[4943]: I0307 14:50:05.535107 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29548244-2zdk7"] Mar 07 14:50:06 crc kubenswrapper[4943]: I0307 14:50:06.074057 4943 patch_prober.go:28] interesting pod/machine-config-daemon-cpdmm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 07 14:50:06 crc kubenswrapper[4943]: I0307 14:50:06.074136 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 07 14:50:06 crc kubenswrapper[4943]: I0307 14:50:06.074202 4943 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" Mar 07 14:50:06 crc kubenswrapper[4943]: I0307 14:50:06.074852 4943 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b2616e0e9d2516da6111a60698ff218442b059990c7d49e009916d0f941b033b"} pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 07 14:50:06 crc kubenswrapper[4943]: I0307 14:50:06.074984 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerName="machine-config-daemon" containerID="cri-o://b2616e0e9d2516da6111a60698ff218442b059990c7d49e009916d0f941b033b" gracePeriod=600 Mar 07 14:50:06 crc kubenswrapper[4943]: I0307 14:50:06.768480 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c43c4c0-8965-43f9-9724-6970c0f15729" path="/var/lib/kubelet/pods/0c43c4c0-8965-43f9-9724-6970c0f15729/volumes" Mar 07 14:50:07 crc kubenswrapper[4943]: I0307 14:50:07.113697 4943 generic.go:334] "Generic (PLEG): container finished" podID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerID="b2616e0e9d2516da6111a60698ff218442b059990c7d49e009916d0f941b033b" exitCode=0 Mar 07 14:50:07 crc kubenswrapper[4943]: I0307 14:50:07.113781 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" event={"ID":"4c1f18ea-e557-41fc-95dd-ce4e1722a61e","Type":"ContainerDied","Data":"b2616e0e9d2516da6111a60698ff218442b059990c7d49e009916d0f941b033b"} Mar 07 14:50:07 crc kubenswrapper[4943]: I0307 14:50:07.113848 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" event={"ID":"4c1f18ea-e557-41fc-95dd-ce4e1722a61e","Type":"ContainerStarted","Data":"8ffd8d858c956cfb1e523c263c554f1c6d621212d5b2c18a1bc047ffbcf47fee"} Mar 07 14:50:07 crc kubenswrapper[4943]: I0307 14:50:07.113885 4943 scope.go:117] "RemoveContainer" containerID="a9980c0c2b00970e1182ba487de7ac1c1a910169389e41db4b54d88fd83453b4" Mar 07 14:50:19 crc kubenswrapper[4943]: I0307 14:50:19.323071 4943 scope.go:117] "RemoveContainer" containerID="a0539578f6fccd6a106c51804570f0018ba0682b32ca76a6ae19a1fcf8ccf643" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.259170 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-s8cw7"] Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.260632 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" podUID="d340687c-faf0-48dc-a1ac-dec5067ac665" containerName="ovn-controller" containerID="cri-o://6a089f76e974f8c12f32f42b96b301fda03651276c599841b9c42fa064ffd5fd" gracePeriod=30 Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.261235 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" podUID="d340687c-faf0-48dc-a1ac-dec5067ac665" containerName="sbdb" containerID="cri-o://f56c9c195453268c830bf92a70a64f70e75ef30e8ca853bb1f3a054a69c791af" gracePeriod=30 Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.261278 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" podUID="d340687c-faf0-48dc-a1ac-dec5067ac665" containerName="nbdb" containerID="cri-o://50e79fc40569f5b3fd73286933ee0242a1edd807a820edf7ba00a856d81c82c9" gracePeriod=30 Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.261312 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" podUID="d340687c-faf0-48dc-a1ac-dec5067ac665" containerName="northd" containerID="cri-o://959458a05302e835122b9db5d1eb4c449749a9e7ba2d85459edaeef2eede12de" gracePeriod=30 Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.261345 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" podUID="d340687c-faf0-48dc-a1ac-dec5067ac665" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://a64d887ab03c882e1a27a4cf36f359a8d18e3c75b0a6d81f33569e1e3b294280" gracePeriod=30 Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.261375 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" podUID="d340687c-faf0-48dc-a1ac-dec5067ac665" containerName="kube-rbac-proxy-node" containerID="cri-o://6fb3f1769f0d1092192c2f5f152c7d1fc52813d2e722687fb3ea1f6f08fba808" gracePeriod=30 Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.261407 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" podUID="d340687c-faf0-48dc-a1ac-dec5067ac665" containerName="ovn-acl-logging" containerID="cri-o://4dd5c06470991a2bcdddf5a5e429adcf4f4af8d3742de9c0a1e9cdd5705f74c5" gracePeriod=30 Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.315423 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" podUID="d340687c-faf0-48dc-a1ac-dec5067ac665" containerName="ovnkube-controller" containerID="cri-o://a200f68bc20411f516725f81a930664ba5baaefd21a2792866262ded7e86fae5" gracePeriod=30 Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.627701 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-s8cw7_d340687c-faf0-48dc-a1ac-dec5067ac665/ovn-acl-logging/0.log" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.628967 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-s8cw7_d340687c-faf0-48dc-a1ac-dec5067ac665/ovn-controller/0.log" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.629633 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.690557 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-npgnc"] Mar 07 14:50:26 crc kubenswrapper[4943]: E0307 14:50:26.690980 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d340687c-faf0-48dc-a1ac-dec5067ac665" containerName="kubecfg-setup" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.691017 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="d340687c-faf0-48dc-a1ac-dec5067ac665" containerName="kubecfg-setup" Mar 07 14:50:26 crc kubenswrapper[4943]: E0307 14:50:26.691050 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d340687c-faf0-48dc-a1ac-dec5067ac665" containerName="kube-rbac-proxy-node" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.691066 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="d340687c-faf0-48dc-a1ac-dec5067ac665" containerName="kube-rbac-proxy-node" Mar 07 14:50:26 crc kubenswrapper[4943]: E0307 14:50:26.691088 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d340687c-faf0-48dc-a1ac-dec5067ac665" containerName="kube-rbac-proxy-ovn-metrics" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.691104 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="d340687c-faf0-48dc-a1ac-dec5067ac665" containerName="kube-rbac-proxy-ovn-metrics" Mar 07 14:50:26 crc kubenswrapper[4943]: E0307 14:50:26.691125 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d340687c-faf0-48dc-a1ac-dec5067ac665" containerName="sbdb" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.691142 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="d340687c-faf0-48dc-a1ac-dec5067ac665" containerName="sbdb" Mar 07 14:50:26 crc kubenswrapper[4943]: E0307 14:50:26.691164 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d340687c-faf0-48dc-a1ac-dec5067ac665" containerName="ovn-controller" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.691180 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="d340687c-faf0-48dc-a1ac-dec5067ac665" containerName="ovn-controller" Mar 07 14:50:26 crc kubenswrapper[4943]: E0307 14:50:26.691207 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce0f9eb4-53fa-4a58-b68f-e4765e56739d" containerName="oc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.691225 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce0f9eb4-53fa-4a58-b68f-e4765e56739d" containerName="oc" Mar 07 14:50:26 crc kubenswrapper[4943]: E0307 14:50:26.691250 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d340687c-faf0-48dc-a1ac-dec5067ac665" containerName="ovn-acl-logging" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.691266 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="d340687c-faf0-48dc-a1ac-dec5067ac665" containerName="ovn-acl-logging" Mar 07 14:50:26 crc kubenswrapper[4943]: E0307 14:50:26.691284 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d340687c-faf0-48dc-a1ac-dec5067ac665" containerName="northd" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.691300 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="d340687c-faf0-48dc-a1ac-dec5067ac665" containerName="northd" Mar 07 14:50:26 crc kubenswrapper[4943]: E0307 14:50:26.691319 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d340687c-faf0-48dc-a1ac-dec5067ac665" containerName="nbdb" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.691334 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="d340687c-faf0-48dc-a1ac-dec5067ac665" containerName="nbdb" Mar 07 14:50:26 crc kubenswrapper[4943]: E0307 14:50:26.691354 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d340687c-faf0-48dc-a1ac-dec5067ac665" containerName="ovnkube-controller" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.691369 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="d340687c-faf0-48dc-a1ac-dec5067ac665" containerName="ovnkube-controller" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.691577 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="d340687c-faf0-48dc-a1ac-dec5067ac665" containerName="ovn-controller" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.691603 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="d340687c-faf0-48dc-a1ac-dec5067ac665" containerName="ovn-acl-logging" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.691627 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="d340687c-faf0-48dc-a1ac-dec5067ac665" containerName="kube-rbac-proxy-node" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.691655 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce0f9eb4-53fa-4a58-b68f-e4765e56739d" containerName="oc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.691675 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="d340687c-faf0-48dc-a1ac-dec5067ac665" containerName="kube-rbac-proxy-ovn-metrics" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.691700 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="d340687c-faf0-48dc-a1ac-dec5067ac665" containerName="northd" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.691723 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="d340687c-faf0-48dc-a1ac-dec5067ac665" containerName="nbdb" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.691756 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="d340687c-faf0-48dc-a1ac-dec5067ac665" containerName="ovnkube-controller" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.691782 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="d340687c-faf0-48dc-a1ac-dec5067ac665" containerName="sbdb" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.695323 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.715666 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d340687c-faf0-48dc-a1ac-dec5067ac665-ovn-node-metrics-cert\") pod \"d340687c-faf0-48dc-a1ac-dec5067ac665\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.715722 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d340687c-faf0-48dc-a1ac-dec5067ac665-ovnkube-config\") pod \"d340687c-faf0-48dc-a1ac-dec5067ac665\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.715745 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d340687c-faf0-48dc-a1ac-dec5067ac665-env-overrides\") pod \"d340687c-faf0-48dc-a1ac-dec5067ac665\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.715761 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-host-kubelet\") pod \"d340687c-faf0-48dc-a1ac-dec5067ac665\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.715783 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d340687c-faf0-48dc-a1ac-dec5067ac665-ovnkube-script-lib\") pod \"d340687c-faf0-48dc-a1ac-dec5067ac665\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.715805 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-host-slash\") pod \"d340687c-faf0-48dc-a1ac-dec5067ac665\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.715826 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-systemd-units\") pod \"d340687c-faf0-48dc-a1ac-dec5067ac665\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.715847 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-run-openvswitch\") pod \"d340687c-faf0-48dc-a1ac-dec5067ac665\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.715876 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-host-cni-bin\") pod \"d340687c-faf0-48dc-a1ac-dec5067ac665\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.715890 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-run-systemd\") pod \"d340687c-faf0-48dc-a1ac-dec5067ac665\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.715904 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-host-run-netns\") pod \"d340687c-faf0-48dc-a1ac-dec5067ac665\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.715942 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-run-ovn\") pod \"d340687c-faf0-48dc-a1ac-dec5067ac665\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.715962 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-var-lib-openvswitch\") pod \"d340687c-faf0-48dc-a1ac-dec5067ac665\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.715982 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-etc-openvswitch\") pod \"d340687c-faf0-48dc-a1ac-dec5067ac665\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.716001 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-node-log\") pod \"d340687c-faf0-48dc-a1ac-dec5067ac665\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.716025 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-host-cni-netd\") pod \"d340687c-faf0-48dc-a1ac-dec5067ac665\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.716016 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-host-slash" (OuterVolumeSpecName: "host-slash") pod "d340687c-faf0-48dc-a1ac-dec5067ac665" (UID: "d340687c-faf0-48dc-a1ac-dec5067ac665"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.716044 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-host-var-lib-cni-networks-ovn-kubernetes\") pod \"d340687c-faf0-48dc-a1ac-dec5067ac665\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.716066 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-host-run-ovn-kubernetes\") pod \"d340687c-faf0-48dc-a1ac-dec5067ac665\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.716098 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p7d2m\" (UniqueName: \"kubernetes.io/projected/d340687c-faf0-48dc-a1ac-dec5067ac665-kube-api-access-p7d2m\") pod \"d340687c-faf0-48dc-a1ac-dec5067ac665\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.716116 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-log-socket\") pod \"d340687c-faf0-48dc-a1ac-dec5067ac665\" (UID: \"d340687c-faf0-48dc-a1ac-dec5067ac665\") " Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.716217 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ed6a5bc3-df22-4779-b20b-df0b86e06d68-systemd-units\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.716247 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ed6a5bc3-df22-4779-b20b-df0b86e06d68-host-kubelet\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.716269 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ed6a5bc3-df22-4779-b20b-df0b86e06d68-log-socket\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.716298 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ed6a5bc3-df22-4779-b20b-df0b86e06d68-etc-openvswitch\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.716322 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ed6a5bc3-df22-4779-b20b-df0b86e06d68-run-ovn\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.716344 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ed6a5bc3-df22-4779-b20b-df0b86e06d68-host-cni-netd\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.716380 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ed6a5bc3-df22-4779-b20b-df0b86e06d68-node-log\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.716401 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ed6a5bc3-df22-4779-b20b-df0b86e06d68-ovn-node-metrics-cert\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.716434 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ed6a5bc3-df22-4779-b20b-df0b86e06d68-host-slash\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.716460 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ed6a5bc3-df22-4779-b20b-df0b86e06d68-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.716475 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ed6a5bc3-df22-4779-b20b-df0b86e06d68-run-openvswitch\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.716489 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2drh2\" (UniqueName: \"kubernetes.io/projected/ed6a5bc3-df22-4779-b20b-df0b86e06d68-kube-api-access-2drh2\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.716504 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ed6a5bc3-df22-4779-b20b-df0b86e06d68-host-run-netns\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.716529 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ed6a5bc3-df22-4779-b20b-df0b86e06d68-host-cni-bin\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.716542 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ed6a5bc3-df22-4779-b20b-df0b86e06d68-ovnkube-script-lib\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.716556 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ed6a5bc3-df22-4779-b20b-df0b86e06d68-ovnkube-config\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.716572 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ed6a5bc3-df22-4779-b20b-df0b86e06d68-run-systemd\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.716619 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ed6a5bc3-df22-4779-b20b-df0b86e06d68-host-run-ovn-kubernetes\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.716635 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ed6a5bc3-df22-4779-b20b-df0b86e06d68-env-overrides\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.716652 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ed6a5bc3-df22-4779-b20b-df0b86e06d68-var-lib-openvswitch\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.716685 4943 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-host-slash\") on node \"crc\" DevicePath \"\"" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.716759 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "d340687c-faf0-48dc-a1ac-dec5067ac665" (UID: "d340687c-faf0-48dc-a1ac-dec5067ac665"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.716822 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "d340687c-faf0-48dc-a1ac-dec5067ac665" (UID: "d340687c-faf0-48dc-a1ac-dec5067ac665"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.716848 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "d340687c-faf0-48dc-a1ac-dec5067ac665" (UID: "d340687c-faf0-48dc-a1ac-dec5067ac665"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.716961 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "d340687c-faf0-48dc-a1ac-dec5067ac665" (UID: "d340687c-faf0-48dc-a1ac-dec5067ac665"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.716981 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "d340687c-faf0-48dc-a1ac-dec5067ac665" (UID: "d340687c-faf0-48dc-a1ac-dec5067ac665"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.716998 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "d340687c-faf0-48dc-a1ac-dec5067ac665" (UID: "d340687c-faf0-48dc-a1ac-dec5067ac665"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.717017 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "d340687c-faf0-48dc-a1ac-dec5067ac665" (UID: "d340687c-faf0-48dc-a1ac-dec5067ac665"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.717033 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "d340687c-faf0-48dc-a1ac-dec5067ac665" (UID: "d340687c-faf0-48dc-a1ac-dec5067ac665"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.717049 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-node-log" (OuterVolumeSpecName: "node-log") pod "d340687c-faf0-48dc-a1ac-dec5067ac665" (UID: "d340687c-faf0-48dc-a1ac-dec5067ac665"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.717064 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "d340687c-faf0-48dc-a1ac-dec5067ac665" (UID: "d340687c-faf0-48dc-a1ac-dec5067ac665"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.717083 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-log-socket" (OuterVolumeSpecName: "log-socket") pod "d340687c-faf0-48dc-a1ac-dec5067ac665" (UID: "d340687c-faf0-48dc-a1ac-dec5067ac665"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.717099 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "d340687c-faf0-48dc-a1ac-dec5067ac665" (UID: "d340687c-faf0-48dc-a1ac-dec5067ac665"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.721042 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d340687c-faf0-48dc-a1ac-dec5067ac665-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "d340687c-faf0-48dc-a1ac-dec5067ac665" (UID: "d340687c-faf0-48dc-a1ac-dec5067ac665"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.721230 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "d340687c-faf0-48dc-a1ac-dec5067ac665" (UID: "d340687c-faf0-48dc-a1ac-dec5067ac665"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.721472 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d340687c-faf0-48dc-a1ac-dec5067ac665-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "d340687c-faf0-48dc-a1ac-dec5067ac665" (UID: "d340687c-faf0-48dc-a1ac-dec5067ac665"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.722498 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d340687c-faf0-48dc-a1ac-dec5067ac665-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "d340687c-faf0-48dc-a1ac-dec5067ac665" (UID: "d340687c-faf0-48dc-a1ac-dec5067ac665"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.726067 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d340687c-faf0-48dc-a1ac-dec5067ac665-kube-api-access-p7d2m" (OuterVolumeSpecName: "kube-api-access-p7d2m") pod "d340687c-faf0-48dc-a1ac-dec5067ac665" (UID: "d340687c-faf0-48dc-a1ac-dec5067ac665"). InnerVolumeSpecName "kube-api-access-p7d2m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.728221 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d340687c-faf0-48dc-a1ac-dec5067ac665-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "d340687c-faf0-48dc-a1ac-dec5067ac665" (UID: "d340687c-faf0-48dc-a1ac-dec5067ac665"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.746880 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "d340687c-faf0-48dc-a1ac-dec5067ac665" (UID: "d340687c-faf0-48dc-a1ac-dec5067ac665"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.818019 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ed6a5bc3-df22-4779-b20b-df0b86e06d68-log-socket\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.818117 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ed6a5bc3-df22-4779-b20b-df0b86e06d68-etc-openvswitch\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.818170 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ed6a5bc3-df22-4779-b20b-df0b86e06d68-run-ovn\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.818189 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ed6a5bc3-df22-4779-b20b-df0b86e06d68-log-socket\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.818202 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ed6a5bc3-df22-4779-b20b-df0b86e06d68-node-log\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.818260 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ed6a5bc3-df22-4779-b20b-df0b86e06d68-node-log\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.818263 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ed6a5bc3-df22-4779-b20b-df0b86e06d68-etc-openvswitch\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.818305 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ed6a5bc3-df22-4779-b20b-df0b86e06d68-host-cni-netd\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.818376 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ed6a5bc3-df22-4779-b20b-df0b86e06d68-ovn-node-metrics-cert\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.818397 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ed6a5bc3-df22-4779-b20b-df0b86e06d68-host-cni-netd\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.818404 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ed6a5bc3-df22-4779-b20b-df0b86e06d68-run-ovn\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.818461 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ed6a5bc3-df22-4779-b20b-df0b86e06d68-host-slash\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.818438 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ed6a5bc3-df22-4779-b20b-df0b86e06d68-host-slash\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.818619 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ed6a5bc3-df22-4779-b20b-df0b86e06d68-run-openvswitch\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.818673 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ed6a5bc3-df22-4779-b20b-df0b86e06d68-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.818717 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2drh2\" (UniqueName: \"kubernetes.io/projected/ed6a5bc3-df22-4779-b20b-df0b86e06d68-kube-api-access-2drh2\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.818759 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ed6a5bc3-df22-4779-b20b-df0b86e06d68-host-run-netns\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.818825 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ed6a5bc3-df22-4779-b20b-df0b86e06d68-run-openvswitch\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.818875 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ed6a5bc3-df22-4779-b20b-df0b86e06d68-host-cni-bin\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.818897 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ed6a5bc3-df22-4779-b20b-df0b86e06d68-host-run-netns\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.818845 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ed6a5bc3-df22-4779-b20b-df0b86e06d68-host-cni-bin\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.819002 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ed6a5bc3-df22-4779-b20b-df0b86e06d68-ovnkube-script-lib\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.819037 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ed6a5bc3-df22-4779-b20b-df0b86e06d68-ovnkube-config\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.819077 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ed6a5bc3-df22-4779-b20b-df0b86e06d68-run-systemd\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.819121 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ed6a5bc3-df22-4779-b20b-df0b86e06d68-host-run-ovn-kubernetes\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.819155 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ed6a5bc3-df22-4779-b20b-df0b86e06d68-env-overrides\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.819214 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ed6a5bc3-df22-4779-b20b-df0b86e06d68-var-lib-openvswitch\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.819275 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ed6a5bc3-df22-4779-b20b-df0b86e06d68-systemd-units\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.819290 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ed6a5bc3-df22-4779-b20b-df0b86e06d68-host-run-ovn-kubernetes\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.819333 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ed6a5bc3-df22-4779-b20b-df0b86e06d68-host-kubelet\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.819364 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ed6a5bc3-df22-4779-b20b-df0b86e06d68-var-lib-openvswitch\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.819421 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ed6a5bc3-df22-4779-b20b-df0b86e06d68-host-kubelet\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.819449 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ed6a5bc3-df22-4779-b20b-df0b86e06d68-systemd-units\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.819346 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ed6a5bc3-df22-4779-b20b-df0b86e06d68-run-systemd\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.819509 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p7d2m\" (UniqueName: \"kubernetes.io/projected/d340687c-faf0-48dc-a1ac-dec5067ac665-kube-api-access-p7d2m\") on node \"crc\" DevicePath \"\"" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.819521 4943 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-log-socket\") on node \"crc\" DevicePath \"\"" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.819532 4943 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d340687c-faf0-48dc-a1ac-dec5067ac665-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.819562 4943 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d340687c-faf0-48dc-a1ac-dec5067ac665-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.819571 4943 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d340687c-faf0-48dc-a1ac-dec5067ac665-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.819581 4943 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-host-kubelet\") on node \"crc\" DevicePath \"\"" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.819590 4943 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d340687c-faf0-48dc-a1ac-dec5067ac665-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.819598 4943 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-systemd-units\") on node \"crc\" DevicePath \"\"" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.819606 4943 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-run-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.819615 4943 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-host-cni-bin\") on node \"crc\" DevicePath \"\"" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.819643 4943 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-run-systemd\") on node \"crc\" DevicePath \"\"" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.819651 4943 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-host-run-netns\") on node \"crc\" DevicePath \"\"" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.819661 4943 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.819670 4943 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.819679 4943 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.819687 4943 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-node-log\") on node \"crc\" DevicePath \"\"" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.819695 4943 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-host-cni-netd\") on node \"crc\" DevicePath \"\"" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.819725 4943 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.819734 4943 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d340687c-faf0-48dc-a1ac-dec5067ac665-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.819514 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ed6a5bc3-df22-4779-b20b-df0b86e06d68-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.820248 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ed6a5bc3-df22-4779-b20b-df0b86e06d68-ovnkube-config\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.820387 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ed6a5bc3-df22-4779-b20b-df0b86e06d68-env-overrides\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.820748 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ed6a5bc3-df22-4779-b20b-df0b86e06d68-ovnkube-script-lib\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.824636 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ed6a5bc3-df22-4779-b20b-df0b86e06d68-ovn-node-metrics-cert\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:26 crc kubenswrapper[4943]: I0307 14:50:26.838279 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2drh2\" (UniqueName: \"kubernetes.io/projected/ed6a5bc3-df22-4779-b20b-df0b86e06d68-kube-api-access-2drh2\") pod \"ovnkube-node-npgnc\" (UID: \"ed6a5bc3-df22-4779-b20b-df0b86e06d68\") " pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.012534 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:27 crc kubenswrapper[4943]: W0307 14:50:27.045674 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poded6a5bc3_df22_4779_b20b_df0b86e06d68.slice/crio-cf067bf8c8e500fd453d5d17a6e3e09d57918f45b86385e64509d2e8b92d1c5f WatchSource:0}: Error finding container cf067bf8c8e500fd453d5d17a6e3e09d57918f45b86385e64509d2e8b92d1c5f: Status 404 returned error can't find the container with id cf067bf8c8e500fd453d5d17a6e3e09d57918f45b86385e64509d2e8b92d1c5f Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.300126 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-qrlpk_a13f1d2e-9974-4c40-85f5-d876d5644993/kube-multus/0.log" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.300592 4943 generic.go:334] "Generic (PLEG): container finished" podID="a13f1d2e-9974-4c40-85f5-d876d5644993" containerID="ea59bca2daa4e410ff69f16a6b5d802dd7375344ebb97d3d5eda572e70bb2db1" exitCode=2 Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.300677 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-qrlpk" event={"ID":"a13f1d2e-9974-4c40-85f5-d876d5644993","Type":"ContainerDied","Data":"ea59bca2daa4e410ff69f16a6b5d802dd7375344ebb97d3d5eda572e70bb2db1"} Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.301392 4943 scope.go:117] "RemoveContainer" containerID="ea59bca2daa4e410ff69f16a6b5d802dd7375344ebb97d3d5eda572e70bb2db1" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.308616 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-s8cw7_d340687c-faf0-48dc-a1ac-dec5067ac665/ovn-acl-logging/0.log" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.309255 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-s8cw7_d340687c-faf0-48dc-a1ac-dec5067ac665/ovn-controller/0.log" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.310659 4943 generic.go:334] "Generic (PLEG): container finished" podID="d340687c-faf0-48dc-a1ac-dec5067ac665" containerID="a200f68bc20411f516725f81a930664ba5baaefd21a2792866262ded7e86fae5" exitCode=0 Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.310705 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.310737 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" event={"ID":"d340687c-faf0-48dc-a1ac-dec5067ac665","Type":"ContainerDied","Data":"a200f68bc20411f516725f81a930664ba5baaefd21a2792866262ded7e86fae5"} Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.310802 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" event={"ID":"d340687c-faf0-48dc-a1ac-dec5067ac665","Type":"ContainerDied","Data":"f56c9c195453268c830bf92a70a64f70e75ef30e8ca853bb1f3a054a69c791af"} Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.310828 4943 scope.go:117] "RemoveContainer" containerID="a200f68bc20411f516725f81a930664ba5baaefd21a2792866262ded7e86fae5" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.310708 4943 generic.go:334] "Generic (PLEG): container finished" podID="d340687c-faf0-48dc-a1ac-dec5067ac665" containerID="f56c9c195453268c830bf92a70a64f70e75ef30e8ca853bb1f3a054a69c791af" exitCode=0 Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.310856 4943 generic.go:334] "Generic (PLEG): container finished" podID="d340687c-faf0-48dc-a1ac-dec5067ac665" containerID="50e79fc40569f5b3fd73286933ee0242a1edd807a820edf7ba00a856d81c82c9" exitCode=0 Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.310872 4943 generic.go:334] "Generic (PLEG): container finished" podID="d340687c-faf0-48dc-a1ac-dec5067ac665" containerID="959458a05302e835122b9db5d1eb4c449749a9e7ba2d85459edaeef2eede12de" exitCode=0 Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.310890 4943 generic.go:334] "Generic (PLEG): container finished" podID="d340687c-faf0-48dc-a1ac-dec5067ac665" containerID="a64d887ab03c882e1a27a4cf36f359a8d18e3c75b0a6d81f33569e1e3b294280" exitCode=0 Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.310973 4943 generic.go:334] "Generic (PLEG): container finished" podID="d340687c-faf0-48dc-a1ac-dec5067ac665" containerID="6fb3f1769f0d1092192c2f5f152c7d1fc52813d2e722687fb3ea1f6f08fba808" exitCode=0 Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.311327 4943 generic.go:334] "Generic (PLEG): container finished" podID="d340687c-faf0-48dc-a1ac-dec5067ac665" containerID="4dd5c06470991a2bcdddf5a5e429adcf4f4af8d3742de9c0a1e9cdd5705f74c5" exitCode=143 Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.311354 4943 generic.go:334] "Generic (PLEG): container finished" podID="d340687c-faf0-48dc-a1ac-dec5067ac665" containerID="6a089f76e974f8c12f32f42b96b301fda03651276c599841b9c42fa064ffd5fd" exitCode=143 Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.311004 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" event={"ID":"d340687c-faf0-48dc-a1ac-dec5067ac665","Type":"ContainerDied","Data":"50e79fc40569f5b3fd73286933ee0242a1edd807a820edf7ba00a856d81c82c9"} Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.311484 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" event={"ID":"d340687c-faf0-48dc-a1ac-dec5067ac665","Type":"ContainerDied","Data":"959458a05302e835122b9db5d1eb4c449749a9e7ba2d85459edaeef2eede12de"} Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.311520 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" event={"ID":"d340687c-faf0-48dc-a1ac-dec5067ac665","Type":"ContainerDied","Data":"a64d887ab03c882e1a27a4cf36f359a8d18e3c75b0a6d81f33569e1e3b294280"} Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.311546 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" event={"ID":"d340687c-faf0-48dc-a1ac-dec5067ac665","Type":"ContainerDied","Data":"6fb3f1769f0d1092192c2f5f152c7d1fc52813d2e722687fb3ea1f6f08fba808"} Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.311574 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4dd5c06470991a2bcdddf5a5e429adcf4f4af8d3742de9c0a1e9cdd5705f74c5"} Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.311594 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6a089f76e974f8c12f32f42b96b301fda03651276c599841b9c42fa064ffd5fd"} Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.311607 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ce64b91eda434340bdacc6a953afe291459d3b18a9cbf842dc8d81723fcdd70e"} Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.311624 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" event={"ID":"d340687c-faf0-48dc-a1ac-dec5067ac665","Type":"ContainerDied","Data":"4dd5c06470991a2bcdddf5a5e429adcf4f4af8d3742de9c0a1e9cdd5705f74c5"} Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.311642 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a200f68bc20411f516725f81a930664ba5baaefd21a2792866262ded7e86fae5"} Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.311655 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f56c9c195453268c830bf92a70a64f70e75ef30e8ca853bb1f3a054a69c791af"} Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.311668 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"50e79fc40569f5b3fd73286933ee0242a1edd807a820edf7ba00a856d81c82c9"} Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.311679 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"959458a05302e835122b9db5d1eb4c449749a9e7ba2d85459edaeef2eede12de"} Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.311691 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a64d887ab03c882e1a27a4cf36f359a8d18e3c75b0a6d81f33569e1e3b294280"} Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.311705 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6fb3f1769f0d1092192c2f5f152c7d1fc52813d2e722687fb3ea1f6f08fba808"} Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.311716 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4dd5c06470991a2bcdddf5a5e429adcf4f4af8d3742de9c0a1e9cdd5705f74c5"} Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.311726 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6a089f76e974f8c12f32f42b96b301fda03651276c599841b9c42fa064ffd5fd"} Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.311737 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ce64b91eda434340bdacc6a953afe291459d3b18a9cbf842dc8d81723fcdd70e"} Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.311751 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" event={"ID":"d340687c-faf0-48dc-a1ac-dec5067ac665","Type":"ContainerDied","Data":"6a089f76e974f8c12f32f42b96b301fda03651276c599841b9c42fa064ffd5fd"} Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.311767 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a200f68bc20411f516725f81a930664ba5baaefd21a2792866262ded7e86fae5"} Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.311779 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f56c9c195453268c830bf92a70a64f70e75ef30e8ca853bb1f3a054a69c791af"} Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.311790 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"50e79fc40569f5b3fd73286933ee0242a1edd807a820edf7ba00a856d81c82c9"} Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.311800 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"959458a05302e835122b9db5d1eb4c449749a9e7ba2d85459edaeef2eede12de"} Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.311811 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a64d887ab03c882e1a27a4cf36f359a8d18e3c75b0a6d81f33569e1e3b294280"} Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.311822 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6fb3f1769f0d1092192c2f5f152c7d1fc52813d2e722687fb3ea1f6f08fba808"} Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.311832 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4dd5c06470991a2bcdddf5a5e429adcf4f4af8d3742de9c0a1e9cdd5705f74c5"} Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.311843 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6a089f76e974f8c12f32f42b96b301fda03651276c599841b9c42fa064ffd5fd"} Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.311853 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ce64b91eda434340bdacc6a953afe291459d3b18a9cbf842dc8d81723fcdd70e"} Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.311867 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8cw7" event={"ID":"d340687c-faf0-48dc-a1ac-dec5067ac665","Type":"ContainerDied","Data":"15fafcc9303d61bcfb6c9516eb858fa860668edc2f434892f764a3af40ac36ed"} Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.311883 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a200f68bc20411f516725f81a930664ba5baaefd21a2792866262ded7e86fae5"} Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.311896 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f56c9c195453268c830bf92a70a64f70e75ef30e8ca853bb1f3a054a69c791af"} Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.311907 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"50e79fc40569f5b3fd73286933ee0242a1edd807a820edf7ba00a856d81c82c9"} Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.311919 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"959458a05302e835122b9db5d1eb4c449749a9e7ba2d85459edaeef2eede12de"} Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.311966 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a64d887ab03c882e1a27a4cf36f359a8d18e3c75b0a6d81f33569e1e3b294280"} Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.311977 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6fb3f1769f0d1092192c2f5f152c7d1fc52813d2e722687fb3ea1f6f08fba808"} Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.311989 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4dd5c06470991a2bcdddf5a5e429adcf4f4af8d3742de9c0a1e9cdd5705f74c5"} Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.312000 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6a089f76e974f8c12f32f42b96b301fda03651276c599841b9c42fa064ffd5fd"} Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.312010 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ce64b91eda434340bdacc6a953afe291459d3b18a9cbf842dc8d81723fcdd70e"} Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.315079 4943 generic.go:334] "Generic (PLEG): container finished" podID="ed6a5bc3-df22-4779-b20b-df0b86e06d68" containerID="729eebd4701a16385745cce0c26c9ebe87023fb6f9b41a2086e70117caf653c7" exitCode=0 Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.315116 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" event={"ID":"ed6a5bc3-df22-4779-b20b-df0b86e06d68","Type":"ContainerDied","Data":"729eebd4701a16385745cce0c26c9ebe87023fb6f9b41a2086e70117caf653c7"} Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.315144 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" event={"ID":"ed6a5bc3-df22-4779-b20b-df0b86e06d68","Type":"ContainerStarted","Data":"cf067bf8c8e500fd453d5d17a6e3e09d57918f45b86385e64509d2e8b92d1c5f"} Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.348577 4943 scope.go:117] "RemoveContainer" containerID="f56c9c195453268c830bf92a70a64f70e75ef30e8ca853bb1f3a054a69c791af" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.387543 4943 scope.go:117] "RemoveContainer" containerID="50e79fc40569f5b3fd73286933ee0242a1edd807a820edf7ba00a856d81c82c9" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.409999 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-s8cw7"] Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.410226 4943 scope.go:117] "RemoveContainer" containerID="959458a05302e835122b9db5d1eb4c449749a9e7ba2d85459edaeef2eede12de" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.415127 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-s8cw7"] Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.484392 4943 scope.go:117] "RemoveContainer" containerID="a64d887ab03c882e1a27a4cf36f359a8d18e3c75b0a6d81f33569e1e3b294280" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.502137 4943 scope.go:117] "RemoveContainer" containerID="6fb3f1769f0d1092192c2f5f152c7d1fc52813d2e722687fb3ea1f6f08fba808" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.524053 4943 scope.go:117] "RemoveContainer" containerID="4dd5c06470991a2bcdddf5a5e429adcf4f4af8d3742de9c0a1e9cdd5705f74c5" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.540682 4943 scope.go:117] "RemoveContainer" containerID="6a089f76e974f8c12f32f42b96b301fda03651276c599841b9c42fa064ffd5fd" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.560593 4943 scope.go:117] "RemoveContainer" containerID="ce64b91eda434340bdacc6a953afe291459d3b18a9cbf842dc8d81723fcdd70e" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.578605 4943 scope.go:117] "RemoveContainer" containerID="a200f68bc20411f516725f81a930664ba5baaefd21a2792866262ded7e86fae5" Mar 07 14:50:27 crc kubenswrapper[4943]: E0307 14:50:27.579089 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a200f68bc20411f516725f81a930664ba5baaefd21a2792866262ded7e86fae5\": container with ID starting with a200f68bc20411f516725f81a930664ba5baaefd21a2792866262ded7e86fae5 not found: ID does not exist" containerID="a200f68bc20411f516725f81a930664ba5baaefd21a2792866262ded7e86fae5" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.579129 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a200f68bc20411f516725f81a930664ba5baaefd21a2792866262ded7e86fae5"} err="failed to get container status \"a200f68bc20411f516725f81a930664ba5baaefd21a2792866262ded7e86fae5\": rpc error: code = NotFound desc = could not find container \"a200f68bc20411f516725f81a930664ba5baaefd21a2792866262ded7e86fae5\": container with ID starting with a200f68bc20411f516725f81a930664ba5baaefd21a2792866262ded7e86fae5 not found: ID does not exist" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.579162 4943 scope.go:117] "RemoveContainer" containerID="f56c9c195453268c830bf92a70a64f70e75ef30e8ca853bb1f3a054a69c791af" Mar 07 14:50:27 crc kubenswrapper[4943]: E0307 14:50:27.579434 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f56c9c195453268c830bf92a70a64f70e75ef30e8ca853bb1f3a054a69c791af\": container with ID starting with f56c9c195453268c830bf92a70a64f70e75ef30e8ca853bb1f3a054a69c791af not found: ID does not exist" containerID="f56c9c195453268c830bf92a70a64f70e75ef30e8ca853bb1f3a054a69c791af" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.579459 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f56c9c195453268c830bf92a70a64f70e75ef30e8ca853bb1f3a054a69c791af"} err="failed to get container status \"f56c9c195453268c830bf92a70a64f70e75ef30e8ca853bb1f3a054a69c791af\": rpc error: code = NotFound desc = could not find container \"f56c9c195453268c830bf92a70a64f70e75ef30e8ca853bb1f3a054a69c791af\": container with ID starting with f56c9c195453268c830bf92a70a64f70e75ef30e8ca853bb1f3a054a69c791af not found: ID does not exist" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.579476 4943 scope.go:117] "RemoveContainer" containerID="50e79fc40569f5b3fd73286933ee0242a1edd807a820edf7ba00a856d81c82c9" Mar 07 14:50:27 crc kubenswrapper[4943]: E0307 14:50:27.580913 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50e79fc40569f5b3fd73286933ee0242a1edd807a820edf7ba00a856d81c82c9\": container with ID starting with 50e79fc40569f5b3fd73286933ee0242a1edd807a820edf7ba00a856d81c82c9 not found: ID does not exist" containerID="50e79fc40569f5b3fd73286933ee0242a1edd807a820edf7ba00a856d81c82c9" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.580997 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50e79fc40569f5b3fd73286933ee0242a1edd807a820edf7ba00a856d81c82c9"} err="failed to get container status \"50e79fc40569f5b3fd73286933ee0242a1edd807a820edf7ba00a856d81c82c9\": rpc error: code = NotFound desc = could not find container \"50e79fc40569f5b3fd73286933ee0242a1edd807a820edf7ba00a856d81c82c9\": container with ID starting with 50e79fc40569f5b3fd73286933ee0242a1edd807a820edf7ba00a856d81c82c9 not found: ID does not exist" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.581019 4943 scope.go:117] "RemoveContainer" containerID="959458a05302e835122b9db5d1eb4c449749a9e7ba2d85459edaeef2eede12de" Mar 07 14:50:27 crc kubenswrapper[4943]: E0307 14:50:27.581510 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"959458a05302e835122b9db5d1eb4c449749a9e7ba2d85459edaeef2eede12de\": container with ID starting with 959458a05302e835122b9db5d1eb4c449749a9e7ba2d85459edaeef2eede12de not found: ID does not exist" containerID="959458a05302e835122b9db5d1eb4c449749a9e7ba2d85459edaeef2eede12de" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.581562 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"959458a05302e835122b9db5d1eb4c449749a9e7ba2d85459edaeef2eede12de"} err="failed to get container status \"959458a05302e835122b9db5d1eb4c449749a9e7ba2d85459edaeef2eede12de\": rpc error: code = NotFound desc = could not find container \"959458a05302e835122b9db5d1eb4c449749a9e7ba2d85459edaeef2eede12de\": container with ID starting with 959458a05302e835122b9db5d1eb4c449749a9e7ba2d85459edaeef2eede12de not found: ID does not exist" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.581581 4943 scope.go:117] "RemoveContainer" containerID="a64d887ab03c882e1a27a4cf36f359a8d18e3c75b0a6d81f33569e1e3b294280" Mar 07 14:50:27 crc kubenswrapper[4943]: E0307 14:50:27.581948 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a64d887ab03c882e1a27a4cf36f359a8d18e3c75b0a6d81f33569e1e3b294280\": container with ID starting with a64d887ab03c882e1a27a4cf36f359a8d18e3c75b0a6d81f33569e1e3b294280 not found: ID does not exist" containerID="a64d887ab03c882e1a27a4cf36f359a8d18e3c75b0a6d81f33569e1e3b294280" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.581980 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a64d887ab03c882e1a27a4cf36f359a8d18e3c75b0a6d81f33569e1e3b294280"} err="failed to get container status \"a64d887ab03c882e1a27a4cf36f359a8d18e3c75b0a6d81f33569e1e3b294280\": rpc error: code = NotFound desc = could not find container \"a64d887ab03c882e1a27a4cf36f359a8d18e3c75b0a6d81f33569e1e3b294280\": container with ID starting with a64d887ab03c882e1a27a4cf36f359a8d18e3c75b0a6d81f33569e1e3b294280 not found: ID does not exist" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.582002 4943 scope.go:117] "RemoveContainer" containerID="6fb3f1769f0d1092192c2f5f152c7d1fc52813d2e722687fb3ea1f6f08fba808" Mar 07 14:50:27 crc kubenswrapper[4943]: E0307 14:50:27.582248 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6fb3f1769f0d1092192c2f5f152c7d1fc52813d2e722687fb3ea1f6f08fba808\": container with ID starting with 6fb3f1769f0d1092192c2f5f152c7d1fc52813d2e722687fb3ea1f6f08fba808 not found: ID does not exist" containerID="6fb3f1769f0d1092192c2f5f152c7d1fc52813d2e722687fb3ea1f6f08fba808" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.582274 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fb3f1769f0d1092192c2f5f152c7d1fc52813d2e722687fb3ea1f6f08fba808"} err="failed to get container status \"6fb3f1769f0d1092192c2f5f152c7d1fc52813d2e722687fb3ea1f6f08fba808\": rpc error: code = NotFound desc = could not find container \"6fb3f1769f0d1092192c2f5f152c7d1fc52813d2e722687fb3ea1f6f08fba808\": container with ID starting with 6fb3f1769f0d1092192c2f5f152c7d1fc52813d2e722687fb3ea1f6f08fba808 not found: ID does not exist" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.582290 4943 scope.go:117] "RemoveContainer" containerID="4dd5c06470991a2bcdddf5a5e429adcf4f4af8d3742de9c0a1e9cdd5705f74c5" Mar 07 14:50:27 crc kubenswrapper[4943]: E0307 14:50:27.582547 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4dd5c06470991a2bcdddf5a5e429adcf4f4af8d3742de9c0a1e9cdd5705f74c5\": container with ID starting with 4dd5c06470991a2bcdddf5a5e429adcf4f4af8d3742de9c0a1e9cdd5705f74c5 not found: ID does not exist" containerID="4dd5c06470991a2bcdddf5a5e429adcf4f4af8d3742de9c0a1e9cdd5705f74c5" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.582573 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4dd5c06470991a2bcdddf5a5e429adcf4f4af8d3742de9c0a1e9cdd5705f74c5"} err="failed to get container status \"4dd5c06470991a2bcdddf5a5e429adcf4f4af8d3742de9c0a1e9cdd5705f74c5\": rpc error: code = NotFound desc = could not find container \"4dd5c06470991a2bcdddf5a5e429adcf4f4af8d3742de9c0a1e9cdd5705f74c5\": container with ID starting with 4dd5c06470991a2bcdddf5a5e429adcf4f4af8d3742de9c0a1e9cdd5705f74c5 not found: ID does not exist" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.582593 4943 scope.go:117] "RemoveContainer" containerID="6a089f76e974f8c12f32f42b96b301fda03651276c599841b9c42fa064ffd5fd" Mar 07 14:50:27 crc kubenswrapper[4943]: E0307 14:50:27.583161 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a089f76e974f8c12f32f42b96b301fda03651276c599841b9c42fa064ffd5fd\": container with ID starting with 6a089f76e974f8c12f32f42b96b301fda03651276c599841b9c42fa064ffd5fd not found: ID does not exist" containerID="6a089f76e974f8c12f32f42b96b301fda03651276c599841b9c42fa064ffd5fd" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.583215 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a089f76e974f8c12f32f42b96b301fda03651276c599841b9c42fa064ffd5fd"} err="failed to get container status \"6a089f76e974f8c12f32f42b96b301fda03651276c599841b9c42fa064ffd5fd\": rpc error: code = NotFound desc = could not find container \"6a089f76e974f8c12f32f42b96b301fda03651276c599841b9c42fa064ffd5fd\": container with ID starting with 6a089f76e974f8c12f32f42b96b301fda03651276c599841b9c42fa064ffd5fd not found: ID does not exist" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.583233 4943 scope.go:117] "RemoveContainer" containerID="ce64b91eda434340bdacc6a953afe291459d3b18a9cbf842dc8d81723fcdd70e" Mar 07 14:50:27 crc kubenswrapper[4943]: E0307 14:50:27.583533 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce64b91eda434340bdacc6a953afe291459d3b18a9cbf842dc8d81723fcdd70e\": container with ID starting with ce64b91eda434340bdacc6a953afe291459d3b18a9cbf842dc8d81723fcdd70e not found: ID does not exist" containerID="ce64b91eda434340bdacc6a953afe291459d3b18a9cbf842dc8d81723fcdd70e" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.583588 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce64b91eda434340bdacc6a953afe291459d3b18a9cbf842dc8d81723fcdd70e"} err="failed to get container status \"ce64b91eda434340bdacc6a953afe291459d3b18a9cbf842dc8d81723fcdd70e\": rpc error: code = NotFound desc = could not find container \"ce64b91eda434340bdacc6a953afe291459d3b18a9cbf842dc8d81723fcdd70e\": container with ID starting with ce64b91eda434340bdacc6a953afe291459d3b18a9cbf842dc8d81723fcdd70e not found: ID does not exist" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.583605 4943 scope.go:117] "RemoveContainer" containerID="a200f68bc20411f516725f81a930664ba5baaefd21a2792866262ded7e86fae5" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.584112 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a200f68bc20411f516725f81a930664ba5baaefd21a2792866262ded7e86fae5"} err="failed to get container status \"a200f68bc20411f516725f81a930664ba5baaefd21a2792866262ded7e86fae5\": rpc error: code = NotFound desc = could not find container \"a200f68bc20411f516725f81a930664ba5baaefd21a2792866262ded7e86fae5\": container with ID starting with a200f68bc20411f516725f81a930664ba5baaefd21a2792866262ded7e86fae5 not found: ID does not exist" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.584137 4943 scope.go:117] "RemoveContainer" containerID="f56c9c195453268c830bf92a70a64f70e75ef30e8ca853bb1f3a054a69c791af" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.584396 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f56c9c195453268c830bf92a70a64f70e75ef30e8ca853bb1f3a054a69c791af"} err="failed to get container status \"f56c9c195453268c830bf92a70a64f70e75ef30e8ca853bb1f3a054a69c791af\": rpc error: code = NotFound desc = could not find container \"f56c9c195453268c830bf92a70a64f70e75ef30e8ca853bb1f3a054a69c791af\": container with ID starting with f56c9c195453268c830bf92a70a64f70e75ef30e8ca853bb1f3a054a69c791af not found: ID does not exist" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.584420 4943 scope.go:117] "RemoveContainer" containerID="50e79fc40569f5b3fd73286933ee0242a1edd807a820edf7ba00a856d81c82c9" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.584709 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50e79fc40569f5b3fd73286933ee0242a1edd807a820edf7ba00a856d81c82c9"} err="failed to get container status \"50e79fc40569f5b3fd73286933ee0242a1edd807a820edf7ba00a856d81c82c9\": rpc error: code = NotFound desc = could not find container \"50e79fc40569f5b3fd73286933ee0242a1edd807a820edf7ba00a856d81c82c9\": container with ID starting with 50e79fc40569f5b3fd73286933ee0242a1edd807a820edf7ba00a856d81c82c9 not found: ID does not exist" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.584743 4943 scope.go:117] "RemoveContainer" containerID="959458a05302e835122b9db5d1eb4c449749a9e7ba2d85459edaeef2eede12de" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.585105 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"959458a05302e835122b9db5d1eb4c449749a9e7ba2d85459edaeef2eede12de"} err="failed to get container status \"959458a05302e835122b9db5d1eb4c449749a9e7ba2d85459edaeef2eede12de\": rpc error: code = NotFound desc = could not find container \"959458a05302e835122b9db5d1eb4c449749a9e7ba2d85459edaeef2eede12de\": container with ID starting with 959458a05302e835122b9db5d1eb4c449749a9e7ba2d85459edaeef2eede12de not found: ID does not exist" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.585139 4943 scope.go:117] "RemoveContainer" containerID="a64d887ab03c882e1a27a4cf36f359a8d18e3c75b0a6d81f33569e1e3b294280" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.586035 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a64d887ab03c882e1a27a4cf36f359a8d18e3c75b0a6d81f33569e1e3b294280"} err="failed to get container status \"a64d887ab03c882e1a27a4cf36f359a8d18e3c75b0a6d81f33569e1e3b294280\": rpc error: code = NotFound desc = could not find container \"a64d887ab03c882e1a27a4cf36f359a8d18e3c75b0a6d81f33569e1e3b294280\": container with ID starting with a64d887ab03c882e1a27a4cf36f359a8d18e3c75b0a6d81f33569e1e3b294280 not found: ID does not exist" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.586055 4943 scope.go:117] "RemoveContainer" containerID="6fb3f1769f0d1092192c2f5f152c7d1fc52813d2e722687fb3ea1f6f08fba808" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.590011 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fb3f1769f0d1092192c2f5f152c7d1fc52813d2e722687fb3ea1f6f08fba808"} err="failed to get container status \"6fb3f1769f0d1092192c2f5f152c7d1fc52813d2e722687fb3ea1f6f08fba808\": rpc error: code = NotFound desc = could not find container \"6fb3f1769f0d1092192c2f5f152c7d1fc52813d2e722687fb3ea1f6f08fba808\": container with ID starting with 6fb3f1769f0d1092192c2f5f152c7d1fc52813d2e722687fb3ea1f6f08fba808 not found: ID does not exist" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.590031 4943 scope.go:117] "RemoveContainer" containerID="4dd5c06470991a2bcdddf5a5e429adcf4f4af8d3742de9c0a1e9cdd5705f74c5" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.590263 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4dd5c06470991a2bcdddf5a5e429adcf4f4af8d3742de9c0a1e9cdd5705f74c5"} err="failed to get container status \"4dd5c06470991a2bcdddf5a5e429adcf4f4af8d3742de9c0a1e9cdd5705f74c5\": rpc error: code = NotFound desc = could not find container \"4dd5c06470991a2bcdddf5a5e429adcf4f4af8d3742de9c0a1e9cdd5705f74c5\": container with ID starting with 4dd5c06470991a2bcdddf5a5e429adcf4f4af8d3742de9c0a1e9cdd5705f74c5 not found: ID does not exist" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.590280 4943 scope.go:117] "RemoveContainer" containerID="6a089f76e974f8c12f32f42b96b301fda03651276c599841b9c42fa064ffd5fd" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.590486 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a089f76e974f8c12f32f42b96b301fda03651276c599841b9c42fa064ffd5fd"} err="failed to get container status \"6a089f76e974f8c12f32f42b96b301fda03651276c599841b9c42fa064ffd5fd\": rpc error: code = NotFound desc = could not find container \"6a089f76e974f8c12f32f42b96b301fda03651276c599841b9c42fa064ffd5fd\": container with ID starting with 6a089f76e974f8c12f32f42b96b301fda03651276c599841b9c42fa064ffd5fd not found: ID does not exist" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.590501 4943 scope.go:117] "RemoveContainer" containerID="ce64b91eda434340bdacc6a953afe291459d3b18a9cbf842dc8d81723fcdd70e" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.590678 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce64b91eda434340bdacc6a953afe291459d3b18a9cbf842dc8d81723fcdd70e"} err="failed to get container status \"ce64b91eda434340bdacc6a953afe291459d3b18a9cbf842dc8d81723fcdd70e\": rpc error: code = NotFound desc = could not find container \"ce64b91eda434340bdacc6a953afe291459d3b18a9cbf842dc8d81723fcdd70e\": container with ID starting with ce64b91eda434340bdacc6a953afe291459d3b18a9cbf842dc8d81723fcdd70e not found: ID does not exist" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.590693 4943 scope.go:117] "RemoveContainer" containerID="a200f68bc20411f516725f81a930664ba5baaefd21a2792866262ded7e86fae5" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.592154 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a200f68bc20411f516725f81a930664ba5baaefd21a2792866262ded7e86fae5"} err="failed to get container status \"a200f68bc20411f516725f81a930664ba5baaefd21a2792866262ded7e86fae5\": rpc error: code = NotFound desc = could not find container \"a200f68bc20411f516725f81a930664ba5baaefd21a2792866262ded7e86fae5\": container with ID starting with a200f68bc20411f516725f81a930664ba5baaefd21a2792866262ded7e86fae5 not found: ID does not exist" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.592218 4943 scope.go:117] "RemoveContainer" containerID="f56c9c195453268c830bf92a70a64f70e75ef30e8ca853bb1f3a054a69c791af" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.592550 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f56c9c195453268c830bf92a70a64f70e75ef30e8ca853bb1f3a054a69c791af"} err="failed to get container status \"f56c9c195453268c830bf92a70a64f70e75ef30e8ca853bb1f3a054a69c791af\": rpc error: code = NotFound desc = could not find container \"f56c9c195453268c830bf92a70a64f70e75ef30e8ca853bb1f3a054a69c791af\": container with ID starting with f56c9c195453268c830bf92a70a64f70e75ef30e8ca853bb1f3a054a69c791af not found: ID does not exist" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.592571 4943 scope.go:117] "RemoveContainer" containerID="50e79fc40569f5b3fd73286933ee0242a1edd807a820edf7ba00a856d81c82c9" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.592974 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50e79fc40569f5b3fd73286933ee0242a1edd807a820edf7ba00a856d81c82c9"} err="failed to get container status \"50e79fc40569f5b3fd73286933ee0242a1edd807a820edf7ba00a856d81c82c9\": rpc error: code = NotFound desc = could not find container \"50e79fc40569f5b3fd73286933ee0242a1edd807a820edf7ba00a856d81c82c9\": container with ID starting with 50e79fc40569f5b3fd73286933ee0242a1edd807a820edf7ba00a856d81c82c9 not found: ID does not exist" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.593028 4943 scope.go:117] "RemoveContainer" containerID="959458a05302e835122b9db5d1eb4c449749a9e7ba2d85459edaeef2eede12de" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.593387 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"959458a05302e835122b9db5d1eb4c449749a9e7ba2d85459edaeef2eede12de"} err="failed to get container status \"959458a05302e835122b9db5d1eb4c449749a9e7ba2d85459edaeef2eede12de\": rpc error: code = NotFound desc = could not find container \"959458a05302e835122b9db5d1eb4c449749a9e7ba2d85459edaeef2eede12de\": container with ID starting with 959458a05302e835122b9db5d1eb4c449749a9e7ba2d85459edaeef2eede12de not found: ID does not exist" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.593411 4943 scope.go:117] "RemoveContainer" containerID="a64d887ab03c882e1a27a4cf36f359a8d18e3c75b0a6d81f33569e1e3b294280" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.593882 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a64d887ab03c882e1a27a4cf36f359a8d18e3c75b0a6d81f33569e1e3b294280"} err="failed to get container status \"a64d887ab03c882e1a27a4cf36f359a8d18e3c75b0a6d81f33569e1e3b294280\": rpc error: code = NotFound desc = could not find container \"a64d887ab03c882e1a27a4cf36f359a8d18e3c75b0a6d81f33569e1e3b294280\": container with ID starting with a64d887ab03c882e1a27a4cf36f359a8d18e3c75b0a6d81f33569e1e3b294280 not found: ID does not exist" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.594011 4943 scope.go:117] "RemoveContainer" containerID="6fb3f1769f0d1092192c2f5f152c7d1fc52813d2e722687fb3ea1f6f08fba808" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.594433 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fb3f1769f0d1092192c2f5f152c7d1fc52813d2e722687fb3ea1f6f08fba808"} err="failed to get container status \"6fb3f1769f0d1092192c2f5f152c7d1fc52813d2e722687fb3ea1f6f08fba808\": rpc error: code = NotFound desc = could not find container \"6fb3f1769f0d1092192c2f5f152c7d1fc52813d2e722687fb3ea1f6f08fba808\": container with ID starting with 6fb3f1769f0d1092192c2f5f152c7d1fc52813d2e722687fb3ea1f6f08fba808 not found: ID does not exist" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.594454 4943 scope.go:117] "RemoveContainer" containerID="4dd5c06470991a2bcdddf5a5e429adcf4f4af8d3742de9c0a1e9cdd5705f74c5" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.594735 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4dd5c06470991a2bcdddf5a5e429adcf4f4af8d3742de9c0a1e9cdd5705f74c5"} err="failed to get container status \"4dd5c06470991a2bcdddf5a5e429adcf4f4af8d3742de9c0a1e9cdd5705f74c5\": rpc error: code = NotFound desc = could not find container \"4dd5c06470991a2bcdddf5a5e429adcf4f4af8d3742de9c0a1e9cdd5705f74c5\": container with ID starting with 4dd5c06470991a2bcdddf5a5e429adcf4f4af8d3742de9c0a1e9cdd5705f74c5 not found: ID does not exist" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.594767 4943 scope.go:117] "RemoveContainer" containerID="6a089f76e974f8c12f32f42b96b301fda03651276c599841b9c42fa064ffd5fd" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.595067 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a089f76e974f8c12f32f42b96b301fda03651276c599841b9c42fa064ffd5fd"} err="failed to get container status \"6a089f76e974f8c12f32f42b96b301fda03651276c599841b9c42fa064ffd5fd\": rpc error: code = NotFound desc = could not find container \"6a089f76e974f8c12f32f42b96b301fda03651276c599841b9c42fa064ffd5fd\": container with ID starting with 6a089f76e974f8c12f32f42b96b301fda03651276c599841b9c42fa064ffd5fd not found: ID does not exist" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.595091 4943 scope.go:117] "RemoveContainer" containerID="ce64b91eda434340bdacc6a953afe291459d3b18a9cbf842dc8d81723fcdd70e" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.595319 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce64b91eda434340bdacc6a953afe291459d3b18a9cbf842dc8d81723fcdd70e"} err="failed to get container status \"ce64b91eda434340bdacc6a953afe291459d3b18a9cbf842dc8d81723fcdd70e\": rpc error: code = NotFound desc = could not find container \"ce64b91eda434340bdacc6a953afe291459d3b18a9cbf842dc8d81723fcdd70e\": container with ID starting with ce64b91eda434340bdacc6a953afe291459d3b18a9cbf842dc8d81723fcdd70e not found: ID does not exist" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.595367 4943 scope.go:117] "RemoveContainer" containerID="a200f68bc20411f516725f81a930664ba5baaefd21a2792866262ded7e86fae5" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.596178 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a200f68bc20411f516725f81a930664ba5baaefd21a2792866262ded7e86fae5"} err="failed to get container status \"a200f68bc20411f516725f81a930664ba5baaefd21a2792866262ded7e86fae5\": rpc error: code = NotFound desc = could not find container \"a200f68bc20411f516725f81a930664ba5baaefd21a2792866262ded7e86fae5\": container with ID starting with a200f68bc20411f516725f81a930664ba5baaefd21a2792866262ded7e86fae5 not found: ID does not exist" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.596208 4943 scope.go:117] "RemoveContainer" containerID="f56c9c195453268c830bf92a70a64f70e75ef30e8ca853bb1f3a054a69c791af" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.596444 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f56c9c195453268c830bf92a70a64f70e75ef30e8ca853bb1f3a054a69c791af"} err="failed to get container status \"f56c9c195453268c830bf92a70a64f70e75ef30e8ca853bb1f3a054a69c791af\": rpc error: code = NotFound desc = could not find container \"f56c9c195453268c830bf92a70a64f70e75ef30e8ca853bb1f3a054a69c791af\": container with ID starting with f56c9c195453268c830bf92a70a64f70e75ef30e8ca853bb1f3a054a69c791af not found: ID does not exist" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.596469 4943 scope.go:117] "RemoveContainer" containerID="50e79fc40569f5b3fd73286933ee0242a1edd807a820edf7ba00a856d81c82c9" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.596686 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50e79fc40569f5b3fd73286933ee0242a1edd807a820edf7ba00a856d81c82c9"} err="failed to get container status \"50e79fc40569f5b3fd73286933ee0242a1edd807a820edf7ba00a856d81c82c9\": rpc error: code = NotFound desc = could not find container \"50e79fc40569f5b3fd73286933ee0242a1edd807a820edf7ba00a856d81c82c9\": container with ID starting with 50e79fc40569f5b3fd73286933ee0242a1edd807a820edf7ba00a856d81c82c9 not found: ID does not exist" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.596713 4943 scope.go:117] "RemoveContainer" containerID="959458a05302e835122b9db5d1eb4c449749a9e7ba2d85459edaeef2eede12de" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.597102 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"959458a05302e835122b9db5d1eb4c449749a9e7ba2d85459edaeef2eede12de"} err="failed to get container status \"959458a05302e835122b9db5d1eb4c449749a9e7ba2d85459edaeef2eede12de\": rpc error: code = NotFound desc = could not find container \"959458a05302e835122b9db5d1eb4c449749a9e7ba2d85459edaeef2eede12de\": container with ID starting with 959458a05302e835122b9db5d1eb4c449749a9e7ba2d85459edaeef2eede12de not found: ID does not exist" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.597147 4943 scope.go:117] "RemoveContainer" containerID="a64d887ab03c882e1a27a4cf36f359a8d18e3c75b0a6d81f33569e1e3b294280" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.597591 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a64d887ab03c882e1a27a4cf36f359a8d18e3c75b0a6d81f33569e1e3b294280"} err="failed to get container status \"a64d887ab03c882e1a27a4cf36f359a8d18e3c75b0a6d81f33569e1e3b294280\": rpc error: code = NotFound desc = could not find container \"a64d887ab03c882e1a27a4cf36f359a8d18e3c75b0a6d81f33569e1e3b294280\": container with ID starting with a64d887ab03c882e1a27a4cf36f359a8d18e3c75b0a6d81f33569e1e3b294280 not found: ID does not exist" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.597626 4943 scope.go:117] "RemoveContainer" containerID="6fb3f1769f0d1092192c2f5f152c7d1fc52813d2e722687fb3ea1f6f08fba808" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.597907 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fb3f1769f0d1092192c2f5f152c7d1fc52813d2e722687fb3ea1f6f08fba808"} err="failed to get container status \"6fb3f1769f0d1092192c2f5f152c7d1fc52813d2e722687fb3ea1f6f08fba808\": rpc error: code = NotFound desc = could not find container \"6fb3f1769f0d1092192c2f5f152c7d1fc52813d2e722687fb3ea1f6f08fba808\": container with ID starting with 6fb3f1769f0d1092192c2f5f152c7d1fc52813d2e722687fb3ea1f6f08fba808 not found: ID does not exist" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.597986 4943 scope.go:117] "RemoveContainer" containerID="4dd5c06470991a2bcdddf5a5e429adcf4f4af8d3742de9c0a1e9cdd5705f74c5" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.602080 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4dd5c06470991a2bcdddf5a5e429adcf4f4af8d3742de9c0a1e9cdd5705f74c5"} err="failed to get container status \"4dd5c06470991a2bcdddf5a5e429adcf4f4af8d3742de9c0a1e9cdd5705f74c5\": rpc error: code = NotFound desc = could not find container \"4dd5c06470991a2bcdddf5a5e429adcf4f4af8d3742de9c0a1e9cdd5705f74c5\": container with ID starting with 4dd5c06470991a2bcdddf5a5e429adcf4f4af8d3742de9c0a1e9cdd5705f74c5 not found: ID does not exist" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.602130 4943 scope.go:117] "RemoveContainer" containerID="6a089f76e974f8c12f32f42b96b301fda03651276c599841b9c42fa064ffd5fd" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.602518 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a089f76e974f8c12f32f42b96b301fda03651276c599841b9c42fa064ffd5fd"} err="failed to get container status \"6a089f76e974f8c12f32f42b96b301fda03651276c599841b9c42fa064ffd5fd\": rpc error: code = NotFound desc = could not find container \"6a089f76e974f8c12f32f42b96b301fda03651276c599841b9c42fa064ffd5fd\": container with ID starting with 6a089f76e974f8c12f32f42b96b301fda03651276c599841b9c42fa064ffd5fd not found: ID does not exist" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.602571 4943 scope.go:117] "RemoveContainer" containerID="ce64b91eda434340bdacc6a953afe291459d3b18a9cbf842dc8d81723fcdd70e" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.602903 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce64b91eda434340bdacc6a953afe291459d3b18a9cbf842dc8d81723fcdd70e"} err="failed to get container status \"ce64b91eda434340bdacc6a953afe291459d3b18a9cbf842dc8d81723fcdd70e\": rpc error: code = NotFound desc = could not find container \"ce64b91eda434340bdacc6a953afe291459d3b18a9cbf842dc8d81723fcdd70e\": container with ID starting with ce64b91eda434340bdacc6a953afe291459d3b18a9cbf842dc8d81723fcdd70e not found: ID does not exist" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.602976 4943 scope.go:117] "RemoveContainer" containerID="a200f68bc20411f516725f81a930664ba5baaefd21a2792866262ded7e86fae5" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.603316 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a200f68bc20411f516725f81a930664ba5baaefd21a2792866262ded7e86fae5"} err="failed to get container status \"a200f68bc20411f516725f81a930664ba5baaefd21a2792866262ded7e86fae5\": rpc error: code = NotFound desc = could not find container \"a200f68bc20411f516725f81a930664ba5baaefd21a2792866262ded7e86fae5\": container with ID starting with a200f68bc20411f516725f81a930664ba5baaefd21a2792866262ded7e86fae5 not found: ID does not exist" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.603358 4943 scope.go:117] "RemoveContainer" containerID="f56c9c195453268c830bf92a70a64f70e75ef30e8ca853bb1f3a054a69c791af" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.603731 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f56c9c195453268c830bf92a70a64f70e75ef30e8ca853bb1f3a054a69c791af"} err="failed to get container status \"f56c9c195453268c830bf92a70a64f70e75ef30e8ca853bb1f3a054a69c791af\": rpc error: code = NotFound desc = could not find container \"f56c9c195453268c830bf92a70a64f70e75ef30e8ca853bb1f3a054a69c791af\": container with ID starting with f56c9c195453268c830bf92a70a64f70e75ef30e8ca853bb1f3a054a69c791af not found: ID does not exist" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.603789 4943 scope.go:117] "RemoveContainer" containerID="50e79fc40569f5b3fd73286933ee0242a1edd807a820edf7ba00a856d81c82c9" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.605332 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50e79fc40569f5b3fd73286933ee0242a1edd807a820edf7ba00a856d81c82c9"} err="failed to get container status \"50e79fc40569f5b3fd73286933ee0242a1edd807a820edf7ba00a856d81c82c9\": rpc error: code = NotFound desc = could not find container \"50e79fc40569f5b3fd73286933ee0242a1edd807a820edf7ba00a856d81c82c9\": container with ID starting with 50e79fc40569f5b3fd73286933ee0242a1edd807a820edf7ba00a856d81c82c9 not found: ID does not exist" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.605377 4943 scope.go:117] "RemoveContainer" containerID="959458a05302e835122b9db5d1eb4c449749a9e7ba2d85459edaeef2eede12de" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.606523 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"959458a05302e835122b9db5d1eb4c449749a9e7ba2d85459edaeef2eede12de"} err="failed to get container status \"959458a05302e835122b9db5d1eb4c449749a9e7ba2d85459edaeef2eede12de\": rpc error: code = NotFound desc = could not find container \"959458a05302e835122b9db5d1eb4c449749a9e7ba2d85459edaeef2eede12de\": container with ID starting with 959458a05302e835122b9db5d1eb4c449749a9e7ba2d85459edaeef2eede12de not found: ID does not exist" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.606575 4943 scope.go:117] "RemoveContainer" containerID="a64d887ab03c882e1a27a4cf36f359a8d18e3c75b0a6d81f33569e1e3b294280" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.606909 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a64d887ab03c882e1a27a4cf36f359a8d18e3c75b0a6d81f33569e1e3b294280"} err="failed to get container status \"a64d887ab03c882e1a27a4cf36f359a8d18e3c75b0a6d81f33569e1e3b294280\": rpc error: code = NotFound desc = could not find container \"a64d887ab03c882e1a27a4cf36f359a8d18e3c75b0a6d81f33569e1e3b294280\": container with ID starting with a64d887ab03c882e1a27a4cf36f359a8d18e3c75b0a6d81f33569e1e3b294280 not found: ID does not exist" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.606976 4943 scope.go:117] "RemoveContainer" containerID="6fb3f1769f0d1092192c2f5f152c7d1fc52813d2e722687fb3ea1f6f08fba808" Mar 07 14:50:27 crc kubenswrapper[4943]: I0307 14:50:27.607261 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fb3f1769f0d1092192c2f5f152c7d1fc52813d2e722687fb3ea1f6f08fba808"} err="failed to get container status \"6fb3f1769f0d1092192c2f5f152c7d1fc52813d2e722687fb3ea1f6f08fba808\": rpc error: code = NotFound desc = could not find container \"6fb3f1769f0d1092192c2f5f152c7d1fc52813d2e722687fb3ea1f6f08fba808\": container with ID starting with 6fb3f1769f0d1092192c2f5f152c7d1fc52813d2e722687fb3ea1f6f08fba808 not found: ID does not exist" Mar 07 14:50:28 crc kubenswrapper[4943]: I0307 14:50:28.326664 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-qrlpk_a13f1d2e-9974-4c40-85f5-d876d5644993/kube-multus/0.log" Mar 07 14:50:28 crc kubenswrapper[4943]: I0307 14:50:28.327186 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-qrlpk" event={"ID":"a13f1d2e-9974-4c40-85f5-d876d5644993","Type":"ContainerStarted","Data":"248cf40cf4e1e54b3fdb19fe2a6ea8d95219bf5ceef5ff7010e644f0ad702274"} Mar 07 14:50:28 crc kubenswrapper[4943]: I0307 14:50:28.336567 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" event={"ID":"ed6a5bc3-df22-4779-b20b-df0b86e06d68","Type":"ContainerStarted","Data":"3d1f6c6ec60bf92bf92d40584a9382930d3af2f91e9a29488418aeb4043886a5"} Mar 07 14:50:28 crc kubenswrapper[4943]: I0307 14:50:28.336641 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" event={"ID":"ed6a5bc3-df22-4779-b20b-df0b86e06d68","Type":"ContainerStarted","Data":"c7caeeff134a58729b2708ca26681e82d24d0bffbb0e4017a6fd052f3255985f"} Mar 07 14:50:28 crc kubenswrapper[4943]: I0307 14:50:28.336670 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" event={"ID":"ed6a5bc3-df22-4779-b20b-df0b86e06d68","Type":"ContainerStarted","Data":"34da60440e8ccddbae01091c09c7eb98882cf9c2c16780f95ffec05b99ba6e4c"} Mar 07 14:50:28 crc kubenswrapper[4943]: I0307 14:50:28.336695 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" event={"ID":"ed6a5bc3-df22-4779-b20b-df0b86e06d68","Type":"ContainerStarted","Data":"d3d2df6f1ac389286f5b15c99f7c51917f0369921f4f078f231681bf311dbbb8"} Mar 07 14:50:28 crc kubenswrapper[4943]: I0307 14:50:28.336719 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" event={"ID":"ed6a5bc3-df22-4779-b20b-df0b86e06d68","Type":"ContainerStarted","Data":"18513313ec3c37877ab5b83820dd180c72af90ce1eba132c165803ecbf681ac3"} Mar 07 14:50:28 crc kubenswrapper[4943]: I0307 14:50:28.336741 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" event={"ID":"ed6a5bc3-df22-4779-b20b-df0b86e06d68","Type":"ContainerStarted","Data":"5c39fc5227d3b53e218dd18ca0ebde5f3a3237bafb8e9f1e441aeae754fa3b95"} Mar 07 14:50:28 crc kubenswrapper[4943]: I0307 14:50:28.765481 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d340687c-faf0-48dc-a1ac-dec5067ac665" path="/var/lib/kubelet/pods/d340687c-faf0-48dc-a1ac-dec5067ac665/volumes" Mar 07 14:50:31 crc kubenswrapper[4943]: I0307 14:50:31.368983 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" event={"ID":"ed6a5bc3-df22-4779-b20b-df0b86e06d68","Type":"ContainerStarted","Data":"d552bbb519622d4381bfd7e80cca1c895a73a216a8362dc318335c5e8b089b6a"} Mar 07 14:50:33 crc kubenswrapper[4943]: I0307 14:50:33.394578 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" event={"ID":"ed6a5bc3-df22-4779-b20b-df0b86e06d68","Type":"ContainerStarted","Data":"ab8fa88f74ec8e6bc7fb32ce0fe695f8bbb1ca2008e5ab5a503f2f3a55a08e3f"} Mar 07 14:50:33 crc kubenswrapper[4943]: I0307 14:50:33.395092 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:33 crc kubenswrapper[4943]: I0307 14:50:33.395129 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:33 crc kubenswrapper[4943]: I0307 14:50:33.421331 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:33 crc kubenswrapper[4943]: I0307 14:50:33.429914 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" podStartSLOduration=7.429894008 podStartE2EDuration="7.429894008s" podCreationTimestamp="2026-03-07 14:50:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:50:33.429471458 +0000 UTC m=+675.381607966" watchObservedRunningTime="2026-03-07 14:50:33.429894008 +0000 UTC m=+675.382030516" Mar 07 14:50:34 crc kubenswrapper[4943]: I0307 14:50:34.399236 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:34 crc kubenswrapper[4943]: I0307 14:50:34.425633 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:52 crc kubenswrapper[4943]: I0307 14:50:52.052032 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f48wksn"] Mar 07 14:50:52 crc kubenswrapper[4943]: I0307 14:50:52.054334 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f48wksn" Mar 07 14:50:52 crc kubenswrapper[4943]: I0307 14:50:52.061221 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 07 14:50:52 crc kubenswrapper[4943]: I0307 14:50:52.075977 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f48wksn"] Mar 07 14:50:52 crc kubenswrapper[4943]: I0307 14:50:52.114447 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/dfbd5db0-a595-4698-85dd-6a6107bc23c5-util\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f48wksn\" (UID: \"dfbd5db0-a595-4698-85dd-6a6107bc23c5\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f48wksn" Mar 07 14:50:52 crc kubenswrapper[4943]: I0307 14:50:52.114621 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/dfbd5db0-a595-4698-85dd-6a6107bc23c5-bundle\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f48wksn\" (UID: \"dfbd5db0-a595-4698-85dd-6a6107bc23c5\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f48wksn" Mar 07 14:50:52 crc kubenswrapper[4943]: I0307 14:50:52.114666 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pv5wg\" (UniqueName: \"kubernetes.io/projected/dfbd5db0-a595-4698-85dd-6a6107bc23c5-kube-api-access-pv5wg\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f48wksn\" (UID: \"dfbd5db0-a595-4698-85dd-6a6107bc23c5\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f48wksn" Mar 07 14:50:52 crc kubenswrapper[4943]: I0307 14:50:52.216646 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pv5wg\" (UniqueName: \"kubernetes.io/projected/dfbd5db0-a595-4698-85dd-6a6107bc23c5-kube-api-access-pv5wg\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f48wksn\" (UID: \"dfbd5db0-a595-4698-85dd-6a6107bc23c5\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f48wksn" Mar 07 14:50:52 crc kubenswrapper[4943]: I0307 14:50:52.217232 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/dfbd5db0-a595-4698-85dd-6a6107bc23c5-util\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f48wksn\" (UID: \"dfbd5db0-a595-4698-85dd-6a6107bc23c5\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f48wksn" Mar 07 14:50:52 crc kubenswrapper[4943]: I0307 14:50:52.217605 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/dfbd5db0-a595-4698-85dd-6a6107bc23c5-bundle\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f48wksn\" (UID: \"dfbd5db0-a595-4698-85dd-6a6107bc23c5\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f48wksn" Mar 07 14:50:52 crc kubenswrapper[4943]: I0307 14:50:52.218382 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/dfbd5db0-a595-4698-85dd-6a6107bc23c5-util\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f48wksn\" (UID: \"dfbd5db0-a595-4698-85dd-6a6107bc23c5\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f48wksn" Mar 07 14:50:52 crc kubenswrapper[4943]: I0307 14:50:52.218557 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/dfbd5db0-a595-4698-85dd-6a6107bc23c5-bundle\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f48wksn\" (UID: \"dfbd5db0-a595-4698-85dd-6a6107bc23c5\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f48wksn" Mar 07 14:50:52 crc kubenswrapper[4943]: I0307 14:50:52.251448 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pv5wg\" (UniqueName: \"kubernetes.io/projected/dfbd5db0-a595-4698-85dd-6a6107bc23c5-kube-api-access-pv5wg\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f48wksn\" (UID: \"dfbd5db0-a595-4698-85dd-6a6107bc23c5\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f48wksn" Mar 07 14:50:52 crc kubenswrapper[4943]: I0307 14:50:52.380387 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f48wksn" Mar 07 14:50:52 crc kubenswrapper[4943]: I0307 14:50:52.703578 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f48wksn"] Mar 07 14:50:53 crc kubenswrapper[4943]: I0307 14:50:53.533889 4943 generic.go:334] "Generic (PLEG): container finished" podID="dfbd5db0-a595-4698-85dd-6a6107bc23c5" containerID="f9f79a7c476d9fc67d2b9f147582a67bda996ed39b022a12e95f1d6b2a4df1ec" exitCode=0 Mar 07 14:50:53 crc kubenswrapper[4943]: I0307 14:50:53.534002 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f48wksn" event={"ID":"dfbd5db0-a595-4698-85dd-6a6107bc23c5","Type":"ContainerDied","Data":"f9f79a7c476d9fc67d2b9f147582a67bda996ed39b022a12e95f1d6b2a4df1ec"} Mar 07 14:50:53 crc kubenswrapper[4943]: I0307 14:50:53.534069 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f48wksn" event={"ID":"dfbd5db0-a595-4698-85dd-6a6107bc23c5","Type":"ContainerStarted","Data":"41231e4ddf3a31f80b0efb2157b0402e1923163bc7bee34b86598dbbca577fa2"} Mar 07 14:50:55 crc kubenswrapper[4943]: I0307 14:50:55.552103 4943 generic.go:334] "Generic (PLEG): container finished" podID="dfbd5db0-a595-4698-85dd-6a6107bc23c5" containerID="d0ea24adedf2298245a776c901b9dca72ecaf256a3738b366982d2644eaca16b" exitCode=0 Mar 07 14:50:55 crc kubenswrapper[4943]: I0307 14:50:55.552158 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f48wksn" event={"ID":"dfbd5db0-a595-4698-85dd-6a6107bc23c5","Type":"ContainerDied","Data":"d0ea24adedf2298245a776c901b9dca72ecaf256a3738b366982d2644eaca16b"} Mar 07 14:50:56 crc kubenswrapper[4943]: I0307 14:50:56.567748 4943 generic.go:334] "Generic (PLEG): container finished" podID="dfbd5db0-a595-4698-85dd-6a6107bc23c5" containerID="afa0ac87cb3466314371d426d02792ed0999692565a83721569d4b702819d0e8" exitCode=0 Mar 07 14:50:56 crc kubenswrapper[4943]: I0307 14:50:56.567839 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f48wksn" event={"ID":"dfbd5db0-a595-4698-85dd-6a6107bc23c5","Type":"ContainerDied","Data":"afa0ac87cb3466314371d426d02792ed0999692565a83721569d4b702819d0e8"} Mar 07 14:50:57 crc kubenswrapper[4943]: I0307 14:50:57.042446 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-npgnc" Mar 07 14:50:58 crc kubenswrapper[4943]: I0307 14:50:58.001174 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f48wksn" Mar 07 14:50:58 crc kubenswrapper[4943]: I0307 14:50:58.160041 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pv5wg\" (UniqueName: \"kubernetes.io/projected/dfbd5db0-a595-4698-85dd-6a6107bc23c5-kube-api-access-pv5wg\") pod \"dfbd5db0-a595-4698-85dd-6a6107bc23c5\" (UID: \"dfbd5db0-a595-4698-85dd-6a6107bc23c5\") " Mar 07 14:50:58 crc kubenswrapper[4943]: I0307 14:50:58.160191 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/dfbd5db0-a595-4698-85dd-6a6107bc23c5-util\") pod \"dfbd5db0-a595-4698-85dd-6a6107bc23c5\" (UID: \"dfbd5db0-a595-4698-85dd-6a6107bc23c5\") " Mar 07 14:50:58 crc kubenswrapper[4943]: I0307 14:50:58.160273 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/dfbd5db0-a595-4698-85dd-6a6107bc23c5-bundle\") pod \"dfbd5db0-a595-4698-85dd-6a6107bc23c5\" (UID: \"dfbd5db0-a595-4698-85dd-6a6107bc23c5\") " Mar 07 14:50:58 crc kubenswrapper[4943]: I0307 14:50:58.162698 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dfbd5db0-a595-4698-85dd-6a6107bc23c5-bundle" (OuterVolumeSpecName: "bundle") pod "dfbd5db0-a595-4698-85dd-6a6107bc23c5" (UID: "dfbd5db0-a595-4698-85dd-6a6107bc23c5"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:50:58 crc kubenswrapper[4943]: I0307 14:50:58.170543 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfbd5db0-a595-4698-85dd-6a6107bc23c5-kube-api-access-pv5wg" (OuterVolumeSpecName: "kube-api-access-pv5wg") pod "dfbd5db0-a595-4698-85dd-6a6107bc23c5" (UID: "dfbd5db0-a595-4698-85dd-6a6107bc23c5"). InnerVolumeSpecName "kube-api-access-pv5wg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:50:58 crc kubenswrapper[4943]: I0307 14:50:58.179421 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dfbd5db0-a595-4698-85dd-6a6107bc23c5-util" (OuterVolumeSpecName: "util") pod "dfbd5db0-a595-4698-85dd-6a6107bc23c5" (UID: "dfbd5db0-a595-4698-85dd-6a6107bc23c5"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:50:58 crc kubenswrapper[4943]: I0307 14:50:58.261675 4943 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/dfbd5db0-a595-4698-85dd-6a6107bc23c5-bundle\") on node \"crc\" DevicePath \"\"" Mar 07 14:50:58 crc kubenswrapper[4943]: I0307 14:50:58.261728 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pv5wg\" (UniqueName: \"kubernetes.io/projected/dfbd5db0-a595-4698-85dd-6a6107bc23c5-kube-api-access-pv5wg\") on node \"crc\" DevicePath \"\"" Mar 07 14:50:58 crc kubenswrapper[4943]: I0307 14:50:58.261749 4943 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/dfbd5db0-a595-4698-85dd-6a6107bc23c5-util\") on node \"crc\" DevicePath \"\"" Mar 07 14:50:58 crc kubenswrapper[4943]: I0307 14:50:58.680056 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f48wksn" event={"ID":"dfbd5db0-a595-4698-85dd-6a6107bc23c5","Type":"ContainerDied","Data":"41231e4ddf3a31f80b0efb2157b0402e1923163bc7bee34b86598dbbca577fa2"} Mar 07 14:50:58 crc kubenswrapper[4943]: I0307 14:50:58.680135 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="41231e4ddf3a31f80b0efb2157b0402e1923163bc7bee34b86598dbbca577fa2" Mar 07 14:50:58 crc kubenswrapper[4943]: I0307 14:50:58.680203 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f48wksn" Mar 07 14:51:07 crc kubenswrapper[4943]: I0307 14:51:07.707691 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-655d5479c5-wnsv8"] Mar 07 14:51:07 crc kubenswrapper[4943]: E0307 14:51:07.708430 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfbd5db0-a595-4698-85dd-6a6107bc23c5" containerName="pull" Mar 07 14:51:07 crc kubenswrapper[4943]: I0307 14:51:07.708441 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfbd5db0-a595-4698-85dd-6a6107bc23c5" containerName="pull" Mar 07 14:51:07 crc kubenswrapper[4943]: E0307 14:51:07.708459 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfbd5db0-a595-4698-85dd-6a6107bc23c5" containerName="extract" Mar 07 14:51:07 crc kubenswrapper[4943]: I0307 14:51:07.708464 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfbd5db0-a595-4698-85dd-6a6107bc23c5" containerName="extract" Mar 07 14:51:07 crc kubenswrapper[4943]: E0307 14:51:07.708474 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfbd5db0-a595-4698-85dd-6a6107bc23c5" containerName="util" Mar 07 14:51:07 crc kubenswrapper[4943]: I0307 14:51:07.708479 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfbd5db0-a595-4698-85dd-6a6107bc23c5" containerName="util" Mar 07 14:51:07 crc kubenswrapper[4943]: I0307 14:51:07.708559 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfbd5db0-a595-4698-85dd-6a6107bc23c5" containerName="extract" Mar 07 14:51:07 crc kubenswrapper[4943]: I0307 14:51:07.708965 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-655d5479c5-wnsv8" Mar 07 14:51:07 crc kubenswrapper[4943]: I0307 14:51:07.714426 4943 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Mar 07 14:51:07 crc kubenswrapper[4943]: I0307 14:51:07.714885 4943 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-tdb98" Mar 07 14:51:07 crc kubenswrapper[4943]: I0307 14:51:07.716967 4943 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Mar 07 14:51:07 crc kubenswrapper[4943]: I0307 14:51:07.717144 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Mar 07 14:51:07 crc kubenswrapper[4943]: I0307 14:51:07.717144 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Mar 07 14:51:07 crc kubenswrapper[4943]: I0307 14:51:07.734689 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-655d5479c5-wnsv8"] Mar 07 14:51:07 crc kubenswrapper[4943]: I0307 14:51:07.885364 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/75e7a24e-c89f-49b4-bae4-cf2627e32068-apiservice-cert\") pod \"metallb-operator-controller-manager-655d5479c5-wnsv8\" (UID: \"75e7a24e-c89f-49b4-bae4-cf2627e32068\") " pod="metallb-system/metallb-operator-controller-manager-655d5479c5-wnsv8" Mar 07 14:51:07 crc kubenswrapper[4943]: I0307 14:51:07.885442 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/75e7a24e-c89f-49b4-bae4-cf2627e32068-webhook-cert\") pod \"metallb-operator-controller-manager-655d5479c5-wnsv8\" (UID: \"75e7a24e-c89f-49b4-bae4-cf2627e32068\") " pod="metallb-system/metallb-operator-controller-manager-655d5479c5-wnsv8" Mar 07 14:51:07 crc kubenswrapper[4943]: I0307 14:51:07.885518 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wk7v7\" (UniqueName: \"kubernetes.io/projected/75e7a24e-c89f-49b4-bae4-cf2627e32068-kube-api-access-wk7v7\") pod \"metallb-operator-controller-manager-655d5479c5-wnsv8\" (UID: \"75e7a24e-c89f-49b4-bae4-cf2627e32068\") " pod="metallb-system/metallb-operator-controller-manager-655d5479c5-wnsv8" Mar 07 14:51:07 crc kubenswrapper[4943]: I0307 14:51:07.987032 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/75e7a24e-c89f-49b4-bae4-cf2627e32068-apiservice-cert\") pod \"metallb-operator-controller-manager-655d5479c5-wnsv8\" (UID: \"75e7a24e-c89f-49b4-bae4-cf2627e32068\") " pod="metallb-system/metallb-operator-controller-manager-655d5479c5-wnsv8" Mar 07 14:51:07 crc kubenswrapper[4943]: I0307 14:51:07.987113 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/75e7a24e-c89f-49b4-bae4-cf2627e32068-webhook-cert\") pod \"metallb-operator-controller-manager-655d5479c5-wnsv8\" (UID: \"75e7a24e-c89f-49b4-bae4-cf2627e32068\") " pod="metallb-system/metallb-operator-controller-manager-655d5479c5-wnsv8" Mar 07 14:51:07 crc kubenswrapper[4943]: I0307 14:51:07.987173 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wk7v7\" (UniqueName: \"kubernetes.io/projected/75e7a24e-c89f-49b4-bae4-cf2627e32068-kube-api-access-wk7v7\") pod \"metallb-operator-controller-manager-655d5479c5-wnsv8\" (UID: \"75e7a24e-c89f-49b4-bae4-cf2627e32068\") " pod="metallb-system/metallb-operator-controller-manager-655d5479c5-wnsv8" Mar 07 14:51:07 crc kubenswrapper[4943]: I0307 14:51:07.995612 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/75e7a24e-c89f-49b4-bae4-cf2627e32068-apiservice-cert\") pod \"metallb-operator-controller-manager-655d5479c5-wnsv8\" (UID: \"75e7a24e-c89f-49b4-bae4-cf2627e32068\") " pod="metallb-system/metallb-operator-controller-manager-655d5479c5-wnsv8" Mar 07 14:51:07 crc kubenswrapper[4943]: I0307 14:51:07.998303 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/75e7a24e-c89f-49b4-bae4-cf2627e32068-webhook-cert\") pod \"metallb-operator-controller-manager-655d5479c5-wnsv8\" (UID: \"75e7a24e-c89f-49b4-bae4-cf2627e32068\") " pod="metallb-system/metallb-operator-controller-manager-655d5479c5-wnsv8" Mar 07 14:51:08 crc kubenswrapper[4943]: I0307 14:51:08.033480 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-5f7c595484-khsdg"] Mar 07 14:51:08 crc kubenswrapper[4943]: I0307 14:51:08.034379 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5f7c595484-khsdg" Mar 07 14:51:08 crc kubenswrapper[4943]: I0307 14:51:08.036733 4943 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Mar 07 14:51:08 crc kubenswrapper[4943]: I0307 14:51:08.036969 4943 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Mar 07 14:51:08 crc kubenswrapper[4943]: I0307 14:51:08.037119 4943 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-c7nqv" Mar 07 14:51:08 crc kubenswrapper[4943]: I0307 14:51:08.052809 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5f7c595484-khsdg"] Mar 07 14:51:08 crc kubenswrapper[4943]: I0307 14:51:08.053912 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wk7v7\" (UniqueName: \"kubernetes.io/projected/75e7a24e-c89f-49b4-bae4-cf2627e32068-kube-api-access-wk7v7\") pod \"metallb-operator-controller-manager-655d5479c5-wnsv8\" (UID: \"75e7a24e-c89f-49b4-bae4-cf2627e32068\") " pod="metallb-system/metallb-operator-controller-manager-655d5479c5-wnsv8" Mar 07 14:51:08 crc kubenswrapper[4943]: I0307 14:51:08.189013 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rh9h\" (UniqueName: \"kubernetes.io/projected/de0f2fd6-4595-498c-9e76-6f8635fe23f6-kube-api-access-2rh9h\") pod \"metallb-operator-webhook-server-5f7c595484-khsdg\" (UID: \"de0f2fd6-4595-498c-9e76-6f8635fe23f6\") " pod="metallb-system/metallb-operator-webhook-server-5f7c595484-khsdg" Mar 07 14:51:08 crc kubenswrapper[4943]: I0307 14:51:08.189073 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/de0f2fd6-4595-498c-9e76-6f8635fe23f6-webhook-cert\") pod \"metallb-operator-webhook-server-5f7c595484-khsdg\" (UID: \"de0f2fd6-4595-498c-9e76-6f8635fe23f6\") " pod="metallb-system/metallb-operator-webhook-server-5f7c595484-khsdg" Mar 07 14:51:08 crc kubenswrapper[4943]: I0307 14:51:08.189101 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/de0f2fd6-4595-498c-9e76-6f8635fe23f6-apiservice-cert\") pod \"metallb-operator-webhook-server-5f7c595484-khsdg\" (UID: \"de0f2fd6-4595-498c-9e76-6f8635fe23f6\") " pod="metallb-system/metallb-operator-webhook-server-5f7c595484-khsdg" Mar 07 14:51:08 crc kubenswrapper[4943]: I0307 14:51:08.290546 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rh9h\" (UniqueName: \"kubernetes.io/projected/de0f2fd6-4595-498c-9e76-6f8635fe23f6-kube-api-access-2rh9h\") pod \"metallb-operator-webhook-server-5f7c595484-khsdg\" (UID: \"de0f2fd6-4595-498c-9e76-6f8635fe23f6\") " pod="metallb-system/metallb-operator-webhook-server-5f7c595484-khsdg" Mar 07 14:51:08 crc kubenswrapper[4943]: I0307 14:51:08.290664 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/de0f2fd6-4595-498c-9e76-6f8635fe23f6-webhook-cert\") pod \"metallb-operator-webhook-server-5f7c595484-khsdg\" (UID: \"de0f2fd6-4595-498c-9e76-6f8635fe23f6\") " pod="metallb-system/metallb-operator-webhook-server-5f7c595484-khsdg" Mar 07 14:51:08 crc kubenswrapper[4943]: I0307 14:51:08.290707 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/de0f2fd6-4595-498c-9e76-6f8635fe23f6-apiservice-cert\") pod \"metallb-operator-webhook-server-5f7c595484-khsdg\" (UID: \"de0f2fd6-4595-498c-9e76-6f8635fe23f6\") " pod="metallb-system/metallb-operator-webhook-server-5f7c595484-khsdg" Mar 07 14:51:08 crc kubenswrapper[4943]: I0307 14:51:08.294766 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/de0f2fd6-4595-498c-9e76-6f8635fe23f6-webhook-cert\") pod \"metallb-operator-webhook-server-5f7c595484-khsdg\" (UID: \"de0f2fd6-4595-498c-9e76-6f8635fe23f6\") " pod="metallb-system/metallb-operator-webhook-server-5f7c595484-khsdg" Mar 07 14:51:08 crc kubenswrapper[4943]: I0307 14:51:08.295799 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/de0f2fd6-4595-498c-9e76-6f8635fe23f6-apiservice-cert\") pod \"metallb-operator-webhook-server-5f7c595484-khsdg\" (UID: \"de0f2fd6-4595-498c-9e76-6f8635fe23f6\") " pod="metallb-system/metallb-operator-webhook-server-5f7c595484-khsdg" Mar 07 14:51:08 crc kubenswrapper[4943]: I0307 14:51:08.318255 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rh9h\" (UniqueName: \"kubernetes.io/projected/de0f2fd6-4595-498c-9e76-6f8635fe23f6-kube-api-access-2rh9h\") pod \"metallb-operator-webhook-server-5f7c595484-khsdg\" (UID: \"de0f2fd6-4595-498c-9e76-6f8635fe23f6\") " pod="metallb-system/metallb-operator-webhook-server-5f7c595484-khsdg" Mar 07 14:51:08 crc kubenswrapper[4943]: I0307 14:51:08.324941 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-655d5479c5-wnsv8" Mar 07 14:51:08 crc kubenswrapper[4943]: I0307 14:51:08.370596 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5f7c595484-khsdg" Mar 07 14:51:08 crc kubenswrapper[4943]: I0307 14:51:08.888330 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-655d5479c5-wnsv8"] Mar 07 14:51:08 crc kubenswrapper[4943]: W0307 14:51:08.987001 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podde0f2fd6_4595_498c_9e76_6f8635fe23f6.slice/crio-2744765928e27371c9a2d18a177454331e7999c27e74da8e8d3560d0b84b3c3d WatchSource:0}: Error finding container 2744765928e27371c9a2d18a177454331e7999c27e74da8e8d3560d0b84b3c3d: Status 404 returned error can't find the container with id 2744765928e27371c9a2d18a177454331e7999c27e74da8e8d3560d0b84b3c3d Mar 07 14:51:08 crc kubenswrapper[4943]: I0307 14:51:08.990653 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5f7c595484-khsdg"] Mar 07 14:51:09 crc kubenswrapper[4943]: I0307 14:51:09.748962 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5f7c595484-khsdg" event={"ID":"de0f2fd6-4595-498c-9e76-6f8635fe23f6","Type":"ContainerStarted","Data":"2744765928e27371c9a2d18a177454331e7999c27e74da8e8d3560d0b84b3c3d"} Mar 07 14:51:09 crc kubenswrapper[4943]: I0307 14:51:09.750097 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-655d5479c5-wnsv8" event={"ID":"75e7a24e-c89f-49b4-bae4-cf2627e32068","Type":"ContainerStarted","Data":"395174b9f60994f3d4cfc7b3e7c086e0bbeb50f605fa63c3e80de61dc636caf2"} Mar 07 14:51:12 crc kubenswrapper[4943]: I0307 14:51:12.785105 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-655d5479c5-wnsv8" event={"ID":"75e7a24e-c89f-49b4-bae4-cf2627e32068","Type":"ContainerStarted","Data":"ded62e4fb2cdde7399e4da2ec79ee1621a21d9070b2fb900d93a524eef0e8997"} Mar 07 14:51:12 crc kubenswrapper[4943]: I0307 14:51:12.788247 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-655d5479c5-wnsv8" Mar 07 14:51:12 crc kubenswrapper[4943]: I0307 14:51:12.812355 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-655d5479c5-wnsv8" podStartSLOduration=2.657573372 podStartE2EDuration="5.812332016s" podCreationTimestamp="2026-03-07 14:51:07 +0000 UTC" firstStartedPulling="2026-03-07 14:51:08.902652695 +0000 UTC m=+710.854789183" lastFinishedPulling="2026-03-07 14:51:12.057411329 +0000 UTC m=+714.009547827" observedRunningTime="2026-03-07 14:51:12.811974207 +0000 UTC m=+714.764110715" watchObservedRunningTime="2026-03-07 14:51:12.812332016 +0000 UTC m=+714.764468514" Mar 07 14:51:14 crc kubenswrapper[4943]: I0307 14:51:14.817299 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5f7c595484-khsdg" event={"ID":"de0f2fd6-4595-498c-9e76-6f8635fe23f6","Type":"ContainerStarted","Data":"b379e3cf014719644490e070a780f0a89035438a250bb926adf2a36af0a4cac7"} Mar 07 14:51:14 crc kubenswrapper[4943]: I0307 14:51:14.817894 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-5f7c595484-khsdg" Mar 07 14:51:14 crc kubenswrapper[4943]: I0307 14:51:14.845176 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-5f7c595484-khsdg" podStartSLOduration=1.983676795 podStartE2EDuration="6.845152721s" podCreationTimestamp="2026-03-07 14:51:08 +0000 UTC" firstStartedPulling="2026-03-07 14:51:08.990006775 +0000 UTC m=+710.942143273" lastFinishedPulling="2026-03-07 14:51:13.851482701 +0000 UTC m=+715.803619199" observedRunningTime="2026-03-07 14:51:14.84428621 +0000 UTC m=+716.796422708" watchObservedRunningTime="2026-03-07 14:51:14.845152721 +0000 UTC m=+716.797289229" Mar 07 14:51:28 crc kubenswrapper[4943]: I0307 14:51:28.391535 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-5f7c595484-khsdg" Mar 07 14:51:32 crc kubenswrapper[4943]: I0307 14:51:32.028588 4943 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 07 14:51:48 crc kubenswrapper[4943]: I0307 14:51:48.328101 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-655d5479c5-wnsv8" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.052869 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-ntxr6"] Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.055575 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-ntxr6" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.059864 4943 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-x4rll" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.060496 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.060637 4943 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.061147 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7f989f654f-h29zk"] Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.061919 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7f989f654f-h29zk" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.064702 4943 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.077853 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7f989f654f-h29zk"] Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.156184 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-22p8b"] Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.164807 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-22p8b" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.165806 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6e721eda-a709-4537-8f7a-4bdc31bf72f8-cert\") pod \"frr-k8s-webhook-server-7f989f654f-h29zk\" (UID: \"6e721eda-a709-4537-8f7a-4bdc31bf72f8\") " pod="metallb-system/frr-k8s-webhook-server-7f989f654f-h29zk" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.165863 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/19220a5d-633c-4051-87d4-246e167e6db5-reloader\") pod \"frr-k8s-ntxr6\" (UID: \"19220a5d-633c-4051-87d4-246e167e6db5\") " pod="metallb-system/frr-k8s-ntxr6" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.165890 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thzt4\" (UniqueName: \"kubernetes.io/projected/19220a5d-633c-4051-87d4-246e167e6db5-kube-api-access-thzt4\") pod \"frr-k8s-ntxr6\" (UID: \"19220a5d-633c-4051-87d4-246e167e6db5\") " pod="metallb-system/frr-k8s-ntxr6" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.165940 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/19220a5d-633c-4051-87d4-246e167e6db5-frr-conf\") pod \"frr-k8s-ntxr6\" (UID: \"19220a5d-633c-4051-87d4-246e167e6db5\") " pod="metallb-system/frr-k8s-ntxr6" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.165970 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/19220a5d-633c-4051-87d4-246e167e6db5-metrics-certs\") pod \"frr-k8s-ntxr6\" (UID: \"19220a5d-633c-4051-87d4-246e167e6db5\") " pod="metallb-system/frr-k8s-ntxr6" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.168694 4943 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.168769 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.169045 4943 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.172484 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/19220a5d-633c-4051-87d4-246e167e6db5-frr-sockets\") pod \"frr-k8s-ntxr6\" (UID: \"19220a5d-633c-4051-87d4-246e167e6db5\") " pod="metallb-system/frr-k8s-ntxr6" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.172676 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/19220a5d-633c-4051-87d4-246e167e6db5-frr-startup\") pod \"frr-k8s-ntxr6\" (UID: \"19220a5d-633c-4051-87d4-246e167e6db5\") " pod="metallb-system/frr-k8s-ntxr6" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.172509 4943 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-6r8cq" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.172800 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/19220a5d-633c-4051-87d4-246e167e6db5-metrics\") pod \"frr-k8s-ntxr6\" (UID: \"19220a5d-633c-4051-87d4-246e167e6db5\") " pod="metallb-system/frr-k8s-ntxr6" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.173010 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgdd9\" (UniqueName: \"kubernetes.io/projected/6e721eda-a709-4537-8f7a-4bdc31bf72f8-kube-api-access-sgdd9\") pod \"frr-k8s-webhook-server-7f989f654f-h29zk\" (UID: \"6e721eda-a709-4537-8f7a-4bdc31bf72f8\") " pod="metallb-system/frr-k8s-webhook-server-7f989f654f-h29zk" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.199135 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-86ddb6bd46-wxgzg"] Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.200308 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-86ddb6bd46-wxgzg" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.203005 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-86ddb6bd46-wxgzg"] Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.206108 4943 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.275099 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/19220a5d-633c-4051-87d4-246e167e6db5-frr-sockets\") pod \"frr-k8s-ntxr6\" (UID: \"19220a5d-633c-4051-87d4-246e167e6db5\") " pod="metallb-system/frr-k8s-ntxr6" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.275184 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e8cc9b78-5dd4-43c6-a1fe-702f4f1c30c9-metrics-certs\") pod \"speaker-22p8b\" (UID: \"e8cc9b78-5dd4-43c6-a1fe-702f4f1c30c9\") " pod="metallb-system/speaker-22p8b" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.275218 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/19220a5d-633c-4051-87d4-246e167e6db5-frr-startup\") pod \"frr-k8s-ntxr6\" (UID: \"19220a5d-633c-4051-87d4-246e167e6db5\") " pod="metallb-system/frr-k8s-ntxr6" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.275241 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/19220a5d-633c-4051-87d4-246e167e6db5-metrics\") pod \"frr-k8s-ntxr6\" (UID: \"19220a5d-633c-4051-87d4-246e167e6db5\") " pod="metallb-system/frr-k8s-ntxr6" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.275265 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgdd9\" (UniqueName: \"kubernetes.io/projected/6e721eda-a709-4537-8f7a-4bdc31bf72f8-kube-api-access-sgdd9\") pod \"frr-k8s-webhook-server-7f989f654f-h29zk\" (UID: \"6e721eda-a709-4537-8f7a-4bdc31bf72f8\") " pod="metallb-system/frr-k8s-webhook-server-7f989f654f-h29zk" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.275301 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hm5r7\" (UniqueName: \"kubernetes.io/projected/e8cc9b78-5dd4-43c6-a1fe-702f4f1c30c9-kube-api-access-hm5r7\") pod \"speaker-22p8b\" (UID: \"e8cc9b78-5dd4-43c6-a1fe-702f4f1c30c9\") " pod="metallb-system/speaker-22p8b" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.275324 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6e721eda-a709-4537-8f7a-4bdc31bf72f8-cert\") pod \"frr-k8s-webhook-server-7f989f654f-h29zk\" (UID: \"6e721eda-a709-4537-8f7a-4bdc31bf72f8\") " pod="metallb-system/frr-k8s-webhook-server-7f989f654f-h29zk" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.275348 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/19220a5d-633c-4051-87d4-246e167e6db5-reloader\") pod \"frr-k8s-ntxr6\" (UID: \"19220a5d-633c-4051-87d4-246e167e6db5\") " pod="metallb-system/frr-k8s-ntxr6" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.275368 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thzt4\" (UniqueName: \"kubernetes.io/projected/19220a5d-633c-4051-87d4-246e167e6db5-kube-api-access-thzt4\") pod \"frr-k8s-ntxr6\" (UID: \"19220a5d-633c-4051-87d4-246e167e6db5\") " pod="metallb-system/frr-k8s-ntxr6" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.275391 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e8cc9b78-5dd4-43c6-a1fe-702f4f1c30c9-memberlist\") pod \"speaker-22p8b\" (UID: \"e8cc9b78-5dd4-43c6-a1fe-702f4f1c30c9\") " pod="metallb-system/speaker-22p8b" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.275432 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/19220a5d-633c-4051-87d4-246e167e6db5-frr-conf\") pod \"frr-k8s-ntxr6\" (UID: \"19220a5d-633c-4051-87d4-246e167e6db5\") " pod="metallb-system/frr-k8s-ntxr6" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.275451 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/e8cc9b78-5dd4-43c6-a1fe-702f4f1c30c9-metallb-excludel2\") pod \"speaker-22p8b\" (UID: \"e8cc9b78-5dd4-43c6-a1fe-702f4f1c30c9\") " pod="metallb-system/speaker-22p8b" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.275483 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/19220a5d-633c-4051-87d4-246e167e6db5-metrics-certs\") pod \"frr-k8s-ntxr6\" (UID: \"19220a5d-633c-4051-87d4-246e167e6db5\") " pod="metallb-system/frr-k8s-ntxr6" Mar 07 14:51:49 crc kubenswrapper[4943]: E0307 14:51:49.277490 4943 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Mar 07 14:51:49 crc kubenswrapper[4943]: E0307 14:51:49.277554 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6e721eda-a709-4537-8f7a-4bdc31bf72f8-cert podName:6e721eda-a709-4537-8f7a-4bdc31bf72f8 nodeName:}" failed. No retries permitted until 2026-03-07 14:51:49.777534412 +0000 UTC m=+751.729670900 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6e721eda-a709-4537-8f7a-4bdc31bf72f8-cert") pod "frr-k8s-webhook-server-7f989f654f-h29zk" (UID: "6e721eda-a709-4537-8f7a-4bdc31bf72f8") : secret "frr-k8s-webhook-server-cert" not found Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.277605 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/19220a5d-633c-4051-87d4-246e167e6db5-frr-sockets\") pod \"frr-k8s-ntxr6\" (UID: \"19220a5d-633c-4051-87d4-246e167e6db5\") " pod="metallb-system/frr-k8s-ntxr6" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.278012 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/19220a5d-633c-4051-87d4-246e167e6db5-reloader\") pod \"frr-k8s-ntxr6\" (UID: \"19220a5d-633c-4051-87d4-246e167e6db5\") " pod="metallb-system/frr-k8s-ntxr6" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.278238 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/19220a5d-633c-4051-87d4-246e167e6db5-metrics\") pod \"frr-k8s-ntxr6\" (UID: \"19220a5d-633c-4051-87d4-246e167e6db5\") " pod="metallb-system/frr-k8s-ntxr6" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.278437 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/19220a5d-633c-4051-87d4-246e167e6db5-frr-conf\") pod \"frr-k8s-ntxr6\" (UID: \"19220a5d-633c-4051-87d4-246e167e6db5\") " pod="metallb-system/frr-k8s-ntxr6" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.278451 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/19220a5d-633c-4051-87d4-246e167e6db5-frr-startup\") pod \"frr-k8s-ntxr6\" (UID: \"19220a5d-633c-4051-87d4-246e167e6db5\") " pod="metallb-system/frr-k8s-ntxr6" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.281392 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/19220a5d-633c-4051-87d4-246e167e6db5-metrics-certs\") pod \"frr-k8s-ntxr6\" (UID: \"19220a5d-633c-4051-87d4-246e167e6db5\") " pod="metallb-system/frr-k8s-ntxr6" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.306512 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thzt4\" (UniqueName: \"kubernetes.io/projected/19220a5d-633c-4051-87d4-246e167e6db5-kube-api-access-thzt4\") pod \"frr-k8s-ntxr6\" (UID: \"19220a5d-633c-4051-87d4-246e167e6db5\") " pod="metallb-system/frr-k8s-ntxr6" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.309651 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgdd9\" (UniqueName: \"kubernetes.io/projected/6e721eda-a709-4537-8f7a-4bdc31bf72f8-kube-api-access-sgdd9\") pod \"frr-k8s-webhook-server-7f989f654f-h29zk\" (UID: \"6e721eda-a709-4537-8f7a-4bdc31bf72f8\") " pod="metallb-system/frr-k8s-webhook-server-7f989f654f-h29zk" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.376485 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-ntxr6" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.376702 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hm5r7\" (UniqueName: \"kubernetes.io/projected/e8cc9b78-5dd4-43c6-a1fe-702f4f1c30c9-kube-api-access-hm5r7\") pod \"speaker-22p8b\" (UID: \"e8cc9b78-5dd4-43c6-a1fe-702f4f1c30c9\") " pod="metallb-system/speaker-22p8b" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.376746 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e8cc9b78-5dd4-43c6-a1fe-702f4f1c30c9-memberlist\") pod \"speaker-22p8b\" (UID: \"e8cc9b78-5dd4-43c6-a1fe-702f4f1c30c9\") " pod="metallb-system/speaker-22p8b" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.376793 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/31d1a300-cbac-4266-8182-300930377412-metrics-certs\") pod \"controller-86ddb6bd46-wxgzg\" (UID: \"31d1a300-cbac-4266-8182-300930377412\") " pod="metallb-system/controller-86ddb6bd46-wxgzg" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.376813 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/e8cc9b78-5dd4-43c6-a1fe-702f4f1c30c9-metallb-excludel2\") pod \"speaker-22p8b\" (UID: \"e8cc9b78-5dd4-43c6-a1fe-702f4f1c30c9\") " pod="metallb-system/speaker-22p8b" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.376855 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/31d1a300-cbac-4266-8182-300930377412-cert\") pod \"controller-86ddb6bd46-wxgzg\" (UID: \"31d1a300-cbac-4266-8182-300930377412\") " pod="metallb-system/controller-86ddb6bd46-wxgzg" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.376877 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-db4lm\" (UniqueName: \"kubernetes.io/projected/31d1a300-cbac-4266-8182-300930377412-kube-api-access-db4lm\") pod \"controller-86ddb6bd46-wxgzg\" (UID: \"31d1a300-cbac-4266-8182-300930377412\") " pod="metallb-system/controller-86ddb6bd46-wxgzg" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.376900 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e8cc9b78-5dd4-43c6-a1fe-702f4f1c30c9-metrics-certs\") pod \"speaker-22p8b\" (UID: \"e8cc9b78-5dd4-43c6-a1fe-702f4f1c30c9\") " pod="metallb-system/speaker-22p8b" Mar 07 14:51:49 crc kubenswrapper[4943]: E0307 14:51:49.377362 4943 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Mar 07 14:51:49 crc kubenswrapper[4943]: E0307 14:51:49.377458 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e8cc9b78-5dd4-43c6-a1fe-702f4f1c30c9-memberlist podName:e8cc9b78-5dd4-43c6-a1fe-702f4f1c30c9 nodeName:}" failed. No retries permitted until 2026-03-07 14:51:49.877432842 +0000 UTC m=+751.829569400 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/e8cc9b78-5dd4-43c6-a1fe-702f4f1c30c9-memberlist") pod "speaker-22p8b" (UID: "e8cc9b78-5dd4-43c6-a1fe-702f4f1c30c9") : secret "metallb-memberlist" not found Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.378214 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/e8cc9b78-5dd4-43c6-a1fe-702f4f1c30c9-metallb-excludel2\") pod \"speaker-22p8b\" (UID: \"e8cc9b78-5dd4-43c6-a1fe-702f4f1c30c9\") " pod="metallb-system/speaker-22p8b" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.381144 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e8cc9b78-5dd4-43c6-a1fe-702f4f1c30c9-metrics-certs\") pod \"speaker-22p8b\" (UID: \"e8cc9b78-5dd4-43c6-a1fe-702f4f1c30c9\") " pod="metallb-system/speaker-22p8b" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.399356 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hm5r7\" (UniqueName: \"kubernetes.io/projected/e8cc9b78-5dd4-43c6-a1fe-702f4f1c30c9-kube-api-access-hm5r7\") pod \"speaker-22p8b\" (UID: \"e8cc9b78-5dd4-43c6-a1fe-702f4f1c30c9\") " pod="metallb-system/speaker-22p8b" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.478012 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/31d1a300-cbac-4266-8182-300930377412-cert\") pod \"controller-86ddb6bd46-wxgzg\" (UID: \"31d1a300-cbac-4266-8182-300930377412\") " pod="metallb-system/controller-86ddb6bd46-wxgzg" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.478062 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-db4lm\" (UniqueName: \"kubernetes.io/projected/31d1a300-cbac-4266-8182-300930377412-kube-api-access-db4lm\") pod \"controller-86ddb6bd46-wxgzg\" (UID: \"31d1a300-cbac-4266-8182-300930377412\") " pod="metallb-system/controller-86ddb6bd46-wxgzg" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.478127 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/31d1a300-cbac-4266-8182-300930377412-metrics-certs\") pod \"controller-86ddb6bd46-wxgzg\" (UID: \"31d1a300-cbac-4266-8182-300930377412\") " pod="metallb-system/controller-86ddb6bd46-wxgzg" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.480278 4943 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.484464 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/31d1a300-cbac-4266-8182-300930377412-metrics-certs\") pod \"controller-86ddb6bd46-wxgzg\" (UID: \"31d1a300-cbac-4266-8182-300930377412\") " pod="metallb-system/controller-86ddb6bd46-wxgzg" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.493941 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/31d1a300-cbac-4266-8182-300930377412-cert\") pod \"controller-86ddb6bd46-wxgzg\" (UID: \"31d1a300-cbac-4266-8182-300930377412\") " pod="metallb-system/controller-86ddb6bd46-wxgzg" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.500518 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-db4lm\" (UniqueName: \"kubernetes.io/projected/31d1a300-cbac-4266-8182-300930377412-kube-api-access-db4lm\") pod \"controller-86ddb6bd46-wxgzg\" (UID: \"31d1a300-cbac-4266-8182-300930377412\") " pod="metallb-system/controller-86ddb6bd46-wxgzg" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.522175 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-86ddb6bd46-wxgzg" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.781759 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6e721eda-a709-4537-8f7a-4bdc31bf72f8-cert\") pod \"frr-k8s-webhook-server-7f989f654f-h29zk\" (UID: \"6e721eda-a709-4537-8f7a-4bdc31bf72f8\") " pod="metallb-system/frr-k8s-webhook-server-7f989f654f-h29zk" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.786142 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6e721eda-a709-4537-8f7a-4bdc31bf72f8-cert\") pod \"frr-k8s-webhook-server-7f989f654f-h29zk\" (UID: \"6e721eda-a709-4537-8f7a-4bdc31bf72f8\") " pod="metallb-system/frr-k8s-webhook-server-7f989f654f-h29zk" Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.883248 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e8cc9b78-5dd4-43c6-a1fe-702f4f1c30c9-memberlist\") pod \"speaker-22p8b\" (UID: \"e8cc9b78-5dd4-43c6-a1fe-702f4f1c30c9\") " pod="metallb-system/speaker-22p8b" Mar 07 14:51:49 crc kubenswrapper[4943]: E0307 14:51:49.883479 4943 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Mar 07 14:51:49 crc kubenswrapper[4943]: E0307 14:51:49.883583 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e8cc9b78-5dd4-43c6-a1fe-702f4f1c30c9-memberlist podName:e8cc9b78-5dd4-43c6-a1fe-702f4f1c30c9 nodeName:}" failed. No retries permitted until 2026-03-07 14:51:50.883561785 +0000 UTC m=+752.835698293 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/e8cc9b78-5dd4-43c6-a1fe-702f4f1c30c9-memberlist") pod "speaker-22p8b" (UID: "e8cc9b78-5dd4-43c6-a1fe-702f4f1c30c9") : secret "metallb-memberlist" not found Mar 07 14:51:49 crc kubenswrapper[4943]: I0307 14:51:49.986637 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7f989f654f-h29zk" Mar 07 14:51:50 crc kubenswrapper[4943]: I0307 14:51:50.088652 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-86ddb6bd46-wxgzg"] Mar 07 14:51:50 crc kubenswrapper[4943]: W0307 14:51:50.098057 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod31d1a300_cbac_4266_8182_300930377412.slice/crio-7e6ba56f768921592a95ce2551dd9d2946796bf0cca11f9c35ce069511c4699d WatchSource:0}: Error finding container 7e6ba56f768921592a95ce2551dd9d2946796bf0cca11f9c35ce069511c4699d: Status 404 returned error can't find the container with id 7e6ba56f768921592a95ce2551dd9d2946796bf0cca11f9c35ce069511c4699d Mar 07 14:51:50 crc kubenswrapper[4943]: I0307 14:51:50.146215 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-86ddb6bd46-wxgzg" event={"ID":"31d1a300-cbac-4266-8182-300930377412","Type":"ContainerStarted","Data":"7e6ba56f768921592a95ce2551dd9d2946796bf0cca11f9c35ce069511c4699d"} Mar 07 14:51:50 crc kubenswrapper[4943]: I0307 14:51:50.147426 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ntxr6" event={"ID":"19220a5d-633c-4051-87d4-246e167e6db5","Type":"ContainerStarted","Data":"7d2f78703c4a54082a5799547d1d6a38dbc25ce333224b742cc7ac4b552b1d4a"} Mar 07 14:51:50 crc kubenswrapper[4943]: I0307 14:51:50.241567 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7f989f654f-h29zk"] Mar 07 14:51:50 crc kubenswrapper[4943]: W0307 14:51:50.250603 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6e721eda_a709_4537_8f7a_4bdc31bf72f8.slice/crio-f2e79e2366ac826a1f1cae98a7daa7e630440202ed451efc341a3a6f3c803bd3 WatchSource:0}: Error finding container f2e79e2366ac826a1f1cae98a7daa7e630440202ed451efc341a3a6f3c803bd3: Status 404 returned error can't find the container with id f2e79e2366ac826a1f1cae98a7daa7e630440202ed451efc341a3a6f3c803bd3 Mar 07 14:51:50 crc kubenswrapper[4943]: I0307 14:51:50.896836 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e8cc9b78-5dd4-43c6-a1fe-702f4f1c30c9-memberlist\") pod \"speaker-22p8b\" (UID: \"e8cc9b78-5dd4-43c6-a1fe-702f4f1c30c9\") " pod="metallb-system/speaker-22p8b" Mar 07 14:51:50 crc kubenswrapper[4943]: I0307 14:51:50.907447 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e8cc9b78-5dd4-43c6-a1fe-702f4f1c30c9-memberlist\") pod \"speaker-22p8b\" (UID: \"e8cc9b78-5dd4-43c6-a1fe-702f4f1c30c9\") " pod="metallb-system/speaker-22p8b" Mar 07 14:51:51 crc kubenswrapper[4943]: I0307 14:51:51.005456 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-22p8b" Mar 07 14:51:51 crc kubenswrapper[4943]: W0307 14:51:51.045824 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode8cc9b78_5dd4_43c6_a1fe_702f4f1c30c9.slice/crio-36480fddaaee5deac90aacfc9dec6e62225757f5e83a59803afd8c379f79df8b WatchSource:0}: Error finding container 36480fddaaee5deac90aacfc9dec6e62225757f5e83a59803afd8c379f79df8b: Status 404 returned error can't find the container with id 36480fddaaee5deac90aacfc9dec6e62225757f5e83a59803afd8c379f79df8b Mar 07 14:51:51 crc kubenswrapper[4943]: I0307 14:51:51.159536 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-86ddb6bd46-wxgzg" event={"ID":"31d1a300-cbac-4266-8182-300930377412","Type":"ContainerStarted","Data":"0cc3827b0a7c0ac1e423eaced76369a6119737a6d92cde270195177af1761d46"} Mar 07 14:51:51 crc kubenswrapper[4943]: I0307 14:51:51.161195 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-22p8b" event={"ID":"e8cc9b78-5dd4-43c6-a1fe-702f4f1c30c9","Type":"ContainerStarted","Data":"36480fddaaee5deac90aacfc9dec6e62225757f5e83a59803afd8c379f79df8b"} Mar 07 14:51:51 crc kubenswrapper[4943]: I0307 14:51:51.163514 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7f989f654f-h29zk" event={"ID":"6e721eda-a709-4537-8f7a-4bdc31bf72f8","Type":"ContainerStarted","Data":"f2e79e2366ac826a1f1cae98a7daa7e630440202ed451efc341a3a6f3c803bd3"} Mar 07 14:51:52 crc kubenswrapper[4943]: I0307 14:51:52.187708 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-22p8b" event={"ID":"e8cc9b78-5dd4-43c6-a1fe-702f4f1c30c9","Type":"ContainerStarted","Data":"f4af5a4172235fbc96164ab8a067369deaab79025d5e75997b7ca08ce895e72c"} Mar 07 14:51:58 crc kubenswrapper[4943]: I0307 14:51:58.228528 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-86ddb6bd46-wxgzg" event={"ID":"31d1a300-cbac-4266-8182-300930377412","Type":"ContainerStarted","Data":"ce8d0fae5f23e46f1a6c5f14a9d483c46b7ac9e6f87ba421d5eccd497462537b"} Mar 07 14:51:58 crc kubenswrapper[4943]: I0307 14:51:58.231032 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-86ddb6bd46-wxgzg" Mar 07 14:51:58 crc kubenswrapper[4943]: I0307 14:51:58.423175 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-86ddb6bd46-wxgzg" podStartSLOduration=1.9645183080000002 podStartE2EDuration="9.423155004s" podCreationTimestamp="2026-03-07 14:51:49 +0000 UTC" firstStartedPulling="2026-03-07 14:51:50.318229553 +0000 UTC m=+752.270366061" lastFinishedPulling="2026-03-07 14:51:57.776866249 +0000 UTC m=+759.729002757" observedRunningTime="2026-03-07 14:51:58.416441247 +0000 UTC m=+760.368577745" watchObservedRunningTime="2026-03-07 14:51:58.423155004 +0000 UTC m=+760.375291502" Mar 07 14:51:59 crc kubenswrapper[4943]: I0307 14:51:59.238780 4943 generic.go:334] "Generic (PLEG): container finished" podID="19220a5d-633c-4051-87d4-246e167e6db5" containerID="6fe5bb85ef1c49401007b94214fa361e4355afae209dc424a519b523ec0a06d6" exitCode=0 Mar 07 14:51:59 crc kubenswrapper[4943]: I0307 14:51:59.238896 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ntxr6" event={"ID":"19220a5d-633c-4051-87d4-246e167e6db5","Type":"ContainerDied","Data":"6fe5bb85ef1c49401007b94214fa361e4355afae209dc424a519b523ec0a06d6"} Mar 07 14:51:59 crc kubenswrapper[4943]: I0307 14:51:59.240707 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-22p8b" event={"ID":"e8cc9b78-5dd4-43c6-a1fe-702f4f1c30c9","Type":"ContainerStarted","Data":"d4e857fd56cd948233a51548e9b606814cecf9b6ec7dce4068fcc0d17f050a0e"} Mar 07 14:51:59 crc kubenswrapper[4943]: I0307 14:51:59.241007 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-22p8b" Mar 07 14:51:59 crc kubenswrapper[4943]: I0307 14:51:59.243458 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7f989f654f-h29zk" event={"ID":"6e721eda-a709-4537-8f7a-4bdc31bf72f8","Type":"ContainerStarted","Data":"cfebf2f6be2013297f7bd0c1c4a401c2b6e87a2c72882d95c2717ccec1fb866f"} Mar 07 14:51:59 crc kubenswrapper[4943]: I0307 14:51:59.276384 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-22p8b" podStartSLOduration=3.933022127 podStartE2EDuration="10.276362108s" podCreationTimestamp="2026-03-07 14:51:49 +0000 UTC" firstStartedPulling="2026-03-07 14:51:51.503900176 +0000 UTC m=+753.456036674" lastFinishedPulling="2026-03-07 14:51:57.847240117 +0000 UTC m=+759.799376655" observedRunningTime="2026-03-07 14:51:59.275868566 +0000 UTC m=+761.228005084" watchObservedRunningTime="2026-03-07 14:51:59.276362108 +0000 UTC m=+761.228498606" Mar 07 14:51:59 crc kubenswrapper[4943]: I0307 14:51:59.294663 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7f989f654f-h29zk" podStartSLOduration=2.63049302 podStartE2EDuration="10.294638852s" podCreationTimestamp="2026-03-07 14:51:49 +0000 UTC" firstStartedPulling="2026-03-07 14:51:50.253004612 +0000 UTC m=+752.205141110" lastFinishedPulling="2026-03-07 14:51:57.917150434 +0000 UTC m=+759.869286942" observedRunningTime="2026-03-07 14:51:59.289988426 +0000 UTC m=+761.242124934" watchObservedRunningTime="2026-03-07 14:51:59.294638852 +0000 UTC m=+761.246775360" Mar 07 14:51:59 crc kubenswrapper[4943]: I0307 14:51:59.987445 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7f989f654f-h29zk" Mar 07 14:52:00 crc kubenswrapper[4943]: I0307 14:52:00.136735 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29548252-brvbf"] Mar 07 14:52:00 crc kubenswrapper[4943]: I0307 14:52:00.138288 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548252-brvbf" Mar 07 14:52:00 crc kubenswrapper[4943]: I0307 14:52:00.148880 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29548252-brvbf"] Mar 07 14:52:00 crc kubenswrapper[4943]: I0307 14:52:00.182239 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 07 14:52:00 crc kubenswrapper[4943]: I0307 14:52:00.184008 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 07 14:52:00 crc kubenswrapper[4943]: I0307 14:52:00.184159 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-qqlk5" Mar 07 14:52:00 crc kubenswrapper[4943]: I0307 14:52:00.240307 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9dsz\" (UniqueName: \"kubernetes.io/projected/ba71982c-c6e1-4247-8cf7-28661b6c4a36-kube-api-access-n9dsz\") pod \"auto-csr-approver-29548252-brvbf\" (UID: \"ba71982c-c6e1-4247-8cf7-28661b6c4a36\") " pod="openshift-infra/auto-csr-approver-29548252-brvbf" Mar 07 14:52:00 crc kubenswrapper[4943]: I0307 14:52:00.250468 4943 generic.go:334] "Generic (PLEG): container finished" podID="19220a5d-633c-4051-87d4-246e167e6db5" containerID="5cd73a27f0883ef79f45e0c7888e5e38f4b9aac54d5fac9ab33fb9e3ab3ce003" exitCode=0 Mar 07 14:52:00 crc kubenswrapper[4943]: I0307 14:52:00.251500 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ntxr6" event={"ID":"19220a5d-633c-4051-87d4-246e167e6db5","Type":"ContainerDied","Data":"5cd73a27f0883ef79f45e0c7888e5e38f4b9aac54d5fac9ab33fb9e3ab3ce003"} Mar 07 14:52:00 crc kubenswrapper[4943]: I0307 14:52:00.258480 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-86ddb6bd46-wxgzg" Mar 07 14:52:00 crc kubenswrapper[4943]: I0307 14:52:00.341082 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9dsz\" (UniqueName: \"kubernetes.io/projected/ba71982c-c6e1-4247-8cf7-28661b6c4a36-kube-api-access-n9dsz\") pod \"auto-csr-approver-29548252-brvbf\" (UID: \"ba71982c-c6e1-4247-8cf7-28661b6c4a36\") " pod="openshift-infra/auto-csr-approver-29548252-brvbf" Mar 07 14:52:00 crc kubenswrapper[4943]: I0307 14:52:00.362375 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9dsz\" (UniqueName: \"kubernetes.io/projected/ba71982c-c6e1-4247-8cf7-28661b6c4a36-kube-api-access-n9dsz\") pod \"auto-csr-approver-29548252-brvbf\" (UID: \"ba71982c-c6e1-4247-8cf7-28661b6c4a36\") " pod="openshift-infra/auto-csr-approver-29548252-brvbf" Mar 07 14:52:00 crc kubenswrapper[4943]: I0307 14:52:00.496335 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548252-brvbf" Mar 07 14:52:00 crc kubenswrapper[4943]: I0307 14:52:00.960178 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29548252-brvbf"] Mar 07 14:52:00 crc kubenswrapper[4943]: W0307 14:52:00.965781 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podba71982c_c6e1_4247_8cf7_28661b6c4a36.slice/crio-085e177491417f04ed7bf07c8648642a0c44d870f37c9b59c670de354a3f7022 WatchSource:0}: Error finding container 085e177491417f04ed7bf07c8648642a0c44d870f37c9b59c670de354a3f7022: Status 404 returned error can't find the container with id 085e177491417f04ed7bf07c8648642a0c44d870f37c9b59c670de354a3f7022 Mar 07 14:52:01 crc kubenswrapper[4943]: I0307 14:52:01.011755 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-22p8b" Mar 07 14:52:01 crc kubenswrapper[4943]: I0307 14:52:01.260708 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548252-brvbf" event={"ID":"ba71982c-c6e1-4247-8cf7-28661b6c4a36","Type":"ContainerStarted","Data":"085e177491417f04ed7bf07c8648642a0c44d870f37c9b59c670de354a3f7022"} Mar 07 14:52:01 crc kubenswrapper[4943]: I0307 14:52:01.264622 4943 generic.go:334] "Generic (PLEG): container finished" podID="19220a5d-633c-4051-87d4-246e167e6db5" containerID="db1192bae542b9571154f4bce11209baded51a5ed7c6799b9520555ffe441d49" exitCode=0 Mar 07 14:52:01 crc kubenswrapper[4943]: I0307 14:52:01.264788 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ntxr6" event={"ID":"19220a5d-633c-4051-87d4-246e167e6db5","Type":"ContainerDied","Data":"db1192bae542b9571154f4bce11209baded51a5ed7c6799b9520555ffe441d49"} Mar 07 14:52:02 crc kubenswrapper[4943]: I0307 14:52:02.279080 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ntxr6" event={"ID":"19220a5d-633c-4051-87d4-246e167e6db5","Type":"ContainerStarted","Data":"24af07ee08dd41c68c332f9ad5b6c7d4cff7f6ef9011009f3dd89dd6217e958b"} Mar 07 14:52:02 crc kubenswrapper[4943]: I0307 14:52:02.279505 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ntxr6" event={"ID":"19220a5d-633c-4051-87d4-246e167e6db5","Type":"ContainerStarted","Data":"a8f69ba207581037aba91d5112173db67772f20f4b56efd4ccd50e88c5315d09"} Mar 07 14:52:02 crc kubenswrapper[4943]: I0307 14:52:02.279527 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ntxr6" event={"ID":"19220a5d-633c-4051-87d4-246e167e6db5","Type":"ContainerStarted","Data":"f50b463437af7dd145ea9a42d092d6fd2b5faa85af7a3af80f2315934e7541cd"} Mar 07 14:52:02 crc kubenswrapper[4943]: I0307 14:52:02.279544 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ntxr6" event={"ID":"19220a5d-633c-4051-87d4-246e167e6db5","Type":"ContainerStarted","Data":"80f81c96761ac92caf45d49f78f4a87d990de4145d130e9e5e5c48da13a66e9f"} Mar 07 14:52:02 crc kubenswrapper[4943]: I0307 14:52:02.279561 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ntxr6" event={"ID":"19220a5d-633c-4051-87d4-246e167e6db5","Type":"ContainerStarted","Data":"0d83e47629f22b180f2006e8f0ed7f99e1be03c750bbb4db705af017abc69a0e"} Mar 07 14:52:02 crc kubenswrapper[4943]: I0307 14:52:02.279577 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ntxr6" event={"ID":"19220a5d-633c-4051-87d4-246e167e6db5","Type":"ContainerStarted","Data":"451a5095196842b20827e38f8546271d407420a970d561ad90c41f0b1de80dea"} Mar 07 14:52:02 crc kubenswrapper[4943]: I0307 14:52:02.279653 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-ntxr6" Mar 07 14:52:02 crc kubenswrapper[4943]: I0307 14:52:02.281493 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548252-brvbf" event={"ID":"ba71982c-c6e1-4247-8cf7-28661b6c4a36","Type":"ContainerStarted","Data":"b554df5bf661f81ff1a5aa8e723d7f0fdf8d057fec6e88cd52bf920ec8b1f872"} Mar 07 14:52:02 crc kubenswrapper[4943]: I0307 14:52:02.306619 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-ntxr6" podStartSLOduration=5.09223076 podStartE2EDuration="13.30659351s" podCreationTimestamp="2026-03-07 14:51:49 +0000 UTC" firstStartedPulling="2026-03-07 14:51:49.734618435 +0000 UTC m=+751.686754963" lastFinishedPulling="2026-03-07 14:51:57.948981165 +0000 UTC m=+759.901117713" observedRunningTime="2026-03-07 14:52:02.301068623 +0000 UTC m=+764.253205121" watchObservedRunningTime="2026-03-07 14:52:02.30659351 +0000 UTC m=+764.258730038" Mar 07 14:52:02 crc kubenswrapper[4943]: I0307 14:52:02.322182 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29548252-brvbf" podStartSLOduration=1.3748934369999999 podStartE2EDuration="2.322158577s" podCreationTimestamp="2026-03-07 14:52:00 +0000 UTC" firstStartedPulling="2026-03-07 14:52:00.969406824 +0000 UTC m=+762.921543332" lastFinishedPulling="2026-03-07 14:52:01.916671984 +0000 UTC m=+763.868808472" observedRunningTime="2026-03-07 14:52:02.319189803 +0000 UTC m=+764.271326341" watchObservedRunningTime="2026-03-07 14:52:02.322158577 +0000 UTC m=+764.274295085" Mar 07 14:52:03 crc kubenswrapper[4943]: I0307 14:52:03.291375 4943 generic.go:334] "Generic (PLEG): container finished" podID="ba71982c-c6e1-4247-8cf7-28661b6c4a36" containerID="b554df5bf661f81ff1a5aa8e723d7f0fdf8d057fec6e88cd52bf920ec8b1f872" exitCode=0 Mar 07 14:52:03 crc kubenswrapper[4943]: I0307 14:52:03.291486 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548252-brvbf" event={"ID":"ba71982c-c6e1-4247-8cf7-28661b6c4a36","Type":"ContainerDied","Data":"b554df5bf661f81ff1a5aa8e723d7f0fdf8d057fec6e88cd52bf920ec8b1f872"} Mar 07 14:52:04 crc kubenswrapper[4943]: I0307 14:52:04.377116 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-ntxr6" Mar 07 14:52:04 crc kubenswrapper[4943]: I0307 14:52:04.427198 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-ntxr6" Mar 07 14:52:04 crc kubenswrapper[4943]: I0307 14:52:04.606582 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548252-brvbf" Mar 07 14:52:04 crc kubenswrapper[4943]: I0307 14:52:04.707945 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n9dsz\" (UniqueName: \"kubernetes.io/projected/ba71982c-c6e1-4247-8cf7-28661b6c4a36-kube-api-access-n9dsz\") pod \"ba71982c-c6e1-4247-8cf7-28661b6c4a36\" (UID: \"ba71982c-c6e1-4247-8cf7-28661b6c4a36\") " Mar 07 14:52:04 crc kubenswrapper[4943]: I0307 14:52:04.714411 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba71982c-c6e1-4247-8cf7-28661b6c4a36-kube-api-access-n9dsz" (OuterVolumeSpecName: "kube-api-access-n9dsz") pod "ba71982c-c6e1-4247-8cf7-28661b6c4a36" (UID: "ba71982c-c6e1-4247-8cf7-28661b6c4a36"). InnerVolumeSpecName "kube-api-access-n9dsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:52:04 crc kubenswrapper[4943]: I0307 14:52:04.809815 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n9dsz\" (UniqueName: \"kubernetes.io/projected/ba71982c-c6e1-4247-8cf7-28661b6c4a36-kube-api-access-n9dsz\") on node \"crc\" DevicePath \"\"" Mar 07 14:52:05 crc kubenswrapper[4943]: I0307 14:52:05.311781 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548252-brvbf" Mar 07 14:52:05 crc kubenswrapper[4943]: I0307 14:52:05.311790 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548252-brvbf" event={"ID":"ba71982c-c6e1-4247-8cf7-28661b6c4a36","Type":"ContainerDied","Data":"085e177491417f04ed7bf07c8648642a0c44d870f37c9b59c670de354a3f7022"} Mar 07 14:52:05 crc kubenswrapper[4943]: I0307 14:52:05.312269 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="085e177491417f04ed7bf07c8648642a0c44d870f37c9b59c670de354a3f7022" Mar 07 14:52:05 crc kubenswrapper[4943]: I0307 14:52:05.685373 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29548246-r7df9"] Mar 07 14:52:05 crc kubenswrapper[4943]: I0307 14:52:05.691917 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29548246-r7df9"] Mar 07 14:52:06 crc kubenswrapper[4943]: I0307 14:52:06.074012 4943 patch_prober.go:28] interesting pod/machine-config-daemon-cpdmm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 07 14:52:06 crc kubenswrapper[4943]: I0307 14:52:06.074115 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 07 14:52:06 crc kubenswrapper[4943]: I0307 14:52:06.627331 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-index-qvwsn"] Mar 07 14:52:06 crc kubenswrapper[4943]: E0307 14:52:06.627739 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba71982c-c6e1-4247-8cf7-28661b6c4a36" containerName="oc" Mar 07 14:52:06 crc kubenswrapper[4943]: I0307 14:52:06.627780 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba71982c-c6e1-4247-8cf7-28661b6c4a36" containerName="oc" Mar 07 14:52:06 crc kubenswrapper[4943]: I0307 14:52:06.628119 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba71982c-c6e1-4247-8cf7-28661b6c4a36" containerName="oc" Mar 07 14:52:06 crc kubenswrapper[4943]: I0307 14:52:06.629026 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-qvwsn" Mar 07 14:52:06 crc kubenswrapper[4943]: I0307 14:52:06.632227 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-index-dockercfg-sp58l" Mar 07 14:52:06 crc kubenswrapper[4943]: I0307 14:52:06.632292 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Mar 07 14:52:06 crc kubenswrapper[4943]: I0307 14:52:06.633405 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Mar 07 14:52:06 crc kubenswrapper[4943]: I0307 14:52:06.637777 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9nqm2\" (UniqueName: \"kubernetes.io/projected/4e86126d-9d45-48ba-b5e2-ce3c77eec3e6-kube-api-access-9nqm2\") pod \"mariadb-operator-index-qvwsn\" (UID: \"4e86126d-9d45-48ba-b5e2-ce3c77eec3e6\") " pod="openstack-operators/mariadb-operator-index-qvwsn" Mar 07 14:52:06 crc kubenswrapper[4943]: I0307 14:52:06.638339 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-qvwsn"] Mar 07 14:52:06 crc kubenswrapper[4943]: I0307 14:52:06.739298 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9nqm2\" (UniqueName: \"kubernetes.io/projected/4e86126d-9d45-48ba-b5e2-ce3c77eec3e6-kube-api-access-9nqm2\") pod \"mariadb-operator-index-qvwsn\" (UID: \"4e86126d-9d45-48ba-b5e2-ce3c77eec3e6\") " pod="openstack-operators/mariadb-operator-index-qvwsn" Mar 07 14:52:06 crc kubenswrapper[4943]: I0307 14:52:06.765074 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5dceb930-7f1e-49cd-a53d-db7045ab6984" path="/var/lib/kubelet/pods/5dceb930-7f1e-49cd-a53d-db7045ab6984/volumes" Mar 07 14:52:06 crc kubenswrapper[4943]: I0307 14:52:06.766290 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9nqm2\" (UniqueName: \"kubernetes.io/projected/4e86126d-9d45-48ba-b5e2-ce3c77eec3e6-kube-api-access-9nqm2\") pod \"mariadb-operator-index-qvwsn\" (UID: \"4e86126d-9d45-48ba-b5e2-ce3c77eec3e6\") " pod="openstack-operators/mariadb-operator-index-qvwsn" Mar 07 14:52:06 crc kubenswrapper[4943]: I0307 14:52:06.987220 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-qvwsn" Mar 07 14:52:07 crc kubenswrapper[4943]: I0307 14:52:07.414410 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-qvwsn"] Mar 07 14:52:08 crc kubenswrapper[4943]: I0307 14:52:08.338539 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-qvwsn" event={"ID":"4e86126d-9d45-48ba-b5e2-ce3c77eec3e6","Type":"ContainerStarted","Data":"bbedbf9de4dee27a531ed88841f581b6f8e140c018538a05c32ec6cb7f02c173"} Mar 07 14:52:09 crc kubenswrapper[4943]: I0307 14:52:09.359624 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-qvwsn" event={"ID":"4e86126d-9d45-48ba-b5e2-ce3c77eec3e6","Type":"ContainerStarted","Data":"cab005ddc3f15752f2d7a58858947256457e246fa4b8e19e3c74c38ea70de625"} Mar 07 14:52:09 crc kubenswrapper[4943]: I0307 14:52:09.382314 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-index-qvwsn" podStartSLOduration=2.61922598 podStartE2EDuration="3.382289745s" podCreationTimestamp="2026-03-07 14:52:06 +0000 UTC" firstStartedPulling="2026-03-07 14:52:07.440167031 +0000 UTC m=+769.392303519" lastFinishedPulling="2026-03-07 14:52:08.203230786 +0000 UTC m=+770.155367284" observedRunningTime="2026-03-07 14:52:09.377714381 +0000 UTC m=+771.329850889" watchObservedRunningTime="2026-03-07 14:52:09.382289745 +0000 UTC m=+771.334426243" Mar 07 14:52:09 crc kubenswrapper[4943]: I0307 14:52:09.998303 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7f989f654f-h29zk" Mar 07 14:52:10 crc kubenswrapper[4943]: I0307 14:52:10.000027 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-qvwsn"] Mar 07 14:52:10 crc kubenswrapper[4943]: I0307 14:52:10.606309 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-index-qdw88"] Mar 07 14:52:10 crc kubenswrapper[4943]: I0307 14:52:10.607443 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-qdw88" Mar 07 14:52:10 crc kubenswrapper[4943]: I0307 14:52:10.621251 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-qdw88"] Mar 07 14:52:10 crc kubenswrapper[4943]: I0307 14:52:10.703061 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x759q\" (UniqueName: \"kubernetes.io/projected/6562db67-0f59-4eab-876f-a60272522e0d-kube-api-access-x759q\") pod \"mariadb-operator-index-qdw88\" (UID: \"6562db67-0f59-4eab-876f-a60272522e0d\") " pod="openstack-operators/mariadb-operator-index-qdw88" Mar 07 14:52:10 crc kubenswrapper[4943]: I0307 14:52:10.804760 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x759q\" (UniqueName: \"kubernetes.io/projected/6562db67-0f59-4eab-876f-a60272522e0d-kube-api-access-x759q\") pod \"mariadb-operator-index-qdw88\" (UID: \"6562db67-0f59-4eab-876f-a60272522e0d\") " pod="openstack-operators/mariadb-operator-index-qdw88" Mar 07 14:52:10 crc kubenswrapper[4943]: I0307 14:52:10.845288 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x759q\" (UniqueName: \"kubernetes.io/projected/6562db67-0f59-4eab-876f-a60272522e0d-kube-api-access-x759q\") pod \"mariadb-operator-index-qdw88\" (UID: \"6562db67-0f59-4eab-876f-a60272522e0d\") " pod="openstack-operators/mariadb-operator-index-qdw88" Mar 07 14:52:10 crc kubenswrapper[4943]: I0307 14:52:10.927916 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-qdw88" Mar 07 14:52:11 crc kubenswrapper[4943]: I0307 14:52:11.268899 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-qdw88"] Mar 07 14:52:11 crc kubenswrapper[4943]: I0307 14:52:11.375608 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-qdw88" event={"ID":"6562db67-0f59-4eab-876f-a60272522e0d","Type":"ContainerStarted","Data":"c69f1fb307a62c95859a40f535082ab312881edb8723012cde1a61c514986759"} Mar 07 14:52:11 crc kubenswrapper[4943]: I0307 14:52:11.375817 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/mariadb-operator-index-qvwsn" podUID="4e86126d-9d45-48ba-b5e2-ce3c77eec3e6" containerName="registry-server" containerID="cri-o://cab005ddc3f15752f2d7a58858947256457e246fa4b8e19e3c74c38ea70de625" gracePeriod=2 Mar 07 14:52:11 crc kubenswrapper[4943]: I0307 14:52:11.725017 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-qvwsn" Mar 07 14:52:11 crc kubenswrapper[4943]: I0307 14:52:11.836085 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9nqm2\" (UniqueName: \"kubernetes.io/projected/4e86126d-9d45-48ba-b5e2-ce3c77eec3e6-kube-api-access-9nqm2\") pod \"4e86126d-9d45-48ba-b5e2-ce3c77eec3e6\" (UID: \"4e86126d-9d45-48ba-b5e2-ce3c77eec3e6\") " Mar 07 14:52:11 crc kubenswrapper[4943]: I0307 14:52:11.882140 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e86126d-9d45-48ba-b5e2-ce3c77eec3e6-kube-api-access-9nqm2" (OuterVolumeSpecName: "kube-api-access-9nqm2") pod "4e86126d-9d45-48ba-b5e2-ce3c77eec3e6" (UID: "4e86126d-9d45-48ba-b5e2-ce3c77eec3e6"). InnerVolumeSpecName "kube-api-access-9nqm2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:52:11 crc kubenswrapper[4943]: I0307 14:52:11.937668 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9nqm2\" (UniqueName: \"kubernetes.io/projected/4e86126d-9d45-48ba-b5e2-ce3c77eec3e6-kube-api-access-9nqm2\") on node \"crc\" DevicePath \"\"" Mar 07 14:52:12 crc kubenswrapper[4943]: I0307 14:52:12.386602 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-qdw88" event={"ID":"6562db67-0f59-4eab-876f-a60272522e0d","Type":"ContainerStarted","Data":"47f1e2687694bcaa01074f68816850e2d2e1ba6ae58328aca5af99fc60504d9d"} Mar 07 14:52:12 crc kubenswrapper[4943]: I0307 14:52:12.389518 4943 generic.go:334] "Generic (PLEG): container finished" podID="4e86126d-9d45-48ba-b5e2-ce3c77eec3e6" containerID="cab005ddc3f15752f2d7a58858947256457e246fa4b8e19e3c74c38ea70de625" exitCode=0 Mar 07 14:52:12 crc kubenswrapper[4943]: I0307 14:52:12.389576 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-qvwsn" event={"ID":"4e86126d-9d45-48ba-b5e2-ce3c77eec3e6","Type":"ContainerDied","Data":"cab005ddc3f15752f2d7a58858947256457e246fa4b8e19e3c74c38ea70de625"} Mar 07 14:52:12 crc kubenswrapper[4943]: I0307 14:52:12.389628 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-qvwsn" Mar 07 14:52:12 crc kubenswrapper[4943]: I0307 14:52:12.389662 4943 scope.go:117] "RemoveContainer" containerID="cab005ddc3f15752f2d7a58858947256457e246fa4b8e19e3c74c38ea70de625" Mar 07 14:52:12 crc kubenswrapper[4943]: I0307 14:52:12.389640 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-qvwsn" event={"ID":"4e86126d-9d45-48ba-b5e2-ce3c77eec3e6","Type":"ContainerDied","Data":"bbedbf9de4dee27a531ed88841f581b6f8e140c018538a05c32ec6cb7f02c173"} Mar 07 14:52:12 crc kubenswrapper[4943]: I0307 14:52:12.412521 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-index-qdw88" podStartSLOduration=1.978978969 podStartE2EDuration="2.412490787s" podCreationTimestamp="2026-03-07 14:52:10 +0000 UTC" firstStartedPulling="2026-03-07 14:52:11.283325398 +0000 UTC m=+773.235461896" lastFinishedPulling="2026-03-07 14:52:11.716837216 +0000 UTC m=+773.668973714" observedRunningTime="2026-03-07 14:52:12.410964859 +0000 UTC m=+774.363101397" watchObservedRunningTime="2026-03-07 14:52:12.412490787 +0000 UTC m=+774.364627315" Mar 07 14:52:12 crc kubenswrapper[4943]: I0307 14:52:12.416174 4943 scope.go:117] "RemoveContainer" containerID="cab005ddc3f15752f2d7a58858947256457e246fa4b8e19e3c74c38ea70de625" Mar 07 14:52:12 crc kubenswrapper[4943]: E0307 14:52:12.416857 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cab005ddc3f15752f2d7a58858947256457e246fa4b8e19e3c74c38ea70de625\": container with ID starting with cab005ddc3f15752f2d7a58858947256457e246fa4b8e19e3c74c38ea70de625 not found: ID does not exist" containerID="cab005ddc3f15752f2d7a58858947256457e246fa4b8e19e3c74c38ea70de625" Mar 07 14:52:12 crc kubenswrapper[4943]: I0307 14:52:12.416949 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cab005ddc3f15752f2d7a58858947256457e246fa4b8e19e3c74c38ea70de625"} err="failed to get container status \"cab005ddc3f15752f2d7a58858947256457e246fa4b8e19e3c74c38ea70de625\": rpc error: code = NotFound desc = could not find container \"cab005ddc3f15752f2d7a58858947256457e246fa4b8e19e3c74c38ea70de625\": container with ID starting with cab005ddc3f15752f2d7a58858947256457e246fa4b8e19e3c74c38ea70de625 not found: ID does not exist" Mar 07 14:52:12 crc kubenswrapper[4943]: I0307 14:52:12.437868 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-qvwsn"] Mar 07 14:52:12 crc kubenswrapper[4943]: I0307 14:52:12.443850 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/mariadb-operator-index-qvwsn"] Mar 07 14:52:12 crc kubenswrapper[4943]: I0307 14:52:12.766509 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e86126d-9d45-48ba-b5e2-ce3c77eec3e6" path="/var/lib/kubelet/pods/4e86126d-9d45-48ba-b5e2-ce3c77eec3e6/volumes" Mar 07 14:52:19 crc kubenswrapper[4943]: I0307 14:52:19.385465 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-ntxr6" Mar 07 14:52:19 crc kubenswrapper[4943]: I0307 14:52:19.428169 4943 scope.go:117] "RemoveContainer" containerID="0571265e58bbf9b45930119ab783ac8030ba47a1e31fd5d62b1f1d9f596b4d55" Mar 07 14:52:20 crc kubenswrapper[4943]: I0307 14:52:20.928483 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/mariadb-operator-index-qdw88" Mar 07 14:52:20 crc kubenswrapper[4943]: I0307 14:52:20.928901 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-index-qdw88" Mar 07 14:52:20 crc kubenswrapper[4943]: I0307 14:52:20.977073 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/mariadb-operator-index-qdw88" Mar 07 14:52:21 crc kubenswrapper[4943]: I0307 14:52:21.514049 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-index-qdw88" Mar 07 14:52:23 crc kubenswrapper[4943]: I0307 14:52:23.050334 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da67slpc"] Mar 07 14:52:23 crc kubenswrapper[4943]: E0307 14:52:23.051075 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e86126d-9d45-48ba-b5e2-ce3c77eec3e6" containerName="registry-server" Mar 07 14:52:23 crc kubenswrapper[4943]: I0307 14:52:23.051098 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e86126d-9d45-48ba-b5e2-ce3c77eec3e6" containerName="registry-server" Mar 07 14:52:23 crc kubenswrapper[4943]: I0307 14:52:23.051307 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e86126d-9d45-48ba-b5e2-ce3c77eec3e6" containerName="registry-server" Mar 07 14:52:23 crc kubenswrapper[4943]: I0307 14:52:23.052779 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da67slpc" Mar 07 14:52:23 crc kubenswrapper[4943]: I0307 14:52:23.056232 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-k6gp2" Mar 07 14:52:23 crc kubenswrapper[4943]: I0307 14:52:23.108078 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da67slpc"] Mar 07 14:52:23 crc kubenswrapper[4943]: I0307 14:52:23.225598 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7g58\" (UniqueName: \"kubernetes.io/projected/a880ad69-0f85-45ec-a5f2-e27ce2ce850f-kube-api-access-s7g58\") pod \"b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da67slpc\" (UID: \"a880ad69-0f85-45ec-a5f2-e27ce2ce850f\") " pod="openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da67slpc" Mar 07 14:52:23 crc kubenswrapper[4943]: I0307 14:52:23.225995 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a880ad69-0f85-45ec-a5f2-e27ce2ce850f-util\") pod \"b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da67slpc\" (UID: \"a880ad69-0f85-45ec-a5f2-e27ce2ce850f\") " pod="openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da67slpc" Mar 07 14:52:23 crc kubenswrapper[4943]: I0307 14:52:23.226284 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a880ad69-0f85-45ec-a5f2-e27ce2ce850f-bundle\") pod \"b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da67slpc\" (UID: \"a880ad69-0f85-45ec-a5f2-e27ce2ce850f\") " pod="openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da67slpc" Mar 07 14:52:23 crc kubenswrapper[4943]: I0307 14:52:23.327338 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7g58\" (UniqueName: \"kubernetes.io/projected/a880ad69-0f85-45ec-a5f2-e27ce2ce850f-kube-api-access-s7g58\") pod \"b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da67slpc\" (UID: \"a880ad69-0f85-45ec-a5f2-e27ce2ce850f\") " pod="openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da67slpc" Mar 07 14:52:23 crc kubenswrapper[4943]: I0307 14:52:23.327731 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a880ad69-0f85-45ec-a5f2-e27ce2ce850f-util\") pod \"b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da67slpc\" (UID: \"a880ad69-0f85-45ec-a5f2-e27ce2ce850f\") " pod="openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da67slpc" Mar 07 14:52:23 crc kubenswrapper[4943]: I0307 14:52:23.328078 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a880ad69-0f85-45ec-a5f2-e27ce2ce850f-bundle\") pod \"b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da67slpc\" (UID: \"a880ad69-0f85-45ec-a5f2-e27ce2ce850f\") " pod="openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da67slpc" Mar 07 14:52:23 crc kubenswrapper[4943]: I0307 14:52:23.328737 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a880ad69-0f85-45ec-a5f2-e27ce2ce850f-util\") pod \"b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da67slpc\" (UID: \"a880ad69-0f85-45ec-a5f2-e27ce2ce850f\") " pod="openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da67slpc" Mar 07 14:52:23 crc kubenswrapper[4943]: I0307 14:52:23.328836 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a880ad69-0f85-45ec-a5f2-e27ce2ce850f-bundle\") pod \"b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da67slpc\" (UID: \"a880ad69-0f85-45ec-a5f2-e27ce2ce850f\") " pod="openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da67slpc" Mar 07 14:52:23 crc kubenswrapper[4943]: I0307 14:52:23.361605 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7g58\" (UniqueName: \"kubernetes.io/projected/a880ad69-0f85-45ec-a5f2-e27ce2ce850f-kube-api-access-s7g58\") pod \"b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da67slpc\" (UID: \"a880ad69-0f85-45ec-a5f2-e27ce2ce850f\") " pod="openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da67slpc" Mar 07 14:52:23 crc kubenswrapper[4943]: I0307 14:52:23.379254 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da67slpc" Mar 07 14:52:23 crc kubenswrapper[4943]: I0307 14:52:23.675643 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da67slpc"] Mar 07 14:52:23 crc kubenswrapper[4943]: W0307 14:52:23.682638 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda880ad69_0f85_45ec_a5f2_e27ce2ce850f.slice/crio-557b7420f4ddb95245cf5d33321cf55ae4561898e575f4e47f1f0d1f3b4d827c WatchSource:0}: Error finding container 557b7420f4ddb95245cf5d33321cf55ae4561898e575f4e47f1f0d1f3b4d827c: Status 404 returned error can't find the container with id 557b7420f4ddb95245cf5d33321cf55ae4561898e575f4e47f1f0d1f3b4d827c Mar 07 14:52:24 crc kubenswrapper[4943]: I0307 14:52:24.509049 4943 generic.go:334] "Generic (PLEG): container finished" podID="a880ad69-0f85-45ec-a5f2-e27ce2ce850f" containerID="f89d1d0b7a1d170f5a66608654c0e2a3f5a0bed64f5308b7b91fdc87ad5a376b" exitCode=0 Mar 07 14:52:24 crc kubenswrapper[4943]: I0307 14:52:24.509167 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da67slpc" event={"ID":"a880ad69-0f85-45ec-a5f2-e27ce2ce850f","Type":"ContainerDied","Data":"f89d1d0b7a1d170f5a66608654c0e2a3f5a0bed64f5308b7b91fdc87ad5a376b"} Mar 07 14:52:24 crc kubenswrapper[4943]: I0307 14:52:24.512723 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da67slpc" event={"ID":"a880ad69-0f85-45ec-a5f2-e27ce2ce850f","Type":"ContainerStarted","Data":"557b7420f4ddb95245cf5d33321cf55ae4561898e575f4e47f1f0d1f3b4d827c"} Mar 07 14:52:25 crc kubenswrapper[4943]: I0307 14:52:25.520701 4943 generic.go:334] "Generic (PLEG): container finished" podID="a880ad69-0f85-45ec-a5f2-e27ce2ce850f" containerID="e16436716b483992bd013d84704ab8fa3b7d593ecac82753c2d567e96220abda" exitCode=0 Mar 07 14:52:25 crc kubenswrapper[4943]: I0307 14:52:25.520805 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da67slpc" event={"ID":"a880ad69-0f85-45ec-a5f2-e27ce2ce850f","Type":"ContainerDied","Data":"e16436716b483992bd013d84704ab8fa3b7d593ecac82753c2d567e96220abda"} Mar 07 14:52:26 crc kubenswrapper[4943]: I0307 14:52:26.532367 4943 generic.go:334] "Generic (PLEG): container finished" podID="a880ad69-0f85-45ec-a5f2-e27ce2ce850f" containerID="ff8b00b8606157d4f8a82c79f995339be56690e15486e6470b67818ea2cec322" exitCode=0 Mar 07 14:52:26 crc kubenswrapper[4943]: I0307 14:52:26.532429 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da67slpc" event={"ID":"a880ad69-0f85-45ec-a5f2-e27ce2ce850f","Type":"ContainerDied","Data":"ff8b00b8606157d4f8a82c79f995339be56690e15486e6470b67818ea2cec322"} Mar 07 14:52:27 crc kubenswrapper[4943]: I0307 14:52:27.792998 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da67slpc" Mar 07 14:52:27 crc kubenswrapper[4943]: I0307 14:52:27.896699 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a880ad69-0f85-45ec-a5f2-e27ce2ce850f-bundle\") pod \"a880ad69-0f85-45ec-a5f2-e27ce2ce850f\" (UID: \"a880ad69-0f85-45ec-a5f2-e27ce2ce850f\") " Mar 07 14:52:27 crc kubenswrapper[4943]: I0307 14:52:27.896922 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s7g58\" (UniqueName: \"kubernetes.io/projected/a880ad69-0f85-45ec-a5f2-e27ce2ce850f-kube-api-access-s7g58\") pod \"a880ad69-0f85-45ec-a5f2-e27ce2ce850f\" (UID: \"a880ad69-0f85-45ec-a5f2-e27ce2ce850f\") " Mar 07 14:52:27 crc kubenswrapper[4943]: I0307 14:52:27.896987 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a880ad69-0f85-45ec-a5f2-e27ce2ce850f-util\") pod \"a880ad69-0f85-45ec-a5f2-e27ce2ce850f\" (UID: \"a880ad69-0f85-45ec-a5f2-e27ce2ce850f\") " Mar 07 14:52:27 crc kubenswrapper[4943]: I0307 14:52:27.899314 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a880ad69-0f85-45ec-a5f2-e27ce2ce850f-bundle" (OuterVolumeSpecName: "bundle") pod "a880ad69-0f85-45ec-a5f2-e27ce2ce850f" (UID: "a880ad69-0f85-45ec-a5f2-e27ce2ce850f"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:52:27 crc kubenswrapper[4943]: I0307 14:52:27.918008 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a880ad69-0f85-45ec-a5f2-e27ce2ce850f-kube-api-access-s7g58" (OuterVolumeSpecName: "kube-api-access-s7g58") pod "a880ad69-0f85-45ec-a5f2-e27ce2ce850f" (UID: "a880ad69-0f85-45ec-a5f2-e27ce2ce850f"). InnerVolumeSpecName "kube-api-access-s7g58". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:52:27 crc kubenswrapper[4943]: I0307 14:52:27.926509 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a880ad69-0f85-45ec-a5f2-e27ce2ce850f-util" (OuterVolumeSpecName: "util") pod "a880ad69-0f85-45ec-a5f2-e27ce2ce850f" (UID: "a880ad69-0f85-45ec-a5f2-e27ce2ce850f"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:52:27 crc kubenswrapper[4943]: I0307 14:52:27.999279 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s7g58\" (UniqueName: \"kubernetes.io/projected/a880ad69-0f85-45ec-a5f2-e27ce2ce850f-kube-api-access-s7g58\") on node \"crc\" DevicePath \"\"" Mar 07 14:52:27 crc kubenswrapper[4943]: I0307 14:52:27.999327 4943 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a880ad69-0f85-45ec-a5f2-e27ce2ce850f-util\") on node \"crc\" DevicePath \"\"" Mar 07 14:52:27 crc kubenswrapper[4943]: I0307 14:52:27.999346 4943 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a880ad69-0f85-45ec-a5f2-e27ce2ce850f-bundle\") on node \"crc\" DevicePath \"\"" Mar 07 14:52:28 crc kubenswrapper[4943]: I0307 14:52:28.549589 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da67slpc" event={"ID":"a880ad69-0f85-45ec-a5f2-e27ce2ce850f","Type":"ContainerDied","Data":"557b7420f4ddb95245cf5d33321cf55ae4561898e575f4e47f1f0d1f3b4d827c"} Mar 07 14:52:28 crc kubenswrapper[4943]: I0307 14:52:28.550002 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="557b7420f4ddb95245cf5d33321cf55ae4561898e575f4e47f1f0d1f3b4d827c" Mar 07 14:52:28 crc kubenswrapper[4943]: I0307 14:52:28.549712 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da67slpc" Mar 07 14:52:36 crc kubenswrapper[4943]: I0307 14:52:36.074766 4943 patch_prober.go:28] interesting pod/machine-config-daemon-cpdmm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 07 14:52:36 crc kubenswrapper[4943]: I0307 14:52:36.076075 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 07 14:52:36 crc kubenswrapper[4943]: I0307 14:52:36.360306 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-749f4cd94-8k4nn"] Mar 07 14:52:36 crc kubenswrapper[4943]: E0307 14:52:36.360637 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a880ad69-0f85-45ec-a5f2-e27ce2ce850f" containerName="util" Mar 07 14:52:36 crc kubenswrapper[4943]: I0307 14:52:36.360656 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="a880ad69-0f85-45ec-a5f2-e27ce2ce850f" containerName="util" Mar 07 14:52:36 crc kubenswrapper[4943]: E0307 14:52:36.360673 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a880ad69-0f85-45ec-a5f2-e27ce2ce850f" containerName="pull" Mar 07 14:52:36 crc kubenswrapper[4943]: I0307 14:52:36.360683 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="a880ad69-0f85-45ec-a5f2-e27ce2ce850f" containerName="pull" Mar 07 14:52:36 crc kubenswrapper[4943]: E0307 14:52:36.360702 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a880ad69-0f85-45ec-a5f2-e27ce2ce850f" containerName="extract" Mar 07 14:52:36 crc kubenswrapper[4943]: I0307 14:52:36.360712 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="a880ad69-0f85-45ec-a5f2-e27ce2ce850f" containerName="extract" Mar 07 14:52:36 crc kubenswrapper[4943]: I0307 14:52:36.360876 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="a880ad69-0f85-45ec-a5f2-e27ce2ce850f" containerName="extract" Mar 07 14:52:36 crc kubenswrapper[4943]: I0307 14:52:36.361485 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-749f4cd94-8k4nn" Mar 07 14:52:36 crc kubenswrapper[4943]: I0307 14:52:36.364697 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-szt8c" Mar 07 14:52:36 crc kubenswrapper[4943]: I0307 14:52:36.365078 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Mar 07 14:52:36 crc kubenswrapper[4943]: I0307 14:52:36.366839 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-service-cert" Mar 07 14:52:36 crc kubenswrapper[4943]: I0307 14:52:36.377781 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-749f4cd94-8k4nn"] Mar 07 14:52:36 crc kubenswrapper[4943]: I0307 14:52:36.455601 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9d6f004b-a409-4960-b377-8e8577597b99-apiservice-cert\") pod \"mariadb-operator-controller-manager-749f4cd94-8k4nn\" (UID: \"9d6f004b-a409-4960-b377-8e8577597b99\") " pod="openstack-operators/mariadb-operator-controller-manager-749f4cd94-8k4nn" Mar 07 14:52:36 crc kubenswrapper[4943]: I0307 14:52:36.455677 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsh66\" (UniqueName: \"kubernetes.io/projected/9d6f004b-a409-4960-b377-8e8577597b99-kube-api-access-jsh66\") pod \"mariadb-operator-controller-manager-749f4cd94-8k4nn\" (UID: \"9d6f004b-a409-4960-b377-8e8577597b99\") " pod="openstack-operators/mariadb-operator-controller-manager-749f4cd94-8k4nn" Mar 07 14:52:36 crc kubenswrapper[4943]: I0307 14:52:36.455765 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9d6f004b-a409-4960-b377-8e8577597b99-webhook-cert\") pod \"mariadb-operator-controller-manager-749f4cd94-8k4nn\" (UID: \"9d6f004b-a409-4960-b377-8e8577597b99\") " pod="openstack-operators/mariadb-operator-controller-manager-749f4cd94-8k4nn" Mar 07 14:52:36 crc kubenswrapper[4943]: I0307 14:52:36.557447 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9d6f004b-a409-4960-b377-8e8577597b99-apiservice-cert\") pod \"mariadb-operator-controller-manager-749f4cd94-8k4nn\" (UID: \"9d6f004b-a409-4960-b377-8e8577597b99\") " pod="openstack-operators/mariadb-operator-controller-manager-749f4cd94-8k4nn" Mar 07 14:52:36 crc kubenswrapper[4943]: I0307 14:52:36.557539 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsh66\" (UniqueName: \"kubernetes.io/projected/9d6f004b-a409-4960-b377-8e8577597b99-kube-api-access-jsh66\") pod \"mariadb-operator-controller-manager-749f4cd94-8k4nn\" (UID: \"9d6f004b-a409-4960-b377-8e8577597b99\") " pod="openstack-operators/mariadb-operator-controller-manager-749f4cd94-8k4nn" Mar 07 14:52:36 crc kubenswrapper[4943]: I0307 14:52:36.557611 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9d6f004b-a409-4960-b377-8e8577597b99-webhook-cert\") pod \"mariadb-operator-controller-manager-749f4cd94-8k4nn\" (UID: \"9d6f004b-a409-4960-b377-8e8577597b99\") " pod="openstack-operators/mariadb-operator-controller-manager-749f4cd94-8k4nn" Mar 07 14:52:36 crc kubenswrapper[4943]: I0307 14:52:36.564343 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9d6f004b-a409-4960-b377-8e8577597b99-apiservice-cert\") pod \"mariadb-operator-controller-manager-749f4cd94-8k4nn\" (UID: \"9d6f004b-a409-4960-b377-8e8577597b99\") " pod="openstack-operators/mariadb-operator-controller-manager-749f4cd94-8k4nn" Mar 07 14:52:36 crc kubenswrapper[4943]: I0307 14:52:36.564648 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9d6f004b-a409-4960-b377-8e8577597b99-webhook-cert\") pod \"mariadb-operator-controller-manager-749f4cd94-8k4nn\" (UID: \"9d6f004b-a409-4960-b377-8e8577597b99\") " pod="openstack-operators/mariadb-operator-controller-manager-749f4cd94-8k4nn" Mar 07 14:52:36 crc kubenswrapper[4943]: I0307 14:52:36.581709 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsh66\" (UniqueName: \"kubernetes.io/projected/9d6f004b-a409-4960-b377-8e8577597b99-kube-api-access-jsh66\") pod \"mariadb-operator-controller-manager-749f4cd94-8k4nn\" (UID: \"9d6f004b-a409-4960-b377-8e8577597b99\") " pod="openstack-operators/mariadb-operator-controller-manager-749f4cd94-8k4nn" Mar 07 14:52:36 crc kubenswrapper[4943]: I0307 14:52:36.680228 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-749f4cd94-8k4nn" Mar 07 14:52:37 crc kubenswrapper[4943]: I0307 14:52:37.110029 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-749f4cd94-8k4nn"] Mar 07 14:52:37 crc kubenswrapper[4943]: W0307 14:52:37.120433 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d6f004b_a409_4960_b377_8e8577597b99.slice/crio-a1ea1d343020e82799c79e2482d03d025fdd326471a164488130641119224d9f WatchSource:0}: Error finding container a1ea1d343020e82799c79e2482d03d025fdd326471a164488130641119224d9f: Status 404 returned error can't find the container with id a1ea1d343020e82799c79e2482d03d025fdd326471a164488130641119224d9f Mar 07 14:52:37 crc kubenswrapper[4943]: I0307 14:52:37.633811 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-749f4cd94-8k4nn" event={"ID":"9d6f004b-a409-4960-b377-8e8577597b99","Type":"ContainerStarted","Data":"a1ea1d343020e82799c79e2482d03d025fdd326471a164488130641119224d9f"} Mar 07 14:52:41 crc kubenswrapper[4943]: I0307 14:52:41.655888 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-749f4cd94-8k4nn" event={"ID":"9d6f004b-a409-4960-b377-8e8577597b99","Type":"ContainerStarted","Data":"90efd2a8e172abef471a9c11fb8bd9a36a87a404e550b924a78d90a31cb619b2"} Mar 07 14:52:41 crc kubenswrapper[4943]: I0307 14:52:41.657754 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-749f4cd94-8k4nn" Mar 07 14:52:46 crc kubenswrapper[4943]: I0307 14:52:46.688629 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-749f4cd94-8k4nn" Mar 07 14:52:46 crc kubenswrapper[4943]: I0307 14:52:46.728960 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-749f4cd94-8k4nn" podStartSLOduration=7.269779109 podStartE2EDuration="10.728908465s" podCreationTimestamp="2026-03-07 14:52:36 +0000 UTC" firstStartedPulling="2026-03-07 14:52:37.124912907 +0000 UTC m=+799.077049455" lastFinishedPulling="2026-03-07 14:52:40.584042313 +0000 UTC m=+802.536178811" observedRunningTime="2026-03-07 14:52:41.68289458 +0000 UTC m=+803.635031108" watchObservedRunningTime="2026-03-07 14:52:46.728908465 +0000 UTC m=+808.681044993" Mar 07 14:52:52 crc kubenswrapper[4943]: I0307 14:52:52.095171 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-index-wxxzb"] Mar 07 14:52:52 crc kubenswrapper[4943]: I0307 14:52:52.096689 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-wxxzb" Mar 07 14:52:52 crc kubenswrapper[4943]: I0307 14:52:52.101170 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-index-dockercfg-b8dkf" Mar 07 14:52:52 crc kubenswrapper[4943]: I0307 14:52:52.108999 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-wxxzb"] Mar 07 14:52:52 crc kubenswrapper[4943]: I0307 14:52:52.272066 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tk46p\" (UniqueName: \"kubernetes.io/projected/74d356f4-d8b0-4d36-9f9a-d23d6bfe8dc7-kube-api-access-tk46p\") pod \"infra-operator-index-wxxzb\" (UID: \"74d356f4-d8b0-4d36-9f9a-d23d6bfe8dc7\") " pod="openstack-operators/infra-operator-index-wxxzb" Mar 07 14:52:52 crc kubenswrapper[4943]: I0307 14:52:52.373802 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tk46p\" (UniqueName: \"kubernetes.io/projected/74d356f4-d8b0-4d36-9f9a-d23d6bfe8dc7-kube-api-access-tk46p\") pod \"infra-operator-index-wxxzb\" (UID: \"74d356f4-d8b0-4d36-9f9a-d23d6bfe8dc7\") " pod="openstack-operators/infra-operator-index-wxxzb" Mar 07 14:52:52 crc kubenswrapper[4943]: I0307 14:52:52.420278 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tk46p\" (UniqueName: \"kubernetes.io/projected/74d356f4-d8b0-4d36-9f9a-d23d6bfe8dc7-kube-api-access-tk46p\") pod \"infra-operator-index-wxxzb\" (UID: \"74d356f4-d8b0-4d36-9f9a-d23d6bfe8dc7\") " pod="openstack-operators/infra-operator-index-wxxzb" Mar 07 14:52:52 crc kubenswrapper[4943]: I0307 14:52:52.712677 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-wxxzb" Mar 07 14:52:53 crc kubenswrapper[4943]: I0307 14:52:53.047643 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-wxxzb"] Mar 07 14:52:53 crc kubenswrapper[4943]: I0307 14:52:53.754364 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-wxxzb" event={"ID":"74d356f4-d8b0-4d36-9f9a-d23d6bfe8dc7","Type":"ContainerStarted","Data":"a28d6950b5b3e465e740579e362b4418921e79dde080f3f330ee329d7863fc18"} Mar 07 14:52:54 crc kubenswrapper[4943]: I0307 14:52:54.765211 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-wxxzb" event={"ID":"74d356f4-d8b0-4d36-9f9a-d23d6bfe8dc7","Type":"ContainerStarted","Data":"0b729143767a7fba297a7576f68b46ae9bc6032d604057f8958297dbcf73fe93"} Mar 07 14:52:54 crc kubenswrapper[4943]: I0307 14:52:54.782207 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-index-wxxzb" podStartSLOduration=1.835163219 podStartE2EDuration="2.782180693s" podCreationTimestamp="2026-03-07 14:52:52 +0000 UTC" firstStartedPulling="2026-03-07 14:52:53.062200238 +0000 UTC m=+815.014336746" lastFinishedPulling="2026-03-07 14:52:54.009217682 +0000 UTC m=+815.961354220" observedRunningTime="2026-03-07 14:52:54.781017874 +0000 UTC m=+816.733154422" watchObservedRunningTime="2026-03-07 14:52:54.782180693 +0000 UTC m=+816.734317231" Mar 07 14:52:56 crc kubenswrapper[4943]: I0307 14:52:56.292982 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-index-wxxzb"] Mar 07 14:52:56 crc kubenswrapper[4943]: I0307 14:52:56.779383 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/infra-operator-index-wxxzb" podUID="74d356f4-d8b0-4d36-9f9a-d23d6bfe8dc7" containerName="registry-server" containerID="cri-o://0b729143767a7fba297a7576f68b46ae9bc6032d604057f8958297dbcf73fe93" gracePeriod=2 Mar 07 14:52:56 crc kubenswrapper[4943]: I0307 14:52:56.913832 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-index-qzp5d"] Mar 07 14:52:56 crc kubenswrapper[4943]: I0307 14:52:56.915114 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-qzp5d" Mar 07 14:52:56 crc kubenswrapper[4943]: I0307 14:52:56.918876 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-qzp5d"] Mar 07 14:52:57 crc kubenswrapper[4943]: I0307 14:52:57.051240 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gh4j\" (UniqueName: \"kubernetes.io/projected/708bf537-729f-4ce7-8a31-ffd3f252cf43-kube-api-access-7gh4j\") pod \"infra-operator-index-qzp5d\" (UID: \"708bf537-729f-4ce7-8a31-ffd3f252cf43\") " pod="openstack-operators/infra-operator-index-qzp5d" Mar 07 14:52:57 crc kubenswrapper[4943]: I0307 14:52:57.153606 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gh4j\" (UniqueName: \"kubernetes.io/projected/708bf537-729f-4ce7-8a31-ffd3f252cf43-kube-api-access-7gh4j\") pod \"infra-operator-index-qzp5d\" (UID: \"708bf537-729f-4ce7-8a31-ffd3f252cf43\") " pod="openstack-operators/infra-operator-index-qzp5d" Mar 07 14:52:57 crc kubenswrapper[4943]: I0307 14:52:57.217398 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gh4j\" (UniqueName: \"kubernetes.io/projected/708bf537-729f-4ce7-8a31-ffd3f252cf43-kube-api-access-7gh4j\") pod \"infra-operator-index-qzp5d\" (UID: \"708bf537-729f-4ce7-8a31-ffd3f252cf43\") " pod="openstack-operators/infra-operator-index-qzp5d" Mar 07 14:52:57 crc kubenswrapper[4943]: I0307 14:52:57.251801 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-qzp5d" Mar 07 14:52:57 crc kubenswrapper[4943]: I0307 14:52:57.325165 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-wxxzb" Mar 07 14:52:57 crc kubenswrapper[4943]: I0307 14:52:57.354888 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk46p\" (UniqueName: \"kubernetes.io/projected/74d356f4-d8b0-4d36-9f9a-d23d6bfe8dc7-kube-api-access-tk46p\") pod \"74d356f4-d8b0-4d36-9f9a-d23d6bfe8dc7\" (UID: \"74d356f4-d8b0-4d36-9f9a-d23d6bfe8dc7\") " Mar 07 14:52:57 crc kubenswrapper[4943]: I0307 14:52:57.358668 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74d356f4-d8b0-4d36-9f9a-d23d6bfe8dc7-kube-api-access-tk46p" (OuterVolumeSpecName: "kube-api-access-tk46p") pod "74d356f4-d8b0-4d36-9f9a-d23d6bfe8dc7" (UID: "74d356f4-d8b0-4d36-9f9a-d23d6bfe8dc7"). InnerVolumeSpecName "kube-api-access-tk46p". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:52:57 crc kubenswrapper[4943]: I0307 14:52:57.456087 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk46p\" (UniqueName: \"kubernetes.io/projected/74d356f4-d8b0-4d36-9f9a-d23d6bfe8dc7-kube-api-access-tk46p\") on node \"crc\" DevicePath \"\"" Mar 07 14:52:57 crc kubenswrapper[4943]: I0307 14:52:57.689186 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-qzp5d"] Mar 07 14:52:57 crc kubenswrapper[4943]: W0307 14:52:57.694699 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod708bf537_729f_4ce7_8a31_ffd3f252cf43.slice/crio-cf5b98e4f43223853600ba5ad98ada8cf0e3240dbca097da640cf455696736c2 WatchSource:0}: Error finding container cf5b98e4f43223853600ba5ad98ada8cf0e3240dbca097da640cf455696736c2: Status 404 returned error can't find the container with id cf5b98e4f43223853600ba5ad98ada8cf0e3240dbca097da640cf455696736c2 Mar 07 14:52:58 crc kubenswrapper[4943]: I0307 14:52:57.786742 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-qzp5d" event={"ID":"708bf537-729f-4ce7-8a31-ffd3f252cf43","Type":"ContainerStarted","Data":"cf5b98e4f43223853600ba5ad98ada8cf0e3240dbca097da640cf455696736c2"} Mar 07 14:52:58 crc kubenswrapper[4943]: I0307 14:52:57.789519 4943 generic.go:334] "Generic (PLEG): container finished" podID="74d356f4-d8b0-4d36-9f9a-d23d6bfe8dc7" containerID="0b729143767a7fba297a7576f68b46ae9bc6032d604057f8958297dbcf73fe93" exitCode=0 Mar 07 14:52:58 crc kubenswrapper[4943]: I0307 14:52:57.789569 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-wxxzb" event={"ID":"74d356f4-d8b0-4d36-9f9a-d23d6bfe8dc7","Type":"ContainerDied","Data":"0b729143767a7fba297a7576f68b46ae9bc6032d604057f8958297dbcf73fe93"} Mar 07 14:52:58 crc kubenswrapper[4943]: I0307 14:52:57.789606 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-wxxzb" Mar 07 14:52:58 crc kubenswrapper[4943]: I0307 14:52:57.789625 4943 scope.go:117] "RemoveContainer" containerID="0b729143767a7fba297a7576f68b46ae9bc6032d604057f8958297dbcf73fe93" Mar 07 14:52:58 crc kubenswrapper[4943]: I0307 14:52:57.789608 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-wxxzb" event={"ID":"74d356f4-d8b0-4d36-9f9a-d23d6bfe8dc7","Type":"ContainerDied","Data":"a28d6950b5b3e465e740579e362b4418921e79dde080f3f330ee329d7863fc18"} Mar 07 14:52:58 crc kubenswrapper[4943]: I0307 14:52:58.271250 4943 scope.go:117] "RemoveContainer" containerID="0b729143767a7fba297a7576f68b46ae9bc6032d604057f8958297dbcf73fe93" Mar 07 14:52:58 crc kubenswrapper[4943]: E0307 14:52:58.272288 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b729143767a7fba297a7576f68b46ae9bc6032d604057f8958297dbcf73fe93\": container with ID starting with 0b729143767a7fba297a7576f68b46ae9bc6032d604057f8958297dbcf73fe93 not found: ID does not exist" containerID="0b729143767a7fba297a7576f68b46ae9bc6032d604057f8958297dbcf73fe93" Mar 07 14:52:58 crc kubenswrapper[4943]: I0307 14:52:58.272333 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b729143767a7fba297a7576f68b46ae9bc6032d604057f8958297dbcf73fe93"} err="failed to get container status \"0b729143767a7fba297a7576f68b46ae9bc6032d604057f8958297dbcf73fe93\": rpc error: code = NotFound desc = could not find container \"0b729143767a7fba297a7576f68b46ae9bc6032d604057f8958297dbcf73fe93\": container with ID starting with 0b729143767a7fba297a7576f68b46ae9bc6032d604057f8958297dbcf73fe93 not found: ID does not exist" Mar 07 14:52:58 crc kubenswrapper[4943]: I0307 14:52:58.276187 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-index-wxxzb"] Mar 07 14:52:58 crc kubenswrapper[4943]: I0307 14:52:58.298738 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/infra-operator-index-wxxzb"] Mar 07 14:52:58 crc kubenswrapper[4943]: I0307 14:52:58.769421 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74d356f4-d8b0-4d36-9f9a-d23d6bfe8dc7" path="/var/lib/kubelet/pods/74d356f4-d8b0-4d36-9f9a-d23d6bfe8dc7/volumes" Mar 07 14:52:58 crc kubenswrapper[4943]: I0307 14:52:58.802248 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-qzp5d" event={"ID":"708bf537-729f-4ce7-8a31-ffd3f252cf43","Type":"ContainerStarted","Data":"b396362735e0ea81d8fe50091ef7e9db2a945cdd74f8141adec8e46e1ee7b50a"} Mar 07 14:52:58 crc kubenswrapper[4943]: I0307 14:52:58.830308 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-index-qzp5d" podStartSLOduration=2.223345233 podStartE2EDuration="2.83028486s" podCreationTimestamp="2026-03-07 14:52:56 +0000 UTC" firstStartedPulling="2026-03-07 14:52:57.699362857 +0000 UTC m=+819.651499385" lastFinishedPulling="2026-03-07 14:52:58.306302514 +0000 UTC m=+820.258439012" observedRunningTime="2026-03-07 14:52:58.829036479 +0000 UTC m=+820.781173027" watchObservedRunningTime="2026-03-07 14:52:58.83028486 +0000 UTC m=+820.782421358" Mar 07 14:53:06 crc kubenswrapper[4943]: I0307 14:53:06.074448 4943 patch_prober.go:28] interesting pod/machine-config-daemon-cpdmm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 07 14:53:06 crc kubenswrapper[4943]: I0307 14:53:06.075414 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 07 14:53:06 crc kubenswrapper[4943]: I0307 14:53:06.075502 4943 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" Mar 07 14:53:06 crc kubenswrapper[4943]: I0307 14:53:06.076546 4943 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8ffd8d858c956cfb1e523c263c554f1c6d621212d5b2c18a1bc047ffbcf47fee"} pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 07 14:53:06 crc kubenswrapper[4943]: I0307 14:53:06.076692 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerName="machine-config-daemon" containerID="cri-o://8ffd8d858c956cfb1e523c263c554f1c6d621212d5b2c18a1bc047ffbcf47fee" gracePeriod=600 Mar 07 14:53:06 crc kubenswrapper[4943]: I0307 14:53:06.869993 4943 generic.go:334] "Generic (PLEG): container finished" podID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerID="8ffd8d858c956cfb1e523c263c554f1c6d621212d5b2c18a1bc047ffbcf47fee" exitCode=0 Mar 07 14:53:06 crc kubenswrapper[4943]: I0307 14:53:06.870370 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" event={"ID":"4c1f18ea-e557-41fc-95dd-ce4e1722a61e","Type":"ContainerDied","Data":"8ffd8d858c956cfb1e523c263c554f1c6d621212d5b2c18a1bc047ffbcf47fee"} Mar 07 14:53:06 crc kubenswrapper[4943]: I0307 14:53:06.870398 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" event={"ID":"4c1f18ea-e557-41fc-95dd-ce4e1722a61e","Type":"ContainerStarted","Data":"fe852e1f10f4ca72f749789ca1e6ab73542eaedb5349732682970babc3f1afd1"} Mar 07 14:53:06 crc kubenswrapper[4943]: I0307 14:53:06.870415 4943 scope.go:117] "RemoveContainer" containerID="b2616e0e9d2516da6111a60698ff218442b059990c7d49e009916d0f941b033b" Mar 07 14:53:07 crc kubenswrapper[4943]: I0307 14:53:07.252400 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/infra-operator-index-qzp5d" Mar 07 14:53:07 crc kubenswrapper[4943]: I0307 14:53:07.252483 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-index-qzp5d" Mar 07 14:53:07 crc kubenswrapper[4943]: I0307 14:53:07.293620 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/infra-operator-index-qzp5d" Mar 07 14:53:07 crc kubenswrapper[4943]: I0307 14:53:07.927847 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-index-qzp5d" Mar 07 14:53:15 crc kubenswrapper[4943]: I0307 14:53:15.961538 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8cwzdw8"] Mar 07 14:53:15 crc kubenswrapper[4943]: E0307 14:53:15.962505 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74d356f4-d8b0-4d36-9f9a-d23d6bfe8dc7" containerName="registry-server" Mar 07 14:53:15 crc kubenswrapper[4943]: I0307 14:53:15.962529 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="74d356f4-d8b0-4d36-9f9a-d23d6bfe8dc7" containerName="registry-server" Mar 07 14:53:15 crc kubenswrapper[4943]: I0307 14:53:15.962765 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="74d356f4-d8b0-4d36-9f9a-d23d6bfe8dc7" containerName="registry-server" Mar 07 14:53:15 crc kubenswrapper[4943]: I0307 14:53:15.964230 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8cwzdw8" Mar 07 14:53:15 crc kubenswrapper[4943]: I0307 14:53:15.967518 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-k6gp2" Mar 07 14:53:15 crc kubenswrapper[4943]: I0307 14:53:15.990970 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8cwzdw8"] Mar 07 14:53:16 crc kubenswrapper[4943]: I0307 14:53:16.038991 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hsnjg\" (UniqueName: \"kubernetes.io/projected/487d337d-5286-4425-b303-e871dcec36e8-kube-api-access-hsnjg\") pod \"c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8cwzdw8\" (UID: \"487d337d-5286-4425-b303-e871dcec36e8\") " pod="openstack-operators/c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8cwzdw8" Mar 07 14:53:16 crc kubenswrapper[4943]: I0307 14:53:16.039156 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/487d337d-5286-4425-b303-e871dcec36e8-util\") pod \"c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8cwzdw8\" (UID: \"487d337d-5286-4425-b303-e871dcec36e8\") " pod="openstack-operators/c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8cwzdw8" Mar 07 14:53:16 crc kubenswrapper[4943]: I0307 14:53:16.039248 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/487d337d-5286-4425-b303-e871dcec36e8-bundle\") pod \"c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8cwzdw8\" (UID: \"487d337d-5286-4425-b303-e871dcec36e8\") " pod="openstack-operators/c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8cwzdw8" Mar 07 14:53:16 crc kubenswrapper[4943]: I0307 14:53:16.141169 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hsnjg\" (UniqueName: \"kubernetes.io/projected/487d337d-5286-4425-b303-e871dcec36e8-kube-api-access-hsnjg\") pod \"c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8cwzdw8\" (UID: \"487d337d-5286-4425-b303-e871dcec36e8\") " pod="openstack-operators/c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8cwzdw8" Mar 07 14:53:16 crc kubenswrapper[4943]: I0307 14:53:16.141464 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/487d337d-5286-4425-b303-e871dcec36e8-util\") pod \"c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8cwzdw8\" (UID: \"487d337d-5286-4425-b303-e871dcec36e8\") " pod="openstack-operators/c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8cwzdw8" Mar 07 14:53:16 crc kubenswrapper[4943]: I0307 14:53:16.141621 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/487d337d-5286-4425-b303-e871dcec36e8-bundle\") pod \"c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8cwzdw8\" (UID: \"487d337d-5286-4425-b303-e871dcec36e8\") " pod="openstack-operators/c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8cwzdw8" Mar 07 14:53:16 crc kubenswrapper[4943]: I0307 14:53:16.142713 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/487d337d-5286-4425-b303-e871dcec36e8-util\") pod \"c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8cwzdw8\" (UID: \"487d337d-5286-4425-b303-e871dcec36e8\") " pod="openstack-operators/c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8cwzdw8" Mar 07 14:53:16 crc kubenswrapper[4943]: I0307 14:53:16.142766 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/487d337d-5286-4425-b303-e871dcec36e8-bundle\") pod \"c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8cwzdw8\" (UID: \"487d337d-5286-4425-b303-e871dcec36e8\") " pod="openstack-operators/c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8cwzdw8" Mar 07 14:53:16 crc kubenswrapper[4943]: I0307 14:53:16.176056 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hsnjg\" (UniqueName: \"kubernetes.io/projected/487d337d-5286-4425-b303-e871dcec36e8-kube-api-access-hsnjg\") pod \"c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8cwzdw8\" (UID: \"487d337d-5286-4425-b303-e871dcec36e8\") " pod="openstack-operators/c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8cwzdw8" Mar 07 14:53:16 crc kubenswrapper[4943]: I0307 14:53:16.290921 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8cwzdw8" Mar 07 14:53:16 crc kubenswrapper[4943]: I0307 14:53:16.822046 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8cwzdw8"] Mar 07 14:53:16 crc kubenswrapper[4943]: I0307 14:53:16.957195 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8cwzdw8" event={"ID":"487d337d-5286-4425-b303-e871dcec36e8","Type":"ContainerStarted","Data":"e5db8434682b6882ac9af58e90769f3bec1f444b93acd7d85594521d85c4414c"} Mar 07 14:53:17 crc kubenswrapper[4943]: I0307 14:53:17.968398 4943 generic.go:334] "Generic (PLEG): container finished" podID="487d337d-5286-4425-b303-e871dcec36e8" containerID="5543f1825dbd5d27d60a70dd7541867919827c48ec12fb5375f56218156ce6b2" exitCode=0 Mar 07 14:53:17 crc kubenswrapper[4943]: I0307 14:53:17.968885 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8cwzdw8" event={"ID":"487d337d-5286-4425-b303-e871dcec36e8","Type":"ContainerDied","Data":"5543f1825dbd5d27d60a70dd7541867919827c48ec12fb5375f56218156ce6b2"} Mar 07 14:53:17 crc kubenswrapper[4943]: I0307 14:53:17.970902 4943 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 07 14:53:18 crc kubenswrapper[4943]: I0307 14:53:18.979544 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8cwzdw8" event={"ID":"487d337d-5286-4425-b303-e871dcec36e8","Type":"ContainerStarted","Data":"67f2d17b9c056941690196f736fdc45dca83423c2890f2072166f76ef1c3c040"} Mar 07 14:53:19 crc kubenswrapper[4943]: I0307 14:53:19.988545 4943 generic.go:334] "Generic (PLEG): container finished" podID="487d337d-5286-4425-b303-e871dcec36e8" containerID="67f2d17b9c056941690196f736fdc45dca83423c2890f2072166f76ef1c3c040" exitCode=0 Mar 07 14:53:19 crc kubenswrapper[4943]: I0307 14:53:19.988607 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8cwzdw8" event={"ID":"487d337d-5286-4425-b303-e871dcec36e8","Type":"ContainerDied","Data":"67f2d17b9c056941690196f736fdc45dca83423c2890f2072166f76ef1c3c040"} Mar 07 14:53:21 crc kubenswrapper[4943]: I0307 14:53:21.000072 4943 generic.go:334] "Generic (PLEG): container finished" podID="487d337d-5286-4425-b303-e871dcec36e8" containerID="75a60bf38484c3b0db8ff0b64c74b88a56135aabce1abb5019478e54e87da7af" exitCode=0 Mar 07 14:53:21 crc kubenswrapper[4943]: I0307 14:53:21.000160 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8cwzdw8" event={"ID":"487d337d-5286-4425-b303-e871dcec36e8","Type":"ContainerDied","Data":"75a60bf38484c3b0db8ff0b64c74b88a56135aabce1abb5019478e54e87da7af"} Mar 07 14:53:22 crc kubenswrapper[4943]: I0307 14:53:22.378356 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8cwzdw8" Mar 07 14:53:22 crc kubenswrapper[4943]: I0307 14:53:22.576522 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hsnjg\" (UniqueName: \"kubernetes.io/projected/487d337d-5286-4425-b303-e871dcec36e8-kube-api-access-hsnjg\") pod \"487d337d-5286-4425-b303-e871dcec36e8\" (UID: \"487d337d-5286-4425-b303-e871dcec36e8\") " Mar 07 14:53:22 crc kubenswrapper[4943]: I0307 14:53:22.576734 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/487d337d-5286-4425-b303-e871dcec36e8-bundle\") pod \"487d337d-5286-4425-b303-e871dcec36e8\" (UID: \"487d337d-5286-4425-b303-e871dcec36e8\") " Mar 07 14:53:22 crc kubenswrapper[4943]: I0307 14:53:22.576885 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/487d337d-5286-4425-b303-e871dcec36e8-util\") pod \"487d337d-5286-4425-b303-e871dcec36e8\" (UID: \"487d337d-5286-4425-b303-e871dcec36e8\") " Mar 07 14:53:22 crc kubenswrapper[4943]: I0307 14:53:22.581361 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/487d337d-5286-4425-b303-e871dcec36e8-bundle" (OuterVolumeSpecName: "bundle") pod "487d337d-5286-4425-b303-e871dcec36e8" (UID: "487d337d-5286-4425-b303-e871dcec36e8"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:53:22 crc kubenswrapper[4943]: I0307 14:53:22.586229 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/487d337d-5286-4425-b303-e871dcec36e8-kube-api-access-hsnjg" (OuterVolumeSpecName: "kube-api-access-hsnjg") pod "487d337d-5286-4425-b303-e871dcec36e8" (UID: "487d337d-5286-4425-b303-e871dcec36e8"). InnerVolumeSpecName "kube-api-access-hsnjg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:53:22 crc kubenswrapper[4943]: I0307 14:53:22.601551 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/487d337d-5286-4425-b303-e871dcec36e8-util" (OuterVolumeSpecName: "util") pod "487d337d-5286-4425-b303-e871dcec36e8" (UID: "487d337d-5286-4425-b303-e871dcec36e8"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:53:22 crc kubenswrapper[4943]: I0307 14:53:22.678390 4943 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/487d337d-5286-4425-b303-e871dcec36e8-util\") on node \"crc\" DevicePath \"\"" Mar 07 14:53:22 crc kubenswrapper[4943]: I0307 14:53:22.678461 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hsnjg\" (UniqueName: \"kubernetes.io/projected/487d337d-5286-4425-b303-e871dcec36e8-kube-api-access-hsnjg\") on node \"crc\" DevicePath \"\"" Mar 07 14:53:22 crc kubenswrapper[4943]: I0307 14:53:22.678489 4943 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/487d337d-5286-4425-b303-e871dcec36e8-bundle\") on node \"crc\" DevicePath \"\"" Mar 07 14:53:23 crc kubenswrapper[4943]: I0307 14:53:23.020231 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8cwzdw8" event={"ID":"487d337d-5286-4425-b303-e871dcec36e8","Type":"ContainerDied","Data":"e5db8434682b6882ac9af58e90769f3bec1f444b93acd7d85594521d85c4414c"} Mar 07 14:53:23 crc kubenswrapper[4943]: I0307 14:53:23.020775 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e5db8434682b6882ac9af58e90769f3bec1f444b93acd7d85594521d85c4414c" Mar 07 14:53:23 crc kubenswrapper[4943]: I0307 14:53:23.020721 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8cwzdw8" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.402866 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/openstack-galera-0"] Mar 07 14:53:32 crc kubenswrapper[4943]: E0307 14:53:32.404483 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="487d337d-5286-4425-b303-e871dcec36e8" containerName="extract" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.404511 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="487d337d-5286-4425-b303-e871dcec36e8" containerName="extract" Mar 07 14:53:32 crc kubenswrapper[4943]: E0307 14:53:32.404543 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="487d337d-5286-4425-b303-e871dcec36e8" containerName="util" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.404555 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="487d337d-5286-4425-b303-e871dcec36e8" containerName="util" Mar 07 14:53:32 crc kubenswrapper[4943]: E0307 14:53:32.404572 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="487d337d-5286-4425-b303-e871dcec36e8" containerName="pull" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.404589 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="487d337d-5286-4425-b303-e871dcec36e8" containerName="pull" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.404791 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="487d337d-5286-4425-b303-e871dcec36e8" containerName="extract" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.405863 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/openstack-galera-0" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.410560 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"openshift-service-ca.crt" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.411350 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"kube-root-ca.crt" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.411633 4943 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"galera-openstack-dockercfg-b475v" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.412041 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"openstack-config-data" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.412084 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"openstack-scripts" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.418457 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/openstack-galera-1"] Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.420414 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/openstack-galera-1" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.437697 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/openstack-galera-0"] Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.448796 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/openstack-galera-2"] Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.451229 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/openstack-galera-2" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.454551 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/openstack-galera-1"] Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.467364 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/openstack-galera-2"] Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.517287 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/92e3c4ec-dbf5-443c-b11a-7d1d961a92c0-operator-scripts\") pod \"openstack-galera-0\" (UID: \"92e3c4ec-dbf5-443c-b11a-7d1d961a92c0\") " pod="swift-kuttl-tests/openstack-galera-0" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.517337 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-0\" (UID: \"92e3c4ec-dbf5-443c-b11a-7d1d961a92c0\") " pod="swift-kuttl-tests/openstack-galera-0" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.517380 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/92e3c4ec-dbf5-443c-b11a-7d1d961a92c0-config-data-default\") pod \"openstack-galera-0\" (UID: \"92e3c4ec-dbf5-443c-b11a-7d1d961a92c0\") " pod="swift-kuttl-tests/openstack-galera-0" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.517398 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjzsz\" (UniqueName: \"kubernetes.io/projected/92e3c4ec-dbf5-443c-b11a-7d1d961a92c0-kube-api-access-fjzsz\") pod \"openstack-galera-0\" (UID: \"92e3c4ec-dbf5-443c-b11a-7d1d961a92c0\") " pod="swift-kuttl-tests/openstack-galera-0" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.517434 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/92e3c4ec-dbf5-443c-b11a-7d1d961a92c0-kolla-config\") pod \"openstack-galera-0\" (UID: \"92e3c4ec-dbf5-443c-b11a-7d1d961a92c0\") " pod="swift-kuttl-tests/openstack-galera-0" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.517631 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/92e3c4ec-dbf5-443c-b11a-7d1d961a92c0-config-data-generated\") pod \"openstack-galera-0\" (UID: \"92e3c4ec-dbf5-443c-b11a-7d1d961a92c0\") " pod="swift-kuttl-tests/openstack-galera-0" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.618537 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-0\" (UID: \"92e3c4ec-dbf5-443c-b11a-7d1d961a92c0\") " pod="swift-kuttl-tests/openstack-galera-0" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.618611 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/64dcacc7-fd15-41e0-ac82-fc37aa562331-operator-scripts\") pod \"openstack-galera-2\" (UID: \"64dcacc7-fd15-41e0-ac82-fc37aa562331\") " pod="swift-kuttl-tests/openstack-galera-2" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.618653 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-2\" (UID: \"64dcacc7-fd15-41e0-ac82-fc37aa562331\") " pod="swift-kuttl-tests/openstack-galera-2" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.618690 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/64dcacc7-fd15-41e0-ac82-fc37aa562331-config-data-generated\") pod \"openstack-galera-2\" (UID: \"64dcacc7-fd15-41e0-ac82-fc37aa562331\") " pod="swift-kuttl-tests/openstack-galera-2" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.618738 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/4f2c813e-73c6-41c7-ba8d-5498fe1d5671-config-data-generated\") pod \"openstack-galera-1\" (UID: \"4f2c813e-73c6-41c7-ba8d-5498fe1d5671\") " pod="swift-kuttl-tests/openstack-galera-1" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.618965 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/92e3c4ec-dbf5-443c-b11a-7d1d961a92c0-config-data-default\") pod \"openstack-galera-0\" (UID: \"92e3c4ec-dbf5-443c-b11a-7d1d961a92c0\") " pod="swift-kuttl-tests/openstack-galera-0" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.619026 4943 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-0\" (UID: \"92e3c4ec-dbf5-443c-b11a-7d1d961a92c0\") device mount path \"/mnt/openstack/pv11\"" pod="swift-kuttl-tests/openstack-galera-0" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.619035 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjzsz\" (UniqueName: \"kubernetes.io/projected/92e3c4ec-dbf5-443c-b11a-7d1d961a92c0-kube-api-access-fjzsz\") pod \"openstack-galera-0\" (UID: \"92e3c4ec-dbf5-443c-b11a-7d1d961a92c0\") " pod="swift-kuttl-tests/openstack-galera-0" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.619180 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/64dcacc7-fd15-41e0-ac82-fc37aa562331-kolla-config\") pod \"openstack-galera-2\" (UID: \"64dcacc7-fd15-41e0-ac82-fc37aa562331\") " pod="swift-kuttl-tests/openstack-galera-2" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.619220 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/4f2c813e-73c6-41c7-ba8d-5498fe1d5671-config-data-default\") pod \"openstack-galera-1\" (UID: \"4f2c813e-73c6-41c7-ba8d-5498fe1d5671\") " pod="swift-kuttl-tests/openstack-galera-1" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.619319 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4f2c813e-73c6-41c7-ba8d-5498fe1d5671-operator-scripts\") pod \"openstack-galera-1\" (UID: \"4f2c813e-73c6-41c7-ba8d-5498fe1d5671\") " pod="swift-kuttl-tests/openstack-galera-1" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.619447 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-knqfn\" (UniqueName: \"kubernetes.io/projected/4f2c813e-73c6-41c7-ba8d-5498fe1d5671-kube-api-access-knqfn\") pod \"openstack-galera-1\" (UID: \"4f2c813e-73c6-41c7-ba8d-5498fe1d5671\") " pod="swift-kuttl-tests/openstack-galera-1" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.619499 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkw6w\" (UniqueName: \"kubernetes.io/projected/64dcacc7-fd15-41e0-ac82-fc37aa562331-kube-api-access-fkw6w\") pod \"openstack-galera-2\" (UID: \"64dcacc7-fd15-41e0-ac82-fc37aa562331\") " pod="swift-kuttl-tests/openstack-galera-2" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.619562 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/92e3c4ec-dbf5-443c-b11a-7d1d961a92c0-kolla-config\") pod \"openstack-galera-0\" (UID: \"92e3c4ec-dbf5-443c-b11a-7d1d961a92c0\") " pod="swift-kuttl-tests/openstack-galera-0" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.619616 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/64dcacc7-fd15-41e0-ac82-fc37aa562331-config-data-default\") pod \"openstack-galera-2\" (UID: \"64dcacc7-fd15-41e0-ac82-fc37aa562331\") " pod="swift-kuttl-tests/openstack-galera-2" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.619688 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/92e3c4ec-dbf5-443c-b11a-7d1d961a92c0-config-data-generated\") pod \"openstack-galera-0\" (UID: \"92e3c4ec-dbf5-443c-b11a-7d1d961a92c0\") " pod="swift-kuttl-tests/openstack-galera-0" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.619732 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4f2c813e-73c6-41c7-ba8d-5498fe1d5671-kolla-config\") pod \"openstack-galera-1\" (UID: \"4f2c813e-73c6-41c7-ba8d-5498fe1d5671\") " pod="swift-kuttl-tests/openstack-galera-1" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.619788 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-1\" (UID: \"4f2c813e-73c6-41c7-ba8d-5498fe1d5671\") " pod="swift-kuttl-tests/openstack-galera-1" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.619820 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/92e3c4ec-dbf5-443c-b11a-7d1d961a92c0-operator-scripts\") pod \"openstack-galera-0\" (UID: \"92e3c4ec-dbf5-443c-b11a-7d1d961a92c0\") " pod="swift-kuttl-tests/openstack-galera-0" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.620343 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/92e3c4ec-dbf5-443c-b11a-7d1d961a92c0-config-data-generated\") pod \"openstack-galera-0\" (UID: \"92e3c4ec-dbf5-443c-b11a-7d1d961a92c0\") " pod="swift-kuttl-tests/openstack-galera-0" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.620489 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/92e3c4ec-dbf5-443c-b11a-7d1d961a92c0-kolla-config\") pod \"openstack-galera-0\" (UID: \"92e3c4ec-dbf5-443c-b11a-7d1d961a92c0\") " pod="swift-kuttl-tests/openstack-galera-0" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.620573 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/92e3c4ec-dbf5-443c-b11a-7d1d961a92c0-config-data-default\") pod \"openstack-galera-0\" (UID: \"92e3c4ec-dbf5-443c-b11a-7d1d961a92c0\") " pod="swift-kuttl-tests/openstack-galera-0" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.622389 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/92e3c4ec-dbf5-443c-b11a-7d1d961a92c0-operator-scripts\") pod \"openstack-galera-0\" (UID: \"92e3c4ec-dbf5-443c-b11a-7d1d961a92c0\") " pod="swift-kuttl-tests/openstack-galera-0" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.645824 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjzsz\" (UniqueName: \"kubernetes.io/projected/92e3c4ec-dbf5-443c-b11a-7d1d961a92c0-kube-api-access-fjzsz\") pod \"openstack-galera-0\" (UID: \"92e3c4ec-dbf5-443c-b11a-7d1d961a92c0\") " pod="swift-kuttl-tests/openstack-galera-0" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.648521 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-0\" (UID: \"92e3c4ec-dbf5-443c-b11a-7d1d961a92c0\") " pod="swift-kuttl-tests/openstack-galera-0" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.720985 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4f2c813e-73c6-41c7-ba8d-5498fe1d5671-kolla-config\") pod \"openstack-galera-1\" (UID: \"4f2c813e-73c6-41c7-ba8d-5498fe1d5671\") " pod="swift-kuttl-tests/openstack-galera-1" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.721074 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-1\" (UID: \"4f2c813e-73c6-41c7-ba8d-5498fe1d5671\") " pod="swift-kuttl-tests/openstack-galera-1" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.721164 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/64dcacc7-fd15-41e0-ac82-fc37aa562331-operator-scripts\") pod \"openstack-galera-2\" (UID: \"64dcacc7-fd15-41e0-ac82-fc37aa562331\") " pod="swift-kuttl-tests/openstack-galera-2" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.721203 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-2\" (UID: \"64dcacc7-fd15-41e0-ac82-fc37aa562331\") " pod="swift-kuttl-tests/openstack-galera-2" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.721242 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/64dcacc7-fd15-41e0-ac82-fc37aa562331-config-data-generated\") pod \"openstack-galera-2\" (UID: \"64dcacc7-fd15-41e0-ac82-fc37aa562331\") " pod="swift-kuttl-tests/openstack-galera-2" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.721294 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/4f2c813e-73c6-41c7-ba8d-5498fe1d5671-config-data-generated\") pod \"openstack-galera-1\" (UID: \"4f2c813e-73c6-41c7-ba8d-5498fe1d5671\") " pod="swift-kuttl-tests/openstack-galera-1" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.721341 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/64dcacc7-fd15-41e0-ac82-fc37aa562331-kolla-config\") pod \"openstack-galera-2\" (UID: \"64dcacc7-fd15-41e0-ac82-fc37aa562331\") " pod="swift-kuttl-tests/openstack-galera-2" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.721375 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/4f2c813e-73c6-41c7-ba8d-5498fe1d5671-config-data-default\") pod \"openstack-galera-1\" (UID: \"4f2c813e-73c6-41c7-ba8d-5498fe1d5671\") " pod="swift-kuttl-tests/openstack-galera-1" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.721416 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4f2c813e-73c6-41c7-ba8d-5498fe1d5671-operator-scripts\") pod \"openstack-galera-1\" (UID: \"4f2c813e-73c6-41c7-ba8d-5498fe1d5671\") " pod="swift-kuttl-tests/openstack-galera-1" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.721460 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-knqfn\" (UniqueName: \"kubernetes.io/projected/4f2c813e-73c6-41c7-ba8d-5498fe1d5671-kube-api-access-knqfn\") pod \"openstack-galera-1\" (UID: \"4f2c813e-73c6-41c7-ba8d-5498fe1d5671\") " pod="swift-kuttl-tests/openstack-galera-1" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.721497 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkw6w\" (UniqueName: \"kubernetes.io/projected/64dcacc7-fd15-41e0-ac82-fc37aa562331-kube-api-access-fkw6w\") pod \"openstack-galera-2\" (UID: \"64dcacc7-fd15-41e0-ac82-fc37aa562331\") " pod="swift-kuttl-tests/openstack-galera-2" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.721545 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/64dcacc7-fd15-41e0-ac82-fc37aa562331-config-data-default\") pod \"openstack-galera-2\" (UID: \"64dcacc7-fd15-41e0-ac82-fc37aa562331\") " pod="swift-kuttl-tests/openstack-galera-2" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.721992 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4f2c813e-73c6-41c7-ba8d-5498fe1d5671-kolla-config\") pod \"openstack-galera-1\" (UID: \"4f2c813e-73c6-41c7-ba8d-5498fe1d5671\") " pod="swift-kuttl-tests/openstack-galera-1" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.722308 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/4f2c813e-73c6-41c7-ba8d-5498fe1d5671-config-data-generated\") pod \"openstack-galera-1\" (UID: \"4f2c813e-73c6-41c7-ba8d-5498fe1d5671\") " pod="swift-kuttl-tests/openstack-galera-1" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.722500 4943 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-1\" (UID: \"4f2c813e-73c6-41c7-ba8d-5498fe1d5671\") device mount path \"/mnt/openstack/pv02\"" pod="swift-kuttl-tests/openstack-galera-1" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.722653 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/64dcacc7-fd15-41e0-ac82-fc37aa562331-config-data-generated\") pod \"openstack-galera-2\" (UID: \"64dcacc7-fd15-41e0-ac82-fc37aa562331\") " pod="swift-kuttl-tests/openstack-galera-2" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.722712 4943 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-2\" (UID: \"64dcacc7-fd15-41e0-ac82-fc37aa562331\") device mount path \"/mnt/openstack/pv10\"" pod="swift-kuttl-tests/openstack-galera-2" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.723026 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/64dcacc7-fd15-41e0-ac82-fc37aa562331-config-data-default\") pod \"openstack-galera-2\" (UID: \"64dcacc7-fd15-41e0-ac82-fc37aa562331\") " pod="swift-kuttl-tests/openstack-galera-2" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.723682 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/64dcacc7-fd15-41e0-ac82-fc37aa562331-kolla-config\") pod \"openstack-galera-2\" (UID: \"64dcacc7-fd15-41e0-ac82-fc37aa562331\") " pod="swift-kuttl-tests/openstack-galera-2" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.724077 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/4f2c813e-73c6-41c7-ba8d-5498fe1d5671-config-data-default\") pod \"openstack-galera-1\" (UID: \"4f2c813e-73c6-41c7-ba8d-5498fe1d5671\") " pod="swift-kuttl-tests/openstack-galera-1" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.724269 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/64dcacc7-fd15-41e0-ac82-fc37aa562331-operator-scripts\") pod \"openstack-galera-2\" (UID: \"64dcacc7-fd15-41e0-ac82-fc37aa562331\") " pod="swift-kuttl-tests/openstack-galera-2" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.725342 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4f2c813e-73c6-41c7-ba8d-5498fe1d5671-operator-scripts\") pod \"openstack-galera-1\" (UID: \"4f2c813e-73c6-41c7-ba8d-5498fe1d5671\") " pod="swift-kuttl-tests/openstack-galera-1" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.745156 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/openstack-galera-0" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.749513 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-2\" (UID: \"64dcacc7-fd15-41e0-ac82-fc37aa562331\") " pod="swift-kuttl-tests/openstack-galera-2" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.757615 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-1\" (UID: \"4f2c813e-73c6-41c7-ba8d-5498fe1d5671\") " pod="swift-kuttl-tests/openstack-galera-1" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.759905 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-knqfn\" (UniqueName: \"kubernetes.io/projected/4f2c813e-73c6-41c7-ba8d-5498fe1d5671-kube-api-access-knqfn\") pod \"openstack-galera-1\" (UID: \"4f2c813e-73c6-41c7-ba8d-5498fe1d5671\") " pod="swift-kuttl-tests/openstack-galera-1" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.767845 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkw6w\" (UniqueName: \"kubernetes.io/projected/64dcacc7-fd15-41e0-ac82-fc37aa562331-kube-api-access-fkw6w\") pod \"openstack-galera-2\" (UID: \"64dcacc7-fd15-41e0-ac82-fc37aa562331\") " pod="swift-kuttl-tests/openstack-galera-2" Mar 07 14:53:32 crc kubenswrapper[4943]: I0307 14:53:32.785384 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/openstack-galera-2" Mar 07 14:53:33 crc kubenswrapper[4943]: I0307 14:53:33.067654 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/openstack-galera-1" Mar 07 14:53:33 crc kubenswrapper[4943]: I0307 14:53:33.228599 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/openstack-galera-2"] Mar 07 14:53:33 crc kubenswrapper[4943]: W0307 14:53:33.244325 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod64dcacc7_fd15_41e0_ac82_fc37aa562331.slice/crio-ac2328c3fb50b500200182d89ae00d95a6eaca6689d8151e8cc1cc335dbca435 WatchSource:0}: Error finding container ac2328c3fb50b500200182d89ae00d95a6eaca6689d8151e8cc1cc335dbca435: Status 404 returned error can't find the container with id ac2328c3fb50b500200182d89ae00d95a6eaca6689d8151e8cc1cc335dbca435 Mar 07 14:53:33 crc kubenswrapper[4943]: I0307 14:53:33.283744 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/openstack-galera-0"] Mar 07 14:53:33 crc kubenswrapper[4943]: I0307 14:53:33.412766 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/openstack-galera-1"] Mar 07 14:53:34 crc kubenswrapper[4943]: I0307 14:53:34.132968 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/openstack-galera-1" event={"ID":"4f2c813e-73c6-41c7-ba8d-5498fe1d5671","Type":"ContainerStarted","Data":"05c2e3a36da8d7bea823ea1488910ba12ebbf5a547304922c9de9832b4b7c1b2"} Mar 07 14:53:34 crc kubenswrapper[4943]: I0307 14:53:34.137007 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/openstack-galera-0" event={"ID":"92e3c4ec-dbf5-443c-b11a-7d1d961a92c0","Type":"ContainerStarted","Data":"0a36a7bf25e48ee762b0237b910deb8d449e32251eb166e6daa73e9503567487"} Mar 07 14:53:34 crc kubenswrapper[4943]: I0307 14:53:34.138429 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/openstack-galera-2" event={"ID":"64dcacc7-fd15-41e0-ac82-fc37aa562331","Type":"ContainerStarted","Data":"ac2328c3fb50b500200182d89ae00d95a6eaca6689d8151e8cc1cc335dbca435"} Mar 07 14:53:34 crc kubenswrapper[4943]: I0307 14:53:34.548630 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-6685f4bd5b-n5ksp"] Mar 07 14:53:34 crc kubenswrapper[4943]: I0307 14:53:34.549572 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-6685f4bd5b-n5ksp" Mar 07 14:53:34 crc kubenswrapper[4943]: I0307 14:53:34.552081 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-service-cert" Mar 07 14:53:34 crc kubenswrapper[4943]: I0307 14:53:34.552318 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-w6qgw" Mar 07 14:53:34 crc kubenswrapper[4943]: I0307 14:53:34.573154 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-6685f4bd5b-n5ksp"] Mar 07 14:53:34 crc kubenswrapper[4943]: I0307 14:53:34.651666 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e9e5a357-affd-43b8-8463-214c981786fa-webhook-cert\") pod \"infra-operator-controller-manager-6685f4bd5b-n5ksp\" (UID: \"e9e5a357-affd-43b8-8463-214c981786fa\") " pod="openstack-operators/infra-operator-controller-manager-6685f4bd5b-n5ksp" Mar 07 14:53:34 crc kubenswrapper[4943]: I0307 14:53:34.651707 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e9e5a357-affd-43b8-8463-214c981786fa-apiservice-cert\") pod \"infra-operator-controller-manager-6685f4bd5b-n5ksp\" (UID: \"e9e5a357-affd-43b8-8463-214c981786fa\") " pod="openstack-operators/infra-operator-controller-manager-6685f4bd5b-n5ksp" Mar 07 14:53:34 crc kubenswrapper[4943]: I0307 14:53:34.651763 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dp4v9\" (UniqueName: \"kubernetes.io/projected/e9e5a357-affd-43b8-8463-214c981786fa-kube-api-access-dp4v9\") pod \"infra-operator-controller-manager-6685f4bd5b-n5ksp\" (UID: \"e9e5a357-affd-43b8-8463-214c981786fa\") " pod="openstack-operators/infra-operator-controller-manager-6685f4bd5b-n5ksp" Mar 07 14:53:34 crc kubenswrapper[4943]: I0307 14:53:34.752648 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e9e5a357-affd-43b8-8463-214c981786fa-webhook-cert\") pod \"infra-operator-controller-manager-6685f4bd5b-n5ksp\" (UID: \"e9e5a357-affd-43b8-8463-214c981786fa\") " pod="openstack-operators/infra-operator-controller-manager-6685f4bd5b-n5ksp" Mar 07 14:53:34 crc kubenswrapper[4943]: I0307 14:53:34.752686 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e9e5a357-affd-43b8-8463-214c981786fa-apiservice-cert\") pod \"infra-operator-controller-manager-6685f4bd5b-n5ksp\" (UID: \"e9e5a357-affd-43b8-8463-214c981786fa\") " pod="openstack-operators/infra-operator-controller-manager-6685f4bd5b-n5ksp" Mar 07 14:53:34 crc kubenswrapper[4943]: I0307 14:53:34.752741 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dp4v9\" (UniqueName: \"kubernetes.io/projected/e9e5a357-affd-43b8-8463-214c981786fa-kube-api-access-dp4v9\") pod \"infra-operator-controller-manager-6685f4bd5b-n5ksp\" (UID: \"e9e5a357-affd-43b8-8463-214c981786fa\") " pod="openstack-operators/infra-operator-controller-manager-6685f4bd5b-n5ksp" Mar 07 14:53:34 crc kubenswrapper[4943]: I0307 14:53:34.760215 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e9e5a357-affd-43b8-8463-214c981786fa-webhook-cert\") pod \"infra-operator-controller-manager-6685f4bd5b-n5ksp\" (UID: \"e9e5a357-affd-43b8-8463-214c981786fa\") " pod="openstack-operators/infra-operator-controller-manager-6685f4bd5b-n5ksp" Mar 07 14:53:34 crc kubenswrapper[4943]: I0307 14:53:34.770736 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dp4v9\" (UniqueName: \"kubernetes.io/projected/e9e5a357-affd-43b8-8463-214c981786fa-kube-api-access-dp4v9\") pod \"infra-operator-controller-manager-6685f4bd5b-n5ksp\" (UID: \"e9e5a357-affd-43b8-8463-214c981786fa\") " pod="openstack-operators/infra-operator-controller-manager-6685f4bd5b-n5ksp" Mar 07 14:53:34 crc kubenswrapper[4943]: I0307 14:53:34.777874 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e9e5a357-affd-43b8-8463-214c981786fa-apiservice-cert\") pod \"infra-operator-controller-manager-6685f4bd5b-n5ksp\" (UID: \"e9e5a357-affd-43b8-8463-214c981786fa\") " pod="openstack-operators/infra-operator-controller-manager-6685f4bd5b-n5ksp" Mar 07 14:53:34 crc kubenswrapper[4943]: I0307 14:53:34.872294 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-6685f4bd5b-n5ksp" Mar 07 14:53:35 crc kubenswrapper[4943]: I0307 14:53:35.055032 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-6685f4bd5b-n5ksp"] Mar 07 14:53:35 crc kubenswrapper[4943]: W0307 14:53:35.073663 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode9e5a357_affd_43b8_8463_214c981786fa.slice/crio-bd9e3b731e79b3f5a0d9d1da2838b74fff5a93900ee54f3dfdaf24fc15d786a0 WatchSource:0}: Error finding container bd9e3b731e79b3f5a0d9d1da2838b74fff5a93900ee54f3dfdaf24fc15d786a0: Status 404 returned error can't find the container with id bd9e3b731e79b3f5a0d9d1da2838b74fff5a93900ee54f3dfdaf24fc15d786a0 Mar 07 14:53:35 crc kubenswrapper[4943]: I0307 14:53:35.148051 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6685f4bd5b-n5ksp" event={"ID":"e9e5a357-affd-43b8-8463-214c981786fa","Type":"ContainerStarted","Data":"bd9e3b731e79b3f5a0d9d1da2838b74fff5a93900ee54f3dfdaf24fc15d786a0"} Mar 07 14:53:43 crc kubenswrapper[4943]: I0307 14:53:43.219843 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/openstack-galera-0" event={"ID":"92e3c4ec-dbf5-443c-b11a-7d1d961a92c0","Type":"ContainerStarted","Data":"2260517bff30181455644b1924f9f271a6de9ec4f512dc6ad1a60c9c898067d0"} Mar 07 14:53:43 crc kubenswrapper[4943]: I0307 14:53:43.222796 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6685f4bd5b-n5ksp" event={"ID":"e9e5a357-affd-43b8-8463-214c981786fa","Type":"ContainerStarted","Data":"6fedee443640792ad02b6b598fe2f1ff0064ab8c7f779d69ef150e92457f4d11"} Mar 07 14:53:43 crc kubenswrapper[4943]: I0307 14:53:43.223048 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-6685f4bd5b-n5ksp" Mar 07 14:53:43 crc kubenswrapper[4943]: I0307 14:53:43.224952 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/openstack-galera-2" event={"ID":"64dcacc7-fd15-41e0-ac82-fc37aa562331","Type":"ContainerStarted","Data":"cc2c4a62b48be7058790907e45548490cb7e521b5fad549cea414ea2f5238661"} Mar 07 14:53:43 crc kubenswrapper[4943]: I0307 14:53:43.226787 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/openstack-galera-1" event={"ID":"4f2c813e-73c6-41c7-ba8d-5498fe1d5671","Type":"ContainerStarted","Data":"e25ce9f6e7d063c73199df9aafd3d5fc38bc62a2041729bb09b709d7a8f5247b"} Mar 07 14:53:43 crc kubenswrapper[4943]: I0307 14:53:43.295214 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-6685f4bd5b-n5ksp" podStartSLOduration=1.602191301 podStartE2EDuration="9.295191234s" podCreationTimestamp="2026-03-07 14:53:34 +0000 UTC" firstStartedPulling="2026-03-07 14:53:35.079785352 +0000 UTC m=+857.031921850" lastFinishedPulling="2026-03-07 14:53:42.772785285 +0000 UTC m=+864.724921783" observedRunningTime="2026-03-07 14:53:43.293720618 +0000 UTC m=+865.245857116" watchObservedRunningTime="2026-03-07 14:53:43.295191234 +0000 UTC m=+865.247327732" Mar 07 14:53:47 crc kubenswrapper[4943]: I0307 14:53:47.257773 4943 generic.go:334] "Generic (PLEG): container finished" podID="64dcacc7-fd15-41e0-ac82-fc37aa562331" containerID="cc2c4a62b48be7058790907e45548490cb7e521b5fad549cea414ea2f5238661" exitCode=0 Mar 07 14:53:47 crc kubenswrapper[4943]: I0307 14:53:47.258295 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/openstack-galera-2" event={"ID":"64dcacc7-fd15-41e0-ac82-fc37aa562331","Type":"ContainerDied","Data":"cc2c4a62b48be7058790907e45548490cb7e521b5fad549cea414ea2f5238661"} Mar 07 14:53:47 crc kubenswrapper[4943]: I0307 14:53:47.268292 4943 generic.go:334] "Generic (PLEG): container finished" podID="4f2c813e-73c6-41c7-ba8d-5498fe1d5671" containerID="e25ce9f6e7d063c73199df9aafd3d5fc38bc62a2041729bb09b709d7a8f5247b" exitCode=0 Mar 07 14:53:47 crc kubenswrapper[4943]: I0307 14:53:47.268380 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/openstack-galera-1" event={"ID":"4f2c813e-73c6-41c7-ba8d-5498fe1d5671","Type":"ContainerDied","Data":"e25ce9f6e7d063c73199df9aafd3d5fc38bc62a2041729bb09b709d7a8f5247b"} Mar 07 14:53:47 crc kubenswrapper[4943]: I0307 14:53:47.274541 4943 generic.go:334] "Generic (PLEG): container finished" podID="92e3c4ec-dbf5-443c-b11a-7d1d961a92c0" containerID="2260517bff30181455644b1924f9f271a6de9ec4f512dc6ad1a60c9c898067d0" exitCode=0 Mar 07 14:53:47 crc kubenswrapper[4943]: I0307 14:53:47.274589 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/openstack-galera-0" event={"ID":"92e3c4ec-dbf5-443c-b11a-7d1d961a92c0","Type":"ContainerDied","Data":"2260517bff30181455644b1924f9f271a6de9ec4f512dc6ad1a60c9c898067d0"} Mar 07 14:53:48 crc kubenswrapper[4943]: I0307 14:53:48.288730 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/openstack-galera-2" event={"ID":"64dcacc7-fd15-41e0-ac82-fc37aa562331","Type":"ContainerStarted","Data":"4b975693ea5742584be186e837486f65d74d3074aebef9792d4119aeb1ff55ec"} Mar 07 14:53:48 crc kubenswrapper[4943]: I0307 14:53:48.291975 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/openstack-galera-1" event={"ID":"4f2c813e-73c6-41c7-ba8d-5498fe1d5671","Type":"ContainerStarted","Data":"9f7dee8eb2f40d8bd6cd97437c8c932fcc104bbdf12bc1926bb6a36e29ec81a9"} Mar 07 14:53:48 crc kubenswrapper[4943]: I0307 14:53:48.295308 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/openstack-galera-0" event={"ID":"92e3c4ec-dbf5-443c-b11a-7d1d961a92c0","Type":"ContainerStarted","Data":"1c8612ce2b08f9bfb2d3e08c46ef9b247e33ae12b070574591af303f6754e797"} Mar 07 14:53:48 crc kubenswrapper[4943]: I0307 14:53:48.321462 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/openstack-galera-2" podStartSLOduration=7.750893693 podStartE2EDuration="17.321426008s" podCreationTimestamp="2026-03-07 14:53:31 +0000 UTC" firstStartedPulling="2026-03-07 14:53:33.251000378 +0000 UTC m=+855.203136876" lastFinishedPulling="2026-03-07 14:53:42.821532693 +0000 UTC m=+864.773669191" observedRunningTime="2026-03-07 14:53:48.317112602 +0000 UTC m=+870.269249150" watchObservedRunningTime="2026-03-07 14:53:48.321426008 +0000 UTC m=+870.273562546" Mar 07 14:53:48 crc kubenswrapper[4943]: I0307 14:53:48.343054 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/openstack-galera-0" podStartSLOduration=7.837978243 podStartE2EDuration="17.343028079s" podCreationTimestamp="2026-03-07 14:53:31 +0000 UTC" firstStartedPulling="2026-03-07 14:53:33.335145856 +0000 UTC m=+855.287282354" lastFinishedPulling="2026-03-07 14:53:42.840195682 +0000 UTC m=+864.792332190" observedRunningTime="2026-03-07 14:53:48.341747467 +0000 UTC m=+870.293884035" watchObservedRunningTime="2026-03-07 14:53:48.343028079 +0000 UTC m=+870.295164607" Mar 07 14:53:48 crc kubenswrapper[4943]: I0307 14:53:48.382884 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/openstack-galera-1" podStartSLOduration=8.025724496 podStartE2EDuration="17.382850937s" podCreationTimestamp="2026-03-07 14:53:31 +0000 UTC" firstStartedPulling="2026-03-07 14:53:33.442394491 +0000 UTC m=+855.394530989" lastFinishedPulling="2026-03-07 14:53:42.799520922 +0000 UTC m=+864.751657430" observedRunningTime="2026-03-07 14:53:48.379006353 +0000 UTC m=+870.331142891" watchObservedRunningTime="2026-03-07 14:53:48.382850937 +0000 UTC m=+870.334987475" Mar 07 14:53:52 crc kubenswrapper[4943]: I0307 14:53:52.745879 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="swift-kuttl-tests/openstack-galera-0" Mar 07 14:53:52 crc kubenswrapper[4943]: I0307 14:53:52.746968 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="swift-kuttl-tests/openstack-galera-0" Mar 07 14:53:52 crc kubenswrapper[4943]: I0307 14:53:52.787852 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="swift-kuttl-tests/openstack-galera-2" Mar 07 14:53:52 crc kubenswrapper[4943]: I0307 14:53:52.787999 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="swift-kuttl-tests/openstack-galera-2" Mar 07 14:53:53 crc kubenswrapper[4943]: I0307 14:53:53.068012 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="swift-kuttl-tests/openstack-galera-1" Mar 07 14:53:53 crc kubenswrapper[4943]: I0307 14:53:53.068496 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="swift-kuttl-tests/openstack-galera-1" Mar 07 14:53:54 crc kubenswrapper[4943]: I0307 14:53:54.877792 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-6685f4bd5b-n5ksp" Mar 07 14:53:56 crc kubenswrapper[4943]: I0307 14:53:56.198795 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/memcached-0"] Mar 07 14:53:56 crc kubenswrapper[4943]: I0307 14:53:56.200408 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/memcached-0" Mar 07 14:53:56 crc kubenswrapper[4943]: I0307 14:53:56.202219 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"memcached-config-data" Mar 07 14:53:56 crc kubenswrapper[4943]: I0307 14:53:56.202411 4943 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"memcached-memcached-dockercfg-9dmvn" Mar 07 14:53:56 crc kubenswrapper[4943]: I0307 14:53:56.221354 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/memcached-0"] Mar 07 14:53:56 crc kubenswrapper[4943]: I0307 14:53:56.312147 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5d85dcb0-3ed6-4fcf-8ee9-49e7c8cf3385-kolla-config\") pod \"memcached-0\" (UID: \"5d85dcb0-3ed6-4fcf-8ee9-49e7c8cf3385\") " pod="swift-kuttl-tests/memcached-0" Mar 07 14:53:56 crc kubenswrapper[4943]: I0307 14:53:56.312472 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5d85dcb0-3ed6-4fcf-8ee9-49e7c8cf3385-config-data\") pod \"memcached-0\" (UID: \"5d85dcb0-3ed6-4fcf-8ee9-49e7c8cf3385\") " pod="swift-kuttl-tests/memcached-0" Mar 07 14:53:56 crc kubenswrapper[4943]: I0307 14:53:56.312608 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5xsc\" (UniqueName: \"kubernetes.io/projected/5d85dcb0-3ed6-4fcf-8ee9-49e7c8cf3385-kube-api-access-h5xsc\") pod \"memcached-0\" (UID: \"5d85dcb0-3ed6-4fcf-8ee9-49e7c8cf3385\") " pod="swift-kuttl-tests/memcached-0" Mar 07 14:53:56 crc kubenswrapper[4943]: I0307 14:53:56.414037 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5xsc\" (UniqueName: \"kubernetes.io/projected/5d85dcb0-3ed6-4fcf-8ee9-49e7c8cf3385-kube-api-access-h5xsc\") pod \"memcached-0\" (UID: \"5d85dcb0-3ed6-4fcf-8ee9-49e7c8cf3385\") " pod="swift-kuttl-tests/memcached-0" Mar 07 14:53:56 crc kubenswrapper[4943]: I0307 14:53:56.414429 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5d85dcb0-3ed6-4fcf-8ee9-49e7c8cf3385-kolla-config\") pod \"memcached-0\" (UID: \"5d85dcb0-3ed6-4fcf-8ee9-49e7c8cf3385\") " pod="swift-kuttl-tests/memcached-0" Mar 07 14:53:56 crc kubenswrapper[4943]: I0307 14:53:56.414590 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5d85dcb0-3ed6-4fcf-8ee9-49e7c8cf3385-config-data\") pod \"memcached-0\" (UID: \"5d85dcb0-3ed6-4fcf-8ee9-49e7c8cf3385\") " pod="swift-kuttl-tests/memcached-0" Mar 07 14:53:56 crc kubenswrapper[4943]: I0307 14:53:56.415458 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5d85dcb0-3ed6-4fcf-8ee9-49e7c8cf3385-config-data\") pod \"memcached-0\" (UID: \"5d85dcb0-3ed6-4fcf-8ee9-49e7c8cf3385\") " pod="swift-kuttl-tests/memcached-0" Mar 07 14:53:56 crc kubenswrapper[4943]: I0307 14:53:56.415460 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5d85dcb0-3ed6-4fcf-8ee9-49e7c8cf3385-kolla-config\") pod \"memcached-0\" (UID: \"5d85dcb0-3ed6-4fcf-8ee9-49e7c8cf3385\") " pod="swift-kuttl-tests/memcached-0" Mar 07 14:53:56 crc kubenswrapper[4943]: I0307 14:53:56.437648 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5xsc\" (UniqueName: \"kubernetes.io/projected/5d85dcb0-3ed6-4fcf-8ee9-49e7c8cf3385-kube-api-access-h5xsc\") pod \"memcached-0\" (UID: \"5d85dcb0-3ed6-4fcf-8ee9-49e7c8cf3385\") " pod="swift-kuttl-tests/memcached-0" Mar 07 14:53:56 crc kubenswrapper[4943]: I0307 14:53:56.514271 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/memcached-0" Mar 07 14:53:56 crc kubenswrapper[4943]: I0307 14:53:56.769773 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/memcached-0"] Mar 07 14:53:56 crc kubenswrapper[4943]: I0307 14:53:56.912059 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="swift-kuttl-tests/openstack-galera-2" Mar 07 14:53:57 crc kubenswrapper[4943]: I0307 14:53:57.012329 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="swift-kuttl-tests/openstack-galera-2" Mar 07 14:53:57 crc kubenswrapper[4943]: I0307 14:53:57.088117 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-dbwn2"] Mar 07 14:53:57 crc kubenswrapper[4943]: I0307 14:53:57.089085 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-dbwn2" Mar 07 14:53:57 crc kubenswrapper[4943]: I0307 14:53:57.091878 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-index-dockercfg-6rwkl" Mar 07 14:53:57 crc kubenswrapper[4943]: I0307 14:53:57.094501 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-dbwn2"] Mar 07 14:53:57 crc kubenswrapper[4943]: I0307 14:53:57.227844 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bph52\" (UniqueName: \"kubernetes.io/projected/634a7b87-70f6-4eb5-ab4a-a520424dcec8-kube-api-access-bph52\") pod \"rabbitmq-cluster-operator-index-dbwn2\" (UID: \"634a7b87-70f6-4eb5-ab4a-a520424dcec8\") " pod="openstack-operators/rabbitmq-cluster-operator-index-dbwn2" Mar 07 14:53:57 crc kubenswrapper[4943]: I0307 14:53:57.329864 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bph52\" (UniqueName: \"kubernetes.io/projected/634a7b87-70f6-4eb5-ab4a-a520424dcec8-kube-api-access-bph52\") pod \"rabbitmq-cluster-operator-index-dbwn2\" (UID: \"634a7b87-70f6-4eb5-ab4a-a520424dcec8\") " pod="openstack-operators/rabbitmq-cluster-operator-index-dbwn2" Mar 07 14:53:57 crc kubenswrapper[4943]: I0307 14:53:57.359147 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/memcached-0" event={"ID":"5d85dcb0-3ed6-4fcf-8ee9-49e7c8cf3385","Type":"ContainerStarted","Data":"75bc68ef6cb4d21259e29551668b4ab1a33b514076d858d98043117df2d55458"} Mar 07 14:53:57 crc kubenswrapper[4943]: I0307 14:53:57.370909 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bph52\" (UniqueName: \"kubernetes.io/projected/634a7b87-70f6-4eb5-ab4a-a520424dcec8-kube-api-access-bph52\") pod \"rabbitmq-cluster-operator-index-dbwn2\" (UID: \"634a7b87-70f6-4eb5-ab4a-a520424dcec8\") " pod="openstack-operators/rabbitmq-cluster-operator-index-dbwn2" Mar 07 14:53:57 crc kubenswrapper[4943]: I0307 14:53:57.409820 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-dbwn2" Mar 07 14:53:57 crc kubenswrapper[4943]: I0307 14:53:57.837124 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-dbwn2"] Mar 07 14:53:57 crc kubenswrapper[4943]: W0307 14:53:57.846764 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod634a7b87_70f6_4eb5_ab4a_a520424dcec8.slice/crio-6c9bc1128b4bc7b7456095b5bb1c91e654167e4a79dc6fcde6582787d981ec17 WatchSource:0}: Error finding container 6c9bc1128b4bc7b7456095b5bb1c91e654167e4a79dc6fcde6582787d981ec17: Status 404 returned error can't find the container with id 6c9bc1128b4bc7b7456095b5bb1c91e654167e4a79dc6fcde6582787d981ec17 Mar 07 14:53:58 crc kubenswrapper[4943]: I0307 14:53:58.367397 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-dbwn2" event={"ID":"634a7b87-70f6-4eb5-ab4a-a520424dcec8","Type":"ContainerStarted","Data":"6c9bc1128b4bc7b7456095b5bb1c91e654167e4a79dc6fcde6582787d981ec17"} Mar 07 14:53:59 crc kubenswrapper[4943]: I0307 14:53:59.378773 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/memcached-0" event={"ID":"5d85dcb0-3ed6-4fcf-8ee9-49e7c8cf3385","Type":"ContainerStarted","Data":"f9a50e6c409650beec8658f58c922e08d9f3608a2587c22381db503ffcc1ea1c"} Mar 07 14:53:59 crc kubenswrapper[4943]: I0307 14:53:59.379230 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="swift-kuttl-tests/memcached-0" Mar 07 14:53:59 crc kubenswrapper[4943]: I0307 14:53:59.398902 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/memcached-0" podStartSLOduration=1.095324505 podStartE2EDuration="3.398882687s" podCreationTimestamp="2026-03-07 14:53:56 +0000 UTC" firstStartedPulling="2026-03-07 14:53:56.770910322 +0000 UTC m=+878.723046820" lastFinishedPulling="2026-03-07 14:53:59.074468494 +0000 UTC m=+881.026605002" observedRunningTime="2026-03-07 14:53:59.396620812 +0000 UTC m=+881.348757310" watchObservedRunningTime="2026-03-07 14:53:59.398882687 +0000 UTC m=+881.351019185" Mar 07 14:54:00 crc kubenswrapper[4943]: I0307 14:54:00.133359 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29548254-zpwrq"] Mar 07 14:54:00 crc kubenswrapper[4943]: I0307 14:54:00.135054 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548254-zpwrq" Mar 07 14:54:00 crc kubenswrapper[4943]: I0307 14:54:00.138374 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 07 14:54:00 crc kubenswrapper[4943]: I0307 14:54:00.138417 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 07 14:54:00 crc kubenswrapper[4943]: I0307 14:54:00.138979 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-qqlk5" Mar 07 14:54:00 crc kubenswrapper[4943]: I0307 14:54:00.139671 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29548254-zpwrq"] Mar 07 14:54:00 crc kubenswrapper[4943]: I0307 14:54:00.178589 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-457jm\" (UniqueName: \"kubernetes.io/projected/81098c4a-ecc7-4fff-a29f-f31365f42a62-kube-api-access-457jm\") pod \"auto-csr-approver-29548254-zpwrq\" (UID: \"81098c4a-ecc7-4fff-a29f-f31365f42a62\") " pod="openshift-infra/auto-csr-approver-29548254-zpwrq" Mar 07 14:54:00 crc kubenswrapper[4943]: I0307 14:54:00.280463 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-457jm\" (UniqueName: \"kubernetes.io/projected/81098c4a-ecc7-4fff-a29f-f31365f42a62-kube-api-access-457jm\") pod \"auto-csr-approver-29548254-zpwrq\" (UID: \"81098c4a-ecc7-4fff-a29f-f31365f42a62\") " pod="openshift-infra/auto-csr-approver-29548254-zpwrq" Mar 07 14:54:00 crc kubenswrapper[4943]: I0307 14:54:00.301131 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-457jm\" (UniqueName: \"kubernetes.io/projected/81098c4a-ecc7-4fff-a29f-f31365f42a62-kube-api-access-457jm\") pod \"auto-csr-approver-29548254-zpwrq\" (UID: \"81098c4a-ecc7-4fff-a29f-f31365f42a62\") " pod="openshift-infra/auto-csr-approver-29548254-zpwrq" Mar 07 14:54:00 crc kubenswrapper[4943]: I0307 14:54:00.460128 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548254-zpwrq" Mar 07 14:54:01 crc kubenswrapper[4943]: I0307 14:54:01.271851 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-dbwn2"] Mar 07 14:54:01 crc kubenswrapper[4943]: I0307 14:54:01.405747 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29548254-zpwrq"] Mar 07 14:54:01 crc kubenswrapper[4943]: W0307 14:54:01.430794 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod81098c4a_ecc7_4fff_a29f_f31365f42a62.slice/crio-75667ba61c1ff1015354700c366b1b2002fc113573b70186077007aa36feda2d WatchSource:0}: Error finding container 75667ba61c1ff1015354700c366b1b2002fc113573b70186077007aa36feda2d: Status 404 returned error can't find the container with id 75667ba61c1ff1015354700c366b1b2002fc113573b70186077007aa36feda2d Mar 07 14:54:01 crc kubenswrapper[4943]: I0307 14:54:01.455664 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/root-account-create-update-qkv6t"] Mar 07 14:54:01 crc kubenswrapper[4943]: I0307 14:54:01.456345 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/root-account-create-update-qkv6t" Mar 07 14:54:01 crc kubenswrapper[4943]: I0307 14:54:01.459264 4943 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"openstack-mariadb-root-db-secret" Mar 07 14:54:01 crc kubenswrapper[4943]: I0307 14:54:01.470912 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/root-account-create-update-qkv6t"] Mar 07 14:54:01 crc kubenswrapper[4943]: I0307 14:54:01.501570 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txt57\" (UniqueName: \"kubernetes.io/projected/21a7b33d-af38-44df-aca9-a602922a101f-kube-api-access-txt57\") pod \"root-account-create-update-qkv6t\" (UID: \"21a7b33d-af38-44df-aca9-a602922a101f\") " pod="swift-kuttl-tests/root-account-create-update-qkv6t" Mar 07 14:54:01 crc kubenswrapper[4943]: I0307 14:54:01.501664 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/21a7b33d-af38-44df-aca9-a602922a101f-operator-scripts\") pod \"root-account-create-update-qkv6t\" (UID: \"21a7b33d-af38-44df-aca9-a602922a101f\") " pod="swift-kuttl-tests/root-account-create-update-qkv6t" Mar 07 14:54:01 crc kubenswrapper[4943]: I0307 14:54:01.603214 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txt57\" (UniqueName: \"kubernetes.io/projected/21a7b33d-af38-44df-aca9-a602922a101f-kube-api-access-txt57\") pod \"root-account-create-update-qkv6t\" (UID: \"21a7b33d-af38-44df-aca9-a602922a101f\") " pod="swift-kuttl-tests/root-account-create-update-qkv6t" Mar 07 14:54:01 crc kubenswrapper[4943]: I0307 14:54:01.603298 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/21a7b33d-af38-44df-aca9-a602922a101f-operator-scripts\") pod \"root-account-create-update-qkv6t\" (UID: \"21a7b33d-af38-44df-aca9-a602922a101f\") " pod="swift-kuttl-tests/root-account-create-update-qkv6t" Mar 07 14:54:01 crc kubenswrapper[4943]: I0307 14:54:01.604165 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/21a7b33d-af38-44df-aca9-a602922a101f-operator-scripts\") pod \"root-account-create-update-qkv6t\" (UID: \"21a7b33d-af38-44df-aca9-a602922a101f\") " pod="swift-kuttl-tests/root-account-create-update-qkv6t" Mar 07 14:54:01 crc kubenswrapper[4943]: I0307 14:54:01.635783 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txt57\" (UniqueName: \"kubernetes.io/projected/21a7b33d-af38-44df-aca9-a602922a101f-kube-api-access-txt57\") pod \"root-account-create-update-qkv6t\" (UID: \"21a7b33d-af38-44df-aca9-a602922a101f\") " pod="swift-kuttl-tests/root-account-create-update-qkv6t" Mar 07 14:54:01 crc kubenswrapper[4943]: I0307 14:54:01.784522 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/root-account-create-update-qkv6t" Mar 07 14:54:01 crc kubenswrapper[4943]: I0307 14:54:01.878585 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-jcj6r"] Mar 07 14:54:01 crc kubenswrapper[4943]: I0307 14:54:01.879694 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-jcj6r" Mar 07 14:54:01 crc kubenswrapper[4943]: I0307 14:54:01.895574 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-jcj6r"] Mar 07 14:54:02 crc kubenswrapper[4943]: I0307 14:54:02.014878 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-csz5p\" (UniqueName: \"kubernetes.io/projected/541ac54d-ab66-40ac-a908-e38fc52e98cc-kube-api-access-csz5p\") pod \"rabbitmq-cluster-operator-index-jcj6r\" (UID: \"541ac54d-ab66-40ac-a908-e38fc52e98cc\") " pod="openstack-operators/rabbitmq-cluster-operator-index-jcj6r" Mar 07 14:54:02 crc kubenswrapper[4943]: I0307 14:54:02.116421 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-csz5p\" (UniqueName: \"kubernetes.io/projected/541ac54d-ab66-40ac-a908-e38fc52e98cc-kube-api-access-csz5p\") pod \"rabbitmq-cluster-operator-index-jcj6r\" (UID: \"541ac54d-ab66-40ac-a908-e38fc52e98cc\") " pod="openstack-operators/rabbitmq-cluster-operator-index-jcj6r" Mar 07 14:54:02 crc kubenswrapper[4943]: I0307 14:54:02.164205 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-csz5p\" (UniqueName: \"kubernetes.io/projected/541ac54d-ab66-40ac-a908-e38fc52e98cc-kube-api-access-csz5p\") pod \"rabbitmq-cluster-operator-index-jcj6r\" (UID: \"541ac54d-ab66-40ac-a908-e38fc52e98cc\") " pod="openstack-operators/rabbitmq-cluster-operator-index-jcj6r" Mar 07 14:54:02 crc kubenswrapper[4943]: I0307 14:54:02.195290 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/root-account-create-update-qkv6t"] Mar 07 14:54:02 crc kubenswrapper[4943]: I0307 14:54:02.204279 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-jcj6r" Mar 07 14:54:02 crc kubenswrapper[4943]: I0307 14:54:02.409500 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548254-zpwrq" event={"ID":"81098c4a-ecc7-4fff-a29f-f31365f42a62","Type":"ContainerStarted","Data":"75667ba61c1ff1015354700c366b1b2002fc113573b70186077007aa36feda2d"} Mar 07 14:54:02 crc kubenswrapper[4943]: I0307 14:54:02.870137 4943 prober.go:107] "Probe failed" probeType="Readiness" pod="swift-kuttl-tests/openstack-galera-2" podUID="64dcacc7-fd15-41e0-ac82-fc37aa562331" containerName="galera" probeResult="failure" output=< Mar 07 14:54:02 crc kubenswrapper[4943]: wsrep_local_state_comment (Donor/Desynced) differs from Synced Mar 07 14:54:02 crc kubenswrapper[4943]: > Mar 07 14:54:05 crc kubenswrapper[4943]: W0307 14:54:05.547264 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod21a7b33d_af38_44df_aca9_a602922a101f.slice/crio-742d090c3a6fbcf92f303df07c8da278aa0947fa7ea6dc86eb2df30a64b8fe0d WatchSource:0}: Error finding container 742d090c3a6fbcf92f303df07c8da278aa0947fa7ea6dc86eb2df30a64b8fe0d: Status 404 returned error can't find the container with id 742d090c3a6fbcf92f303df07c8da278aa0947fa7ea6dc86eb2df30a64b8fe0d Mar 07 14:54:06 crc kubenswrapper[4943]: I0307 14:54:06.065585 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-jcj6r"] Mar 07 14:54:06 crc kubenswrapper[4943]: W0307 14:54:06.070201 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod541ac54d_ab66_40ac_a908_e38fc52e98cc.slice/crio-cf8b688cdb980329663d522814e1fc3c4c84dc1ababd782fd975efb1bcbb3ef7 WatchSource:0}: Error finding container cf8b688cdb980329663d522814e1fc3c4c84dc1ababd782fd975efb1bcbb3ef7: Status 404 returned error can't find the container with id cf8b688cdb980329663d522814e1fc3c4c84dc1ababd782fd975efb1bcbb3ef7 Mar 07 14:54:06 crc kubenswrapper[4943]: I0307 14:54:06.434696 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/root-account-create-update-qkv6t" event={"ID":"21a7b33d-af38-44df-aca9-a602922a101f","Type":"ContainerStarted","Data":"9aeef101e414d6daf2c8c8b10d657d3c5e34ec1c372f98d99be7fc3609de1900"} Mar 07 14:54:06 crc kubenswrapper[4943]: I0307 14:54:06.434774 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/root-account-create-update-qkv6t" event={"ID":"21a7b33d-af38-44df-aca9-a602922a101f","Type":"ContainerStarted","Data":"742d090c3a6fbcf92f303df07c8da278aa0947fa7ea6dc86eb2df30a64b8fe0d"} Mar 07 14:54:06 crc kubenswrapper[4943]: I0307 14:54:06.438352 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-dbwn2" event={"ID":"634a7b87-70f6-4eb5-ab4a-a520424dcec8","Type":"ContainerStarted","Data":"143cec4f42756aa41405ad0b4f402dd67e037561e616b6c43336e700a8b2de01"} Mar 07 14:54:06 crc kubenswrapper[4943]: I0307 14:54:06.438572 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/rabbitmq-cluster-operator-index-dbwn2" podUID="634a7b87-70f6-4eb5-ab4a-a520424dcec8" containerName="registry-server" containerID="cri-o://143cec4f42756aa41405ad0b4f402dd67e037561e616b6c43336e700a8b2de01" gracePeriod=2 Mar 07 14:54:06 crc kubenswrapper[4943]: I0307 14:54:06.443477 4943 generic.go:334] "Generic (PLEG): container finished" podID="81098c4a-ecc7-4fff-a29f-f31365f42a62" containerID="09ac5748f6e67412ea6c6600ed76341524eb969dba0d55c3ffac79f034e4588a" exitCode=0 Mar 07 14:54:06 crc kubenswrapper[4943]: I0307 14:54:06.443567 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548254-zpwrq" event={"ID":"81098c4a-ecc7-4fff-a29f-f31365f42a62","Type":"ContainerDied","Data":"09ac5748f6e67412ea6c6600ed76341524eb969dba0d55c3ffac79f034e4588a"} Mar 07 14:54:06 crc kubenswrapper[4943]: I0307 14:54:06.445826 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-jcj6r" event={"ID":"541ac54d-ab66-40ac-a908-e38fc52e98cc","Type":"ContainerStarted","Data":"cf8b688cdb980329663d522814e1fc3c4c84dc1ababd782fd975efb1bcbb3ef7"} Mar 07 14:54:06 crc kubenswrapper[4943]: I0307 14:54:06.461198 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/root-account-create-update-qkv6t" podStartSLOduration=5.4611786 podStartE2EDuration="5.4611786s" podCreationTimestamp="2026-03-07 14:54:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:54:06.456507685 +0000 UTC m=+888.408644193" watchObservedRunningTime="2026-03-07 14:54:06.4611786 +0000 UTC m=+888.413315108" Mar 07 14:54:06 crc kubenswrapper[4943]: I0307 14:54:06.492699 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-index-dbwn2" podStartSLOduration=1.645513883 podStartE2EDuration="9.492676024s" podCreationTimestamp="2026-03-07 14:53:57 +0000 UTC" firstStartedPulling="2026-03-07 14:53:57.850555236 +0000 UTC m=+879.802691734" lastFinishedPulling="2026-03-07 14:54:05.697717367 +0000 UTC m=+887.649853875" observedRunningTime="2026-03-07 14:54:06.490792488 +0000 UTC m=+888.442929046" watchObservedRunningTime="2026-03-07 14:54:06.492676024 +0000 UTC m=+888.444812532" Mar 07 14:54:06 crc kubenswrapper[4943]: I0307 14:54:06.516316 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="swift-kuttl-tests/memcached-0" Mar 07 14:54:06 crc kubenswrapper[4943]: I0307 14:54:06.883437 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-dbwn2" Mar 07 14:54:06 crc kubenswrapper[4943]: I0307 14:54:06.997864 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bph52\" (UniqueName: \"kubernetes.io/projected/634a7b87-70f6-4eb5-ab4a-a520424dcec8-kube-api-access-bph52\") pod \"634a7b87-70f6-4eb5-ab4a-a520424dcec8\" (UID: \"634a7b87-70f6-4eb5-ab4a-a520424dcec8\") " Mar 07 14:54:07 crc kubenswrapper[4943]: I0307 14:54:07.004981 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/634a7b87-70f6-4eb5-ab4a-a520424dcec8-kube-api-access-bph52" (OuterVolumeSpecName: "kube-api-access-bph52") pod "634a7b87-70f6-4eb5-ab4a-a520424dcec8" (UID: "634a7b87-70f6-4eb5-ab4a-a520424dcec8"). InnerVolumeSpecName "kube-api-access-bph52". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:54:07 crc kubenswrapper[4943]: I0307 14:54:07.099626 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bph52\" (UniqueName: \"kubernetes.io/projected/634a7b87-70f6-4eb5-ab4a-a520424dcec8-kube-api-access-bph52\") on node \"crc\" DevicePath \"\"" Mar 07 14:54:07 crc kubenswrapper[4943]: I0307 14:54:07.455301 4943 generic.go:334] "Generic (PLEG): container finished" podID="21a7b33d-af38-44df-aca9-a602922a101f" containerID="9aeef101e414d6daf2c8c8b10d657d3c5e34ec1c372f98d99be7fc3609de1900" exitCode=0 Mar 07 14:54:07 crc kubenswrapper[4943]: I0307 14:54:07.455515 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/root-account-create-update-qkv6t" event={"ID":"21a7b33d-af38-44df-aca9-a602922a101f","Type":"ContainerDied","Data":"9aeef101e414d6daf2c8c8b10d657d3c5e34ec1c372f98d99be7fc3609de1900"} Mar 07 14:54:07 crc kubenswrapper[4943]: I0307 14:54:07.458102 4943 generic.go:334] "Generic (PLEG): container finished" podID="634a7b87-70f6-4eb5-ab4a-a520424dcec8" containerID="143cec4f42756aa41405ad0b4f402dd67e037561e616b6c43336e700a8b2de01" exitCode=0 Mar 07 14:54:07 crc kubenswrapper[4943]: I0307 14:54:07.458210 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-dbwn2" event={"ID":"634a7b87-70f6-4eb5-ab4a-a520424dcec8","Type":"ContainerDied","Data":"143cec4f42756aa41405ad0b4f402dd67e037561e616b6c43336e700a8b2de01"} Mar 07 14:54:07 crc kubenswrapper[4943]: I0307 14:54:07.458271 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-dbwn2" event={"ID":"634a7b87-70f6-4eb5-ab4a-a520424dcec8","Type":"ContainerDied","Data":"6c9bc1128b4bc7b7456095b5bb1c91e654167e4a79dc6fcde6582787d981ec17"} Mar 07 14:54:07 crc kubenswrapper[4943]: I0307 14:54:07.458301 4943 scope.go:117] "RemoveContainer" containerID="143cec4f42756aa41405ad0b4f402dd67e037561e616b6c43336e700a8b2de01" Mar 07 14:54:07 crc kubenswrapper[4943]: I0307 14:54:07.458190 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-dbwn2" Mar 07 14:54:07 crc kubenswrapper[4943]: I0307 14:54:07.460500 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-jcj6r" event={"ID":"541ac54d-ab66-40ac-a908-e38fc52e98cc","Type":"ContainerStarted","Data":"ff80516f9d8270adbbfc3d706372562f26bebcd57bdabcea7e82202bd20dc22a"} Mar 07 14:54:07 crc kubenswrapper[4943]: I0307 14:54:07.493500 4943 scope.go:117] "RemoveContainer" containerID="143cec4f42756aa41405ad0b4f402dd67e037561e616b6c43336e700a8b2de01" Mar 07 14:54:07 crc kubenswrapper[4943]: E0307 14:54:07.494004 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"143cec4f42756aa41405ad0b4f402dd67e037561e616b6c43336e700a8b2de01\": container with ID starting with 143cec4f42756aa41405ad0b4f402dd67e037561e616b6c43336e700a8b2de01 not found: ID does not exist" containerID="143cec4f42756aa41405ad0b4f402dd67e037561e616b6c43336e700a8b2de01" Mar 07 14:54:07 crc kubenswrapper[4943]: I0307 14:54:07.494048 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"143cec4f42756aa41405ad0b4f402dd67e037561e616b6c43336e700a8b2de01"} err="failed to get container status \"143cec4f42756aa41405ad0b4f402dd67e037561e616b6c43336e700a8b2de01\": rpc error: code = NotFound desc = could not find container \"143cec4f42756aa41405ad0b4f402dd67e037561e616b6c43336e700a8b2de01\": container with ID starting with 143cec4f42756aa41405ad0b4f402dd67e037561e616b6c43336e700a8b2de01 not found: ID does not exist" Mar 07 14:54:07 crc kubenswrapper[4943]: I0307 14:54:07.494856 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-index-jcj6r" podStartSLOduration=5.929469719 podStartE2EDuration="6.494841964s" podCreationTimestamp="2026-03-07 14:54:01 +0000 UTC" firstStartedPulling="2026-03-07 14:54:06.073682417 +0000 UTC m=+888.025818925" lastFinishedPulling="2026-03-07 14:54:06.639054662 +0000 UTC m=+888.591191170" observedRunningTime="2026-03-07 14:54:07.491264976 +0000 UTC m=+889.443401484" watchObservedRunningTime="2026-03-07 14:54:07.494841964 +0000 UTC m=+889.446978482" Mar 07 14:54:07 crc kubenswrapper[4943]: I0307 14:54:07.515309 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-dbwn2"] Mar 07 14:54:07 crc kubenswrapper[4943]: I0307 14:54:07.521097 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-dbwn2"] Mar 07 14:54:07 crc kubenswrapper[4943]: I0307 14:54:07.803745 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548254-zpwrq" Mar 07 14:54:07 crc kubenswrapper[4943]: I0307 14:54:07.911487 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-457jm\" (UniqueName: \"kubernetes.io/projected/81098c4a-ecc7-4fff-a29f-f31365f42a62-kube-api-access-457jm\") pod \"81098c4a-ecc7-4fff-a29f-f31365f42a62\" (UID: \"81098c4a-ecc7-4fff-a29f-f31365f42a62\") " Mar 07 14:54:07 crc kubenswrapper[4943]: I0307 14:54:07.918295 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81098c4a-ecc7-4fff-a29f-f31365f42a62-kube-api-access-457jm" (OuterVolumeSpecName: "kube-api-access-457jm") pod "81098c4a-ecc7-4fff-a29f-f31365f42a62" (UID: "81098c4a-ecc7-4fff-a29f-f31365f42a62"). InnerVolumeSpecName "kube-api-access-457jm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:54:08 crc kubenswrapper[4943]: I0307 14:54:08.013694 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-457jm\" (UniqueName: \"kubernetes.io/projected/81098c4a-ecc7-4fff-a29f-f31365f42a62-kube-api-access-457jm\") on node \"crc\" DevicePath \"\"" Mar 07 14:54:08 crc kubenswrapper[4943]: I0307 14:54:08.471233 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548254-zpwrq" event={"ID":"81098c4a-ecc7-4fff-a29f-f31365f42a62","Type":"ContainerDied","Data":"75667ba61c1ff1015354700c366b1b2002fc113573b70186077007aa36feda2d"} Mar 07 14:54:08 crc kubenswrapper[4943]: I0307 14:54:08.471295 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="75667ba61c1ff1015354700c366b1b2002fc113573b70186077007aa36feda2d" Mar 07 14:54:08 crc kubenswrapper[4943]: I0307 14:54:08.471341 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548254-zpwrq" Mar 07 14:54:08 crc kubenswrapper[4943]: I0307 14:54:08.774687 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="634a7b87-70f6-4eb5-ab4a-a520424dcec8" path="/var/lib/kubelet/pods/634a7b87-70f6-4eb5-ab4a-a520424dcec8/volumes" Mar 07 14:54:08 crc kubenswrapper[4943]: I0307 14:54:08.807687 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/root-account-create-update-qkv6t" Mar 07 14:54:08 crc kubenswrapper[4943]: I0307 14:54:08.862989 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29548248-2459n"] Mar 07 14:54:08 crc kubenswrapper[4943]: I0307 14:54:08.870416 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29548248-2459n"] Mar 07 14:54:08 crc kubenswrapper[4943]: I0307 14:54:08.927341 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-txt57\" (UniqueName: \"kubernetes.io/projected/21a7b33d-af38-44df-aca9-a602922a101f-kube-api-access-txt57\") pod \"21a7b33d-af38-44df-aca9-a602922a101f\" (UID: \"21a7b33d-af38-44df-aca9-a602922a101f\") " Mar 07 14:54:08 crc kubenswrapper[4943]: I0307 14:54:08.927398 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/21a7b33d-af38-44df-aca9-a602922a101f-operator-scripts\") pod \"21a7b33d-af38-44df-aca9-a602922a101f\" (UID: \"21a7b33d-af38-44df-aca9-a602922a101f\") " Mar 07 14:54:08 crc kubenswrapper[4943]: I0307 14:54:08.928925 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21a7b33d-af38-44df-aca9-a602922a101f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "21a7b33d-af38-44df-aca9-a602922a101f" (UID: "21a7b33d-af38-44df-aca9-a602922a101f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:54:08 crc kubenswrapper[4943]: I0307 14:54:08.941137 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21a7b33d-af38-44df-aca9-a602922a101f-kube-api-access-txt57" (OuterVolumeSpecName: "kube-api-access-txt57") pod "21a7b33d-af38-44df-aca9-a602922a101f" (UID: "21a7b33d-af38-44df-aca9-a602922a101f"). InnerVolumeSpecName "kube-api-access-txt57". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:54:09 crc kubenswrapper[4943]: I0307 14:54:09.029083 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-txt57\" (UniqueName: \"kubernetes.io/projected/21a7b33d-af38-44df-aca9-a602922a101f-kube-api-access-txt57\") on node \"crc\" DevicePath \"\"" Mar 07 14:54:09 crc kubenswrapper[4943]: I0307 14:54:09.029123 4943 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/21a7b33d-af38-44df-aca9-a602922a101f-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 14:54:09 crc kubenswrapper[4943]: I0307 14:54:09.479947 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/root-account-create-update-qkv6t" event={"ID":"21a7b33d-af38-44df-aca9-a602922a101f","Type":"ContainerDied","Data":"742d090c3a6fbcf92f303df07c8da278aa0947fa7ea6dc86eb2df30a64b8fe0d"} Mar 07 14:54:09 crc kubenswrapper[4943]: I0307 14:54:09.479989 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="742d090c3a6fbcf92f303df07c8da278aa0947fa7ea6dc86eb2df30a64b8fe0d" Mar 07 14:54:09 crc kubenswrapper[4943]: I0307 14:54:09.480009 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/root-account-create-update-qkv6t" Mar 07 14:54:10 crc kubenswrapper[4943]: I0307 14:54:10.682253 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="swift-kuttl-tests/openstack-galera-1" Mar 07 14:54:10 crc kubenswrapper[4943]: I0307 14:54:10.766231 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1349e3eb-9576-43fe-bd36-6e2832255ae4" path="/var/lib/kubelet/pods/1349e3eb-9576-43fe-bd36-6e2832255ae4/volumes" Mar 07 14:54:10 crc kubenswrapper[4943]: I0307 14:54:10.780462 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="swift-kuttl-tests/openstack-galera-1" Mar 07 14:54:12 crc kubenswrapper[4943]: I0307 14:54:12.206047 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/rabbitmq-cluster-operator-index-jcj6r" Mar 07 14:54:12 crc kubenswrapper[4943]: I0307 14:54:12.206458 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/rabbitmq-cluster-operator-index-jcj6r" Mar 07 14:54:12 crc kubenswrapper[4943]: I0307 14:54:12.238220 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/rabbitmq-cluster-operator-index-jcj6r" Mar 07 14:54:12 crc kubenswrapper[4943]: I0307 14:54:12.550286 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/rabbitmq-cluster-operator-index-jcj6r" Mar 07 14:54:13 crc kubenswrapper[4943]: I0307 14:54:13.095798 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="swift-kuttl-tests/openstack-galera-0" Mar 07 14:54:13 crc kubenswrapper[4943]: I0307 14:54:13.350589 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="swift-kuttl-tests/openstack-galera-0" Mar 07 14:54:13 crc kubenswrapper[4943]: I0307 14:54:13.527332 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590mkxwv"] Mar 07 14:54:13 crc kubenswrapper[4943]: E0307 14:54:13.527683 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="634a7b87-70f6-4eb5-ab4a-a520424dcec8" containerName="registry-server" Mar 07 14:54:13 crc kubenswrapper[4943]: I0307 14:54:13.527711 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="634a7b87-70f6-4eb5-ab4a-a520424dcec8" containerName="registry-server" Mar 07 14:54:13 crc kubenswrapper[4943]: E0307 14:54:13.527740 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81098c4a-ecc7-4fff-a29f-f31365f42a62" containerName="oc" Mar 07 14:54:13 crc kubenswrapper[4943]: I0307 14:54:13.527753 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="81098c4a-ecc7-4fff-a29f-f31365f42a62" containerName="oc" Mar 07 14:54:13 crc kubenswrapper[4943]: E0307 14:54:13.527783 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21a7b33d-af38-44df-aca9-a602922a101f" containerName="mariadb-account-create-update" Mar 07 14:54:13 crc kubenswrapper[4943]: I0307 14:54:13.527798 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="21a7b33d-af38-44df-aca9-a602922a101f" containerName="mariadb-account-create-update" Mar 07 14:54:13 crc kubenswrapper[4943]: I0307 14:54:13.528006 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="21a7b33d-af38-44df-aca9-a602922a101f" containerName="mariadb-account-create-update" Mar 07 14:54:13 crc kubenswrapper[4943]: I0307 14:54:13.528051 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="81098c4a-ecc7-4fff-a29f-f31365f42a62" containerName="oc" Mar 07 14:54:13 crc kubenswrapper[4943]: I0307 14:54:13.528069 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="634a7b87-70f6-4eb5-ab4a-a520424dcec8" containerName="registry-server" Mar 07 14:54:13 crc kubenswrapper[4943]: I0307 14:54:13.529405 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590mkxwv" Mar 07 14:54:13 crc kubenswrapper[4943]: I0307 14:54:13.531447 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-k6gp2" Mar 07 14:54:13 crc kubenswrapper[4943]: I0307 14:54:13.547791 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590mkxwv"] Mar 07 14:54:13 crc kubenswrapper[4943]: I0307 14:54:13.597318 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e6788dcd-2850-4113-94e7-204d8028d32a-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590mkxwv\" (UID: \"e6788dcd-2850-4113-94e7-204d8028d32a\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590mkxwv" Mar 07 14:54:13 crc kubenswrapper[4943]: I0307 14:54:13.597394 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krrcv\" (UniqueName: \"kubernetes.io/projected/e6788dcd-2850-4113-94e7-204d8028d32a-kube-api-access-krrcv\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590mkxwv\" (UID: \"e6788dcd-2850-4113-94e7-204d8028d32a\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590mkxwv" Mar 07 14:54:13 crc kubenswrapper[4943]: I0307 14:54:13.597591 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e6788dcd-2850-4113-94e7-204d8028d32a-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590mkxwv\" (UID: \"e6788dcd-2850-4113-94e7-204d8028d32a\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590mkxwv" Mar 07 14:54:13 crc kubenswrapper[4943]: I0307 14:54:13.699465 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e6788dcd-2850-4113-94e7-204d8028d32a-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590mkxwv\" (UID: \"e6788dcd-2850-4113-94e7-204d8028d32a\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590mkxwv" Mar 07 14:54:13 crc kubenswrapper[4943]: I0307 14:54:13.699535 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e6788dcd-2850-4113-94e7-204d8028d32a-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590mkxwv\" (UID: \"e6788dcd-2850-4113-94e7-204d8028d32a\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590mkxwv" Mar 07 14:54:13 crc kubenswrapper[4943]: I0307 14:54:13.699570 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krrcv\" (UniqueName: \"kubernetes.io/projected/e6788dcd-2850-4113-94e7-204d8028d32a-kube-api-access-krrcv\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590mkxwv\" (UID: \"e6788dcd-2850-4113-94e7-204d8028d32a\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590mkxwv" Mar 07 14:54:13 crc kubenswrapper[4943]: I0307 14:54:13.700164 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e6788dcd-2850-4113-94e7-204d8028d32a-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590mkxwv\" (UID: \"e6788dcd-2850-4113-94e7-204d8028d32a\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590mkxwv" Mar 07 14:54:13 crc kubenswrapper[4943]: I0307 14:54:13.700281 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e6788dcd-2850-4113-94e7-204d8028d32a-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590mkxwv\" (UID: \"e6788dcd-2850-4113-94e7-204d8028d32a\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590mkxwv" Mar 07 14:54:13 crc kubenswrapper[4943]: I0307 14:54:13.732736 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krrcv\" (UniqueName: \"kubernetes.io/projected/e6788dcd-2850-4113-94e7-204d8028d32a-kube-api-access-krrcv\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590mkxwv\" (UID: \"e6788dcd-2850-4113-94e7-204d8028d32a\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590mkxwv" Mar 07 14:54:13 crc kubenswrapper[4943]: I0307 14:54:13.847906 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590mkxwv" Mar 07 14:54:14 crc kubenswrapper[4943]: I0307 14:54:14.322047 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590mkxwv"] Mar 07 14:54:14 crc kubenswrapper[4943]: I0307 14:54:14.533974 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590mkxwv" event={"ID":"e6788dcd-2850-4113-94e7-204d8028d32a","Type":"ContainerStarted","Data":"1a56dd64c639fe586e582a7197cfcc6bf1bfcf645ce765a66e06a9bb16bd8420"} Mar 07 14:54:14 crc kubenswrapper[4943]: I0307 14:54:14.536128 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590mkxwv" event={"ID":"e6788dcd-2850-4113-94e7-204d8028d32a","Type":"ContainerStarted","Data":"40edf65fc5fc04262bd973139ab183f6502f06883cba4d57b5bb1bc589d2af0c"} Mar 07 14:54:15 crc kubenswrapper[4943]: I0307 14:54:15.544694 4943 generic.go:334] "Generic (PLEG): container finished" podID="e6788dcd-2850-4113-94e7-204d8028d32a" containerID="1a56dd64c639fe586e582a7197cfcc6bf1bfcf645ce765a66e06a9bb16bd8420" exitCode=0 Mar 07 14:54:15 crc kubenswrapper[4943]: I0307 14:54:15.544793 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590mkxwv" event={"ID":"e6788dcd-2850-4113-94e7-204d8028d32a","Type":"ContainerDied","Data":"1a56dd64c639fe586e582a7197cfcc6bf1bfcf645ce765a66e06a9bb16bd8420"} Mar 07 14:54:17 crc kubenswrapper[4943]: I0307 14:54:17.569387 4943 generic.go:334] "Generic (PLEG): container finished" podID="e6788dcd-2850-4113-94e7-204d8028d32a" containerID="27c72d7dcec14ee10b120e55200eb184b000d2e522aad3ad6a593914f9c0d267" exitCode=0 Mar 07 14:54:17 crc kubenswrapper[4943]: I0307 14:54:17.569474 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590mkxwv" event={"ID":"e6788dcd-2850-4113-94e7-204d8028d32a","Type":"ContainerDied","Data":"27c72d7dcec14ee10b120e55200eb184b000d2e522aad3ad6a593914f9c0d267"} Mar 07 14:54:18 crc kubenswrapper[4943]: I0307 14:54:18.581828 4943 generic.go:334] "Generic (PLEG): container finished" podID="e6788dcd-2850-4113-94e7-204d8028d32a" containerID="d8fe08319a6b9cd101056d6b766c467383770b54a69483b9541a46885431a923" exitCode=0 Mar 07 14:54:18 crc kubenswrapper[4943]: I0307 14:54:18.581921 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590mkxwv" event={"ID":"e6788dcd-2850-4113-94e7-204d8028d32a","Type":"ContainerDied","Data":"d8fe08319a6b9cd101056d6b766c467383770b54a69483b9541a46885431a923"} Mar 07 14:54:19 crc kubenswrapper[4943]: I0307 14:54:19.568255 4943 scope.go:117] "RemoveContainer" containerID="c3c1ce363e4f0f7a4f265024b98311e29e9ea1ea4b506f5fff498cb14b3ee2fd" Mar 07 14:54:19 crc kubenswrapper[4943]: I0307 14:54:19.997196 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590mkxwv" Mar 07 14:54:20 crc kubenswrapper[4943]: I0307 14:54:20.121233 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e6788dcd-2850-4113-94e7-204d8028d32a-bundle\") pod \"e6788dcd-2850-4113-94e7-204d8028d32a\" (UID: \"e6788dcd-2850-4113-94e7-204d8028d32a\") " Mar 07 14:54:20 crc kubenswrapper[4943]: I0307 14:54:20.121287 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e6788dcd-2850-4113-94e7-204d8028d32a-util\") pod \"e6788dcd-2850-4113-94e7-204d8028d32a\" (UID: \"e6788dcd-2850-4113-94e7-204d8028d32a\") " Mar 07 14:54:20 crc kubenswrapper[4943]: I0307 14:54:20.121339 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-krrcv\" (UniqueName: \"kubernetes.io/projected/e6788dcd-2850-4113-94e7-204d8028d32a-kube-api-access-krrcv\") pod \"e6788dcd-2850-4113-94e7-204d8028d32a\" (UID: \"e6788dcd-2850-4113-94e7-204d8028d32a\") " Mar 07 14:54:20 crc kubenswrapper[4943]: I0307 14:54:20.123418 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e6788dcd-2850-4113-94e7-204d8028d32a-bundle" (OuterVolumeSpecName: "bundle") pod "e6788dcd-2850-4113-94e7-204d8028d32a" (UID: "e6788dcd-2850-4113-94e7-204d8028d32a"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:54:20 crc kubenswrapper[4943]: I0307 14:54:20.140152 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6788dcd-2850-4113-94e7-204d8028d32a-kube-api-access-krrcv" (OuterVolumeSpecName: "kube-api-access-krrcv") pod "e6788dcd-2850-4113-94e7-204d8028d32a" (UID: "e6788dcd-2850-4113-94e7-204d8028d32a"). InnerVolumeSpecName "kube-api-access-krrcv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:54:20 crc kubenswrapper[4943]: I0307 14:54:20.222955 4943 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e6788dcd-2850-4113-94e7-204d8028d32a-bundle\") on node \"crc\" DevicePath \"\"" Mar 07 14:54:20 crc kubenswrapper[4943]: I0307 14:54:20.222999 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-krrcv\" (UniqueName: \"kubernetes.io/projected/e6788dcd-2850-4113-94e7-204d8028d32a-kube-api-access-krrcv\") on node \"crc\" DevicePath \"\"" Mar 07 14:54:20 crc kubenswrapper[4943]: I0307 14:54:20.364371 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e6788dcd-2850-4113-94e7-204d8028d32a-util" (OuterVolumeSpecName: "util") pod "e6788dcd-2850-4113-94e7-204d8028d32a" (UID: "e6788dcd-2850-4113-94e7-204d8028d32a"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:54:20 crc kubenswrapper[4943]: I0307 14:54:20.426128 4943 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e6788dcd-2850-4113-94e7-204d8028d32a-util\") on node \"crc\" DevicePath \"\"" Mar 07 14:54:20 crc kubenswrapper[4943]: I0307 14:54:20.602139 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590mkxwv" event={"ID":"e6788dcd-2850-4113-94e7-204d8028d32a","Type":"ContainerDied","Data":"40edf65fc5fc04262bd973139ab183f6502f06883cba4d57b5bb1bc589d2af0c"} Mar 07 14:54:20 crc kubenswrapper[4943]: I0307 14:54:20.602193 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="40edf65fc5fc04262bd973139ab183f6502f06883cba4d57b5bb1bc589d2af0c" Mar 07 14:54:20 crc kubenswrapper[4943]: I0307 14:54:20.602238 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590mkxwv" Mar 07 14:54:21 crc kubenswrapper[4943]: I0307 14:54:21.483231 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lfm7s"] Mar 07 14:54:21 crc kubenswrapper[4943]: E0307 14:54:21.485100 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6788dcd-2850-4113-94e7-204d8028d32a" containerName="extract" Mar 07 14:54:21 crc kubenswrapper[4943]: I0307 14:54:21.485246 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6788dcd-2850-4113-94e7-204d8028d32a" containerName="extract" Mar 07 14:54:21 crc kubenswrapper[4943]: E0307 14:54:21.485359 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6788dcd-2850-4113-94e7-204d8028d32a" containerName="util" Mar 07 14:54:21 crc kubenswrapper[4943]: I0307 14:54:21.485451 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6788dcd-2850-4113-94e7-204d8028d32a" containerName="util" Mar 07 14:54:21 crc kubenswrapper[4943]: E0307 14:54:21.485569 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6788dcd-2850-4113-94e7-204d8028d32a" containerName="pull" Mar 07 14:54:21 crc kubenswrapper[4943]: I0307 14:54:21.485670 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6788dcd-2850-4113-94e7-204d8028d32a" containerName="pull" Mar 07 14:54:21 crc kubenswrapper[4943]: I0307 14:54:21.485961 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6788dcd-2850-4113-94e7-204d8028d32a" containerName="extract" Mar 07 14:54:21 crc kubenswrapper[4943]: I0307 14:54:21.489836 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lfm7s" Mar 07 14:54:21 crc kubenswrapper[4943]: I0307 14:54:21.517692 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lfm7s"] Mar 07 14:54:21 crc kubenswrapper[4943]: I0307 14:54:21.540848 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wppnr\" (UniqueName: \"kubernetes.io/projected/79dd2902-63b6-4e27-b50e-bc5e5553af5e-kube-api-access-wppnr\") pod \"redhat-marketplace-lfm7s\" (UID: \"79dd2902-63b6-4e27-b50e-bc5e5553af5e\") " pod="openshift-marketplace/redhat-marketplace-lfm7s" Mar 07 14:54:21 crc kubenswrapper[4943]: I0307 14:54:21.540904 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79dd2902-63b6-4e27-b50e-bc5e5553af5e-utilities\") pod \"redhat-marketplace-lfm7s\" (UID: \"79dd2902-63b6-4e27-b50e-bc5e5553af5e\") " pod="openshift-marketplace/redhat-marketplace-lfm7s" Mar 07 14:54:21 crc kubenswrapper[4943]: I0307 14:54:21.540973 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79dd2902-63b6-4e27-b50e-bc5e5553af5e-catalog-content\") pod \"redhat-marketplace-lfm7s\" (UID: \"79dd2902-63b6-4e27-b50e-bc5e5553af5e\") " pod="openshift-marketplace/redhat-marketplace-lfm7s" Mar 07 14:54:21 crc kubenswrapper[4943]: I0307 14:54:21.642297 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79dd2902-63b6-4e27-b50e-bc5e5553af5e-catalog-content\") pod \"redhat-marketplace-lfm7s\" (UID: \"79dd2902-63b6-4e27-b50e-bc5e5553af5e\") " pod="openshift-marketplace/redhat-marketplace-lfm7s" Mar 07 14:54:21 crc kubenswrapper[4943]: I0307 14:54:21.642393 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wppnr\" (UniqueName: \"kubernetes.io/projected/79dd2902-63b6-4e27-b50e-bc5e5553af5e-kube-api-access-wppnr\") pod \"redhat-marketplace-lfm7s\" (UID: \"79dd2902-63b6-4e27-b50e-bc5e5553af5e\") " pod="openshift-marketplace/redhat-marketplace-lfm7s" Mar 07 14:54:21 crc kubenswrapper[4943]: I0307 14:54:21.642417 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79dd2902-63b6-4e27-b50e-bc5e5553af5e-utilities\") pod \"redhat-marketplace-lfm7s\" (UID: \"79dd2902-63b6-4e27-b50e-bc5e5553af5e\") " pod="openshift-marketplace/redhat-marketplace-lfm7s" Mar 07 14:54:21 crc kubenswrapper[4943]: I0307 14:54:21.642971 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79dd2902-63b6-4e27-b50e-bc5e5553af5e-catalog-content\") pod \"redhat-marketplace-lfm7s\" (UID: \"79dd2902-63b6-4e27-b50e-bc5e5553af5e\") " pod="openshift-marketplace/redhat-marketplace-lfm7s" Mar 07 14:54:21 crc kubenswrapper[4943]: I0307 14:54:21.643015 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79dd2902-63b6-4e27-b50e-bc5e5553af5e-utilities\") pod \"redhat-marketplace-lfm7s\" (UID: \"79dd2902-63b6-4e27-b50e-bc5e5553af5e\") " pod="openshift-marketplace/redhat-marketplace-lfm7s" Mar 07 14:54:21 crc kubenswrapper[4943]: I0307 14:54:21.670222 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wppnr\" (UniqueName: \"kubernetes.io/projected/79dd2902-63b6-4e27-b50e-bc5e5553af5e-kube-api-access-wppnr\") pod \"redhat-marketplace-lfm7s\" (UID: \"79dd2902-63b6-4e27-b50e-bc5e5553af5e\") " pod="openshift-marketplace/redhat-marketplace-lfm7s" Mar 07 14:54:21 crc kubenswrapper[4943]: I0307 14:54:21.825269 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lfm7s" Mar 07 14:54:22 crc kubenswrapper[4943]: I0307 14:54:22.080767 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lfm7s"] Mar 07 14:54:22 crc kubenswrapper[4943]: I0307 14:54:22.620226 4943 generic.go:334] "Generic (PLEG): container finished" podID="79dd2902-63b6-4e27-b50e-bc5e5553af5e" containerID="c891f1cd14b68e7e0fdd5695de2818629a7d39504759045f1f486ff491d7338d" exitCode=0 Mar 07 14:54:22 crc kubenswrapper[4943]: I0307 14:54:22.620293 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lfm7s" event={"ID":"79dd2902-63b6-4e27-b50e-bc5e5553af5e","Type":"ContainerDied","Data":"c891f1cd14b68e7e0fdd5695de2818629a7d39504759045f1f486ff491d7338d"} Mar 07 14:54:22 crc kubenswrapper[4943]: I0307 14:54:22.620339 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lfm7s" event={"ID":"79dd2902-63b6-4e27-b50e-bc5e5553af5e","Type":"ContainerStarted","Data":"0eb7de7c6bed0c49027fdacbef373caf99828d9f8a366c04e6b703bb773e2c76"} Mar 07 14:54:24 crc kubenswrapper[4943]: I0307 14:54:24.651467 4943 generic.go:334] "Generic (PLEG): container finished" podID="79dd2902-63b6-4e27-b50e-bc5e5553af5e" containerID="bde325f01af3d4e520034311c5b4c3b00dd00dcdd0e47b104d4eda480787b38e" exitCode=0 Mar 07 14:54:24 crc kubenswrapper[4943]: I0307 14:54:24.651541 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lfm7s" event={"ID":"79dd2902-63b6-4e27-b50e-bc5e5553af5e","Type":"ContainerDied","Data":"bde325f01af3d4e520034311c5b4c3b00dd00dcdd0e47b104d4eda480787b38e"} Mar 07 14:54:25 crc kubenswrapper[4943]: I0307 14:54:25.663590 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lfm7s" event={"ID":"79dd2902-63b6-4e27-b50e-bc5e5553af5e","Type":"ContainerStarted","Data":"5a6fdad7eb933814aa708c0cf0b0542d6b46d6a679a15e886339de6d1d223b39"} Mar 07 14:54:25 crc kubenswrapper[4943]: I0307 14:54:25.692419 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lfm7s" podStartSLOduration=2.246225199 podStartE2EDuration="4.692393336s" podCreationTimestamp="2026-03-07 14:54:21 +0000 UTC" firstStartedPulling="2026-03-07 14:54:22.623062524 +0000 UTC m=+904.575199052" lastFinishedPulling="2026-03-07 14:54:25.069230691 +0000 UTC m=+907.021367189" observedRunningTime="2026-03-07 14:54:25.684304737 +0000 UTC m=+907.636441265" watchObservedRunningTime="2026-03-07 14:54:25.692393336 +0000 UTC m=+907.644529834" Mar 07 14:54:29 crc kubenswrapper[4943]: I0307 14:54:29.925483 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-42t4c"] Mar 07 14:54:29 crc kubenswrapper[4943]: I0307 14:54:29.927230 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-42t4c" Mar 07 14:54:29 crc kubenswrapper[4943]: I0307 14:54:29.932086 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-dockercfg-459cw" Mar 07 14:54:29 crc kubenswrapper[4943]: I0307 14:54:29.936243 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-42t4c"] Mar 07 14:54:29 crc kubenswrapper[4943]: I0307 14:54:29.960495 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjzzb\" (UniqueName: \"kubernetes.io/projected/0b3e1865-67b8-4e70-ae06-e0c41256e303-kube-api-access-zjzzb\") pod \"rabbitmq-cluster-operator-779fc9694b-42t4c\" (UID: \"0b3e1865-67b8-4e70-ae06-e0c41256e303\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-42t4c" Mar 07 14:54:30 crc kubenswrapper[4943]: I0307 14:54:30.062466 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjzzb\" (UniqueName: \"kubernetes.io/projected/0b3e1865-67b8-4e70-ae06-e0c41256e303-kube-api-access-zjzzb\") pod \"rabbitmq-cluster-operator-779fc9694b-42t4c\" (UID: \"0b3e1865-67b8-4e70-ae06-e0c41256e303\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-42t4c" Mar 07 14:54:30 crc kubenswrapper[4943]: I0307 14:54:30.093093 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjzzb\" (UniqueName: \"kubernetes.io/projected/0b3e1865-67b8-4e70-ae06-e0c41256e303-kube-api-access-zjzzb\") pod \"rabbitmq-cluster-operator-779fc9694b-42t4c\" (UID: \"0b3e1865-67b8-4e70-ae06-e0c41256e303\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-42t4c" Mar 07 14:54:30 crc kubenswrapper[4943]: I0307 14:54:30.266269 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-42t4c" Mar 07 14:54:30 crc kubenswrapper[4943]: I0307 14:54:30.752651 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-42t4c"] Mar 07 14:54:31 crc kubenswrapper[4943]: I0307 14:54:31.722640 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-42t4c" event={"ID":"0b3e1865-67b8-4e70-ae06-e0c41256e303","Type":"ContainerStarted","Data":"8319f5c3116610df202d6335eee60b807aa82f64336aaf6e16c5db31f416af90"} Mar 07 14:54:31 crc kubenswrapper[4943]: I0307 14:54:31.827293 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lfm7s" Mar 07 14:54:31 crc kubenswrapper[4943]: I0307 14:54:31.827362 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lfm7s" Mar 07 14:54:31 crc kubenswrapper[4943]: I0307 14:54:31.874459 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lfm7s" Mar 07 14:54:32 crc kubenswrapper[4943]: I0307 14:54:32.789220 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lfm7s" Mar 07 14:54:34 crc kubenswrapper[4943]: I0307 14:54:34.746696 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-42t4c" event={"ID":"0b3e1865-67b8-4e70-ae06-e0c41256e303","Type":"ContainerStarted","Data":"5ba77e1ed72c5b5bb727904f3a2bd28a91ea3171b98336d4e644086861772f5a"} Mar 07 14:54:34 crc kubenswrapper[4943]: I0307 14:54:34.767737 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-42t4c" podStartSLOduration=2.564724764 podStartE2EDuration="5.76771358s" podCreationTimestamp="2026-03-07 14:54:29 +0000 UTC" firstStartedPulling="2026-03-07 14:54:30.772003752 +0000 UTC m=+912.724140280" lastFinishedPulling="2026-03-07 14:54:33.974992598 +0000 UTC m=+915.927129096" observedRunningTime="2026-03-07 14:54:34.766588643 +0000 UTC m=+916.718725181" watchObservedRunningTime="2026-03-07 14:54:34.76771358 +0000 UTC m=+916.719850108" Mar 07 14:54:34 crc kubenswrapper[4943]: I0307 14:54:34.918672 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8nwmm"] Mar 07 14:54:34 crc kubenswrapper[4943]: I0307 14:54:34.922447 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8nwmm" Mar 07 14:54:34 crc kubenswrapper[4943]: I0307 14:54:34.935761 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8nwmm"] Mar 07 14:54:34 crc kubenswrapper[4943]: I0307 14:54:34.944966 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e18c052b-1baa-4e77-8554-48bc2f05679b-utilities\") pod \"community-operators-8nwmm\" (UID: \"e18c052b-1baa-4e77-8554-48bc2f05679b\") " pod="openshift-marketplace/community-operators-8nwmm" Mar 07 14:54:34 crc kubenswrapper[4943]: I0307 14:54:34.945159 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48hqx\" (UniqueName: \"kubernetes.io/projected/e18c052b-1baa-4e77-8554-48bc2f05679b-kube-api-access-48hqx\") pod \"community-operators-8nwmm\" (UID: \"e18c052b-1baa-4e77-8554-48bc2f05679b\") " pod="openshift-marketplace/community-operators-8nwmm" Mar 07 14:54:34 crc kubenswrapper[4943]: I0307 14:54:34.945336 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e18c052b-1baa-4e77-8554-48bc2f05679b-catalog-content\") pod \"community-operators-8nwmm\" (UID: \"e18c052b-1baa-4e77-8554-48bc2f05679b\") " pod="openshift-marketplace/community-operators-8nwmm" Mar 07 14:54:35 crc kubenswrapper[4943]: I0307 14:54:35.046873 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e18c052b-1baa-4e77-8554-48bc2f05679b-utilities\") pod \"community-operators-8nwmm\" (UID: \"e18c052b-1baa-4e77-8554-48bc2f05679b\") " pod="openshift-marketplace/community-operators-8nwmm" Mar 07 14:54:35 crc kubenswrapper[4943]: I0307 14:54:35.046944 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48hqx\" (UniqueName: \"kubernetes.io/projected/e18c052b-1baa-4e77-8554-48bc2f05679b-kube-api-access-48hqx\") pod \"community-operators-8nwmm\" (UID: \"e18c052b-1baa-4e77-8554-48bc2f05679b\") " pod="openshift-marketplace/community-operators-8nwmm" Mar 07 14:54:35 crc kubenswrapper[4943]: I0307 14:54:35.046991 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e18c052b-1baa-4e77-8554-48bc2f05679b-catalog-content\") pod \"community-operators-8nwmm\" (UID: \"e18c052b-1baa-4e77-8554-48bc2f05679b\") " pod="openshift-marketplace/community-operators-8nwmm" Mar 07 14:54:35 crc kubenswrapper[4943]: I0307 14:54:35.047412 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e18c052b-1baa-4e77-8554-48bc2f05679b-catalog-content\") pod \"community-operators-8nwmm\" (UID: \"e18c052b-1baa-4e77-8554-48bc2f05679b\") " pod="openshift-marketplace/community-operators-8nwmm" Mar 07 14:54:35 crc kubenswrapper[4943]: I0307 14:54:35.047621 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e18c052b-1baa-4e77-8554-48bc2f05679b-utilities\") pod \"community-operators-8nwmm\" (UID: \"e18c052b-1baa-4e77-8554-48bc2f05679b\") " pod="openshift-marketplace/community-operators-8nwmm" Mar 07 14:54:35 crc kubenswrapper[4943]: I0307 14:54:35.069863 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48hqx\" (UniqueName: \"kubernetes.io/projected/e18c052b-1baa-4e77-8554-48bc2f05679b-kube-api-access-48hqx\") pod \"community-operators-8nwmm\" (UID: \"e18c052b-1baa-4e77-8554-48bc2f05679b\") " pod="openshift-marketplace/community-operators-8nwmm" Mar 07 14:54:35 crc kubenswrapper[4943]: I0307 14:54:35.242313 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8nwmm" Mar 07 14:54:35 crc kubenswrapper[4943]: I0307 14:54:35.753310 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8nwmm"] Mar 07 14:54:36 crc kubenswrapper[4943]: I0307 14:54:36.477011 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lfm7s"] Mar 07 14:54:36 crc kubenswrapper[4943]: I0307 14:54:36.477683 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-lfm7s" podUID="79dd2902-63b6-4e27-b50e-bc5e5553af5e" containerName="registry-server" containerID="cri-o://5a6fdad7eb933814aa708c0cf0b0542d6b46d6a679a15e886339de6d1d223b39" gracePeriod=2 Mar 07 14:54:36 crc kubenswrapper[4943]: I0307 14:54:36.779942 4943 generic.go:334] "Generic (PLEG): container finished" podID="79dd2902-63b6-4e27-b50e-bc5e5553af5e" containerID="5a6fdad7eb933814aa708c0cf0b0542d6b46d6a679a15e886339de6d1d223b39" exitCode=0 Mar 07 14:54:36 crc kubenswrapper[4943]: I0307 14:54:36.780370 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lfm7s" event={"ID":"79dd2902-63b6-4e27-b50e-bc5e5553af5e","Type":"ContainerDied","Data":"5a6fdad7eb933814aa708c0cf0b0542d6b46d6a679a15e886339de6d1d223b39"} Mar 07 14:54:36 crc kubenswrapper[4943]: I0307 14:54:36.782568 4943 generic.go:334] "Generic (PLEG): container finished" podID="e18c052b-1baa-4e77-8554-48bc2f05679b" containerID="8b3a7cf00e74c1c923ca3f8a5143d2848f929e6afac683e626afc0df4428053a" exitCode=0 Mar 07 14:54:36 crc kubenswrapper[4943]: I0307 14:54:36.782622 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8nwmm" event={"ID":"e18c052b-1baa-4e77-8554-48bc2f05679b","Type":"ContainerDied","Data":"8b3a7cf00e74c1c923ca3f8a5143d2848f929e6afac683e626afc0df4428053a"} Mar 07 14:54:36 crc kubenswrapper[4943]: I0307 14:54:36.782661 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8nwmm" event={"ID":"e18c052b-1baa-4e77-8554-48bc2f05679b","Type":"ContainerStarted","Data":"31b06e05201d28281ac46e6af54ba7a4db2d46a920944b3d8b86060b656f44ea"} Mar 07 14:54:36 crc kubenswrapper[4943]: I0307 14:54:36.916990 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lfm7s" Mar 07 14:54:37 crc kubenswrapper[4943]: I0307 14:54:37.081160 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wppnr\" (UniqueName: \"kubernetes.io/projected/79dd2902-63b6-4e27-b50e-bc5e5553af5e-kube-api-access-wppnr\") pod \"79dd2902-63b6-4e27-b50e-bc5e5553af5e\" (UID: \"79dd2902-63b6-4e27-b50e-bc5e5553af5e\") " Mar 07 14:54:37 crc kubenswrapper[4943]: I0307 14:54:37.081239 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79dd2902-63b6-4e27-b50e-bc5e5553af5e-utilities\") pod \"79dd2902-63b6-4e27-b50e-bc5e5553af5e\" (UID: \"79dd2902-63b6-4e27-b50e-bc5e5553af5e\") " Mar 07 14:54:37 crc kubenswrapper[4943]: I0307 14:54:37.081315 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79dd2902-63b6-4e27-b50e-bc5e5553af5e-catalog-content\") pod \"79dd2902-63b6-4e27-b50e-bc5e5553af5e\" (UID: \"79dd2902-63b6-4e27-b50e-bc5e5553af5e\") " Mar 07 14:54:37 crc kubenswrapper[4943]: I0307 14:54:37.083188 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79dd2902-63b6-4e27-b50e-bc5e5553af5e-utilities" (OuterVolumeSpecName: "utilities") pod "79dd2902-63b6-4e27-b50e-bc5e5553af5e" (UID: "79dd2902-63b6-4e27-b50e-bc5e5553af5e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:54:37 crc kubenswrapper[4943]: I0307 14:54:37.094414 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79dd2902-63b6-4e27-b50e-bc5e5553af5e-kube-api-access-wppnr" (OuterVolumeSpecName: "kube-api-access-wppnr") pod "79dd2902-63b6-4e27-b50e-bc5e5553af5e" (UID: "79dd2902-63b6-4e27-b50e-bc5e5553af5e"). InnerVolumeSpecName "kube-api-access-wppnr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:54:37 crc kubenswrapper[4943]: I0307 14:54:37.136507 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79dd2902-63b6-4e27-b50e-bc5e5553af5e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "79dd2902-63b6-4e27-b50e-bc5e5553af5e" (UID: "79dd2902-63b6-4e27-b50e-bc5e5553af5e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:54:37 crc kubenswrapper[4943]: I0307 14:54:37.182583 4943 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79dd2902-63b6-4e27-b50e-bc5e5553af5e-utilities\") on node \"crc\" DevicePath \"\"" Mar 07 14:54:37 crc kubenswrapper[4943]: I0307 14:54:37.182624 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wppnr\" (UniqueName: \"kubernetes.io/projected/79dd2902-63b6-4e27-b50e-bc5e5553af5e-kube-api-access-wppnr\") on node \"crc\" DevicePath \"\"" Mar 07 14:54:37 crc kubenswrapper[4943]: I0307 14:54:37.182638 4943 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79dd2902-63b6-4e27-b50e-bc5e5553af5e-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 07 14:54:37 crc kubenswrapper[4943]: I0307 14:54:37.798725 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8nwmm" event={"ID":"e18c052b-1baa-4e77-8554-48bc2f05679b","Type":"ContainerStarted","Data":"e0af7b9c1b8fc768adda7d70a85ab0e011adca80fed19d67c27a14ca1d91a6d2"} Mar 07 14:54:37 crc kubenswrapper[4943]: I0307 14:54:37.801245 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lfm7s" event={"ID":"79dd2902-63b6-4e27-b50e-bc5e5553af5e","Type":"ContainerDied","Data":"0eb7de7c6bed0c49027fdacbef373caf99828d9f8a366c04e6b703bb773e2c76"} Mar 07 14:54:37 crc kubenswrapper[4943]: I0307 14:54:37.801317 4943 scope.go:117] "RemoveContainer" containerID="5a6fdad7eb933814aa708c0cf0b0542d6b46d6a679a15e886339de6d1d223b39" Mar 07 14:54:37 crc kubenswrapper[4943]: I0307 14:54:37.801393 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lfm7s" Mar 07 14:54:37 crc kubenswrapper[4943]: I0307 14:54:37.826206 4943 scope.go:117] "RemoveContainer" containerID="bde325f01af3d4e520034311c5b4c3b00dd00dcdd0e47b104d4eda480787b38e" Mar 07 14:54:37 crc kubenswrapper[4943]: I0307 14:54:37.838147 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lfm7s"] Mar 07 14:54:37 crc kubenswrapper[4943]: I0307 14:54:37.844309 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-lfm7s"] Mar 07 14:54:37 crc kubenswrapper[4943]: I0307 14:54:37.858180 4943 scope.go:117] "RemoveContainer" containerID="c891f1cd14b68e7e0fdd5695de2818629a7d39504759045f1f486ff491d7338d" Mar 07 14:54:38 crc kubenswrapper[4943]: I0307 14:54:38.510299 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/rabbitmq-server-0"] Mar 07 14:54:38 crc kubenswrapper[4943]: E0307 14:54:38.511044 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79dd2902-63b6-4e27-b50e-bc5e5553af5e" containerName="extract-utilities" Mar 07 14:54:38 crc kubenswrapper[4943]: I0307 14:54:38.511072 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="79dd2902-63b6-4e27-b50e-bc5e5553af5e" containerName="extract-utilities" Mar 07 14:54:38 crc kubenswrapper[4943]: E0307 14:54:38.511099 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79dd2902-63b6-4e27-b50e-bc5e5553af5e" containerName="registry-server" Mar 07 14:54:38 crc kubenswrapper[4943]: I0307 14:54:38.511112 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="79dd2902-63b6-4e27-b50e-bc5e5553af5e" containerName="registry-server" Mar 07 14:54:38 crc kubenswrapper[4943]: E0307 14:54:38.511134 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79dd2902-63b6-4e27-b50e-bc5e5553af5e" containerName="extract-content" Mar 07 14:54:38 crc kubenswrapper[4943]: I0307 14:54:38.511148 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="79dd2902-63b6-4e27-b50e-bc5e5553af5e" containerName="extract-content" Mar 07 14:54:38 crc kubenswrapper[4943]: I0307 14:54:38.511367 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="79dd2902-63b6-4e27-b50e-bc5e5553af5e" containerName="registry-server" Mar 07 14:54:38 crc kubenswrapper[4943]: I0307 14:54:38.512474 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/rabbitmq-server-0" Mar 07 14:54:38 crc kubenswrapper[4943]: I0307 14:54:38.514899 4943 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"rabbitmq-default-user" Mar 07 14:54:38 crc kubenswrapper[4943]: I0307 14:54:38.515432 4943 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"rabbitmq-server-dockercfg-2fkxw" Mar 07 14:54:38 crc kubenswrapper[4943]: I0307 14:54:38.515746 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"rabbitmq-plugins-conf" Mar 07 14:54:38 crc kubenswrapper[4943]: I0307 14:54:38.516073 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"rabbitmq-server-conf" Mar 07 14:54:38 crc kubenswrapper[4943]: I0307 14:54:38.516170 4943 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"rabbitmq-erlang-cookie" Mar 07 14:54:38 crc kubenswrapper[4943]: I0307 14:54:38.538979 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/rabbitmq-server-0"] Mar 07 14:54:38 crc kubenswrapper[4943]: I0307 14:54:38.715558 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-61626df6-5c9d-4106-bda9-af5b194a9d5b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-61626df6-5c9d-4106-bda9-af5b194a9d5b\") pod \"rabbitmq-server-0\" (UID: \"f73c32b4-3904-4bf3-9bb4-13750cb45e88\") " pod="swift-kuttl-tests/rabbitmq-server-0" Mar 07 14:54:38 crc kubenswrapper[4943]: I0307 14:54:38.715653 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f73c32b4-3904-4bf3-9bb4-13750cb45e88-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f73c32b4-3904-4bf3-9bb4-13750cb45e88\") " pod="swift-kuttl-tests/rabbitmq-server-0" Mar 07 14:54:38 crc kubenswrapper[4943]: I0307 14:54:38.715707 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f73c32b4-3904-4bf3-9bb4-13750cb45e88-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f73c32b4-3904-4bf3-9bb4-13750cb45e88\") " pod="swift-kuttl-tests/rabbitmq-server-0" Mar 07 14:54:38 crc kubenswrapper[4943]: I0307 14:54:38.715758 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbj26\" (UniqueName: \"kubernetes.io/projected/f73c32b4-3904-4bf3-9bb4-13750cb45e88-kube-api-access-pbj26\") pod \"rabbitmq-server-0\" (UID: \"f73c32b4-3904-4bf3-9bb4-13750cb45e88\") " pod="swift-kuttl-tests/rabbitmq-server-0" Mar 07 14:54:38 crc kubenswrapper[4943]: I0307 14:54:38.715811 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f73c32b4-3904-4bf3-9bb4-13750cb45e88-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f73c32b4-3904-4bf3-9bb4-13750cb45e88\") " pod="swift-kuttl-tests/rabbitmq-server-0" Mar 07 14:54:38 crc kubenswrapper[4943]: I0307 14:54:38.716079 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f73c32b4-3904-4bf3-9bb4-13750cb45e88-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f73c32b4-3904-4bf3-9bb4-13750cb45e88\") " pod="swift-kuttl-tests/rabbitmq-server-0" Mar 07 14:54:38 crc kubenswrapper[4943]: I0307 14:54:38.716207 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f73c32b4-3904-4bf3-9bb4-13750cb45e88-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f73c32b4-3904-4bf3-9bb4-13750cb45e88\") " pod="swift-kuttl-tests/rabbitmq-server-0" Mar 07 14:54:38 crc kubenswrapper[4943]: I0307 14:54:38.716268 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f73c32b4-3904-4bf3-9bb4-13750cb45e88-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f73c32b4-3904-4bf3-9bb4-13750cb45e88\") " pod="swift-kuttl-tests/rabbitmq-server-0" Mar 07 14:54:38 crc kubenswrapper[4943]: I0307 14:54:38.777904 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79dd2902-63b6-4e27-b50e-bc5e5553af5e" path="/var/lib/kubelet/pods/79dd2902-63b6-4e27-b50e-bc5e5553af5e/volumes" Mar 07 14:54:38 crc kubenswrapper[4943]: I0307 14:54:38.811572 4943 generic.go:334] "Generic (PLEG): container finished" podID="e18c052b-1baa-4e77-8554-48bc2f05679b" containerID="e0af7b9c1b8fc768adda7d70a85ab0e011adca80fed19d67c27a14ca1d91a6d2" exitCode=0 Mar 07 14:54:38 crc kubenswrapper[4943]: I0307 14:54:38.811638 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8nwmm" event={"ID":"e18c052b-1baa-4e77-8554-48bc2f05679b","Type":"ContainerDied","Data":"e0af7b9c1b8fc768adda7d70a85ab0e011adca80fed19d67c27a14ca1d91a6d2"} Mar 07 14:54:38 crc kubenswrapper[4943]: I0307 14:54:38.817215 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f73c32b4-3904-4bf3-9bb4-13750cb45e88-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f73c32b4-3904-4bf3-9bb4-13750cb45e88\") " pod="swift-kuttl-tests/rabbitmq-server-0" Mar 07 14:54:38 crc kubenswrapper[4943]: I0307 14:54:38.817341 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f73c32b4-3904-4bf3-9bb4-13750cb45e88-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f73c32b4-3904-4bf3-9bb4-13750cb45e88\") " pod="swift-kuttl-tests/rabbitmq-server-0" Mar 07 14:54:38 crc kubenswrapper[4943]: I0307 14:54:38.817412 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f73c32b4-3904-4bf3-9bb4-13750cb45e88-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f73c32b4-3904-4bf3-9bb4-13750cb45e88\") " pod="swift-kuttl-tests/rabbitmq-server-0" Mar 07 14:54:38 crc kubenswrapper[4943]: I0307 14:54:38.817478 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-61626df6-5c9d-4106-bda9-af5b194a9d5b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-61626df6-5c9d-4106-bda9-af5b194a9d5b\") pod \"rabbitmq-server-0\" (UID: \"f73c32b4-3904-4bf3-9bb4-13750cb45e88\") " pod="swift-kuttl-tests/rabbitmq-server-0" Mar 07 14:54:38 crc kubenswrapper[4943]: I0307 14:54:38.817522 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f73c32b4-3904-4bf3-9bb4-13750cb45e88-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f73c32b4-3904-4bf3-9bb4-13750cb45e88\") " pod="swift-kuttl-tests/rabbitmq-server-0" Mar 07 14:54:38 crc kubenswrapper[4943]: I0307 14:54:38.817561 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f73c32b4-3904-4bf3-9bb4-13750cb45e88-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f73c32b4-3904-4bf3-9bb4-13750cb45e88\") " pod="swift-kuttl-tests/rabbitmq-server-0" Mar 07 14:54:38 crc kubenswrapper[4943]: I0307 14:54:38.817605 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbj26\" (UniqueName: \"kubernetes.io/projected/f73c32b4-3904-4bf3-9bb4-13750cb45e88-kube-api-access-pbj26\") pod \"rabbitmq-server-0\" (UID: \"f73c32b4-3904-4bf3-9bb4-13750cb45e88\") " pod="swift-kuttl-tests/rabbitmq-server-0" Mar 07 14:54:38 crc kubenswrapper[4943]: I0307 14:54:38.817646 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f73c32b4-3904-4bf3-9bb4-13750cb45e88-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f73c32b4-3904-4bf3-9bb4-13750cb45e88\") " pod="swift-kuttl-tests/rabbitmq-server-0" Mar 07 14:54:38 crc kubenswrapper[4943]: I0307 14:54:38.818490 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f73c32b4-3904-4bf3-9bb4-13750cb45e88-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f73c32b4-3904-4bf3-9bb4-13750cb45e88\") " pod="swift-kuttl-tests/rabbitmq-server-0" Mar 07 14:54:38 crc kubenswrapper[4943]: I0307 14:54:38.819094 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f73c32b4-3904-4bf3-9bb4-13750cb45e88-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f73c32b4-3904-4bf3-9bb4-13750cb45e88\") " pod="swift-kuttl-tests/rabbitmq-server-0" Mar 07 14:54:38 crc kubenswrapper[4943]: I0307 14:54:38.820418 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f73c32b4-3904-4bf3-9bb4-13750cb45e88-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f73c32b4-3904-4bf3-9bb4-13750cb45e88\") " pod="swift-kuttl-tests/rabbitmq-server-0" Mar 07 14:54:38 crc kubenswrapper[4943]: I0307 14:54:38.827000 4943 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 07 14:54:38 crc kubenswrapper[4943]: I0307 14:54:38.827062 4943 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-61626df6-5c9d-4106-bda9-af5b194a9d5b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-61626df6-5c9d-4106-bda9-af5b194a9d5b\") pod \"rabbitmq-server-0\" (UID: \"f73c32b4-3904-4bf3-9bb4-13750cb45e88\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/31725801f97b9cab4e71acb3d856d66ffbcaba2b7a73a122f2d4d816697071e3/globalmount\"" pod="swift-kuttl-tests/rabbitmq-server-0" Mar 07 14:54:38 crc kubenswrapper[4943]: I0307 14:54:38.829316 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f73c32b4-3904-4bf3-9bb4-13750cb45e88-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f73c32b4-3904-4bf3-9bb4-13750cb45e88\") " pod="swift-kuttl-tests/rabbitmq-server-0" Mar 07 14:54:38 crc kubenswrapper[4943]: I0307 14:54:38.830080 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f73c32b4-3904-4bf3-9bb4-13750cb45e88-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f73c32b4-3904-4bf3-9bb4-13750cb45e88\") " pod="swift-kuttl-tests/rabbitmq-server-0" Mar 07 14:54:38 crc kubenswrapper[4943]: I0307 14:54:38.831324 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f73c32b4-3904-4bf3-9bb4-13750cb45e88-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f73c32b4-3904-4bf3-9bb4-13750cb45e88\") " pod="swift-kuttl-tests/rabbitmq-server-0" Mar 07 14:54:38 crc kubenswrapper[4943]: I0307 14:54:38.849491 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbj26\" (UniqueName: \"kubernetes.io/projected/f73c32b4-3904-4bf3-9bb4-13750cb45e88-kube-api-access-pbj26\") pod \"rabbitmq-server-0\" (UID: \"f73c32b4-3904-4bf3-9bb4-13750cb45e88\") " pod="swift-kuttl-tests/rabbitmq-server-0" Mar 07 14:54:38 crc kubenswrapper[4943]: I0307 14:54:38.903035 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-61626df6-5c9d-4106-bda9-af5b194a9d5b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-61626df6-5c9d-4106-bda9-af5b194a9d5b\") pod \"rabbitmq-server-0\" (UID: \"f73c32b4-3904-4bf3-9bb4-13750cb45e88\") " pod="swift-kuttl-tests/rabbitmq-server-0" Mar 07 14:54:39 crc kubenswrapper[4943]: I0307 14:54:39.138508 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/rabbitmq-server-0" Mar 07 14:54:39 crc kubenswrapper[4943]: I0307 14:54:39.624726 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/rabbitmq-server-0"] Mar 07 14:54:39 crc kubenswrapper[4943]: I0307 14:54:39.824059 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/rabbitmq-server-0" event={"ID":"f73c32b4-3904-4bf3-9bb4-13750cb45e88","Type":"ContainerStarted","Data":"7a946ddc281fd756f70ecda43fa116579d2b913686ab5850c6854551dce31752"} Mar 07 14:54:39 crc kubenswrapper[4943]: I0307 14:54:39.828231 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8nwmm" event={"ID":"e18c052b-1baa-4e77-8554-48bc2f05679b","Type":"ContainerStarted","Data":"54d7879c92deadbd4a46ddb9d96375df62a16c3d8292afd296acff3f8b59b0bd"} Mar 07 14:54:39 crc kubenswrapper[4943]: I0307 14:54:39.860678 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8nwmm" podStartSLOduration=3.430487911 podStartE2EDuration="5.860654344s" podCreationTimestamp="2026-03-07 14:54:34 +0000 UTC" firstStartedPulling="2026-03-07 14:54:36.785302635 +0000 UTC m=+918.737439143" lastFinishedPulling="2026-03-07 14:54:39.215469068 +0000 UTC m=+921.167605576" observedRunningTime="2026-03-07 14:54:39.854948744 +0000 UTC m=+921.807085252" watchObservedRunningTime="2026-03-07 14:54:39.860654344 +0000 UTC m=+921.812790852" Mar 07 14:54:43 crc kubenswrapper[4943]: I0307 14:54:43.476554 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-index-cnq8m"] Mar 07 14:54:43 crc kubenswrapper[4943]: I0307 14:54:43.480668 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-cnq8m" Mar 07 14:54:43 crc kubenswrapper[4943]: I0307 14:54:43.485277 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-index-dockercfg-5q6xb" Mar 07 14:54:43 crc kubenswrapper[4943]: I0307 14:54:43.487249 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-cnq8m"] Mar 07 14:54:43 crc kubenswrapper[4943]: I0307 14:54:43.585230 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5sqs4\" (UniqueName: \"kubernetes.io/projected/690e82f5-4a59-4926-95b5-b13362cb1a97-kube-api-access-5sqs4\") pod \"keystone-operator-index-cnq8m\" (UID: \"690e82f5-4a59-4926-95b5-b13362cb1a97\") " pod="openstack-operators/keystone-operator-index-cnq8m" Mar 07 14:54:43 crc kubenswrapper[4943]: I0307 14:54:43.687207 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5sqs4\" (UniqueName: \"kubernetes.io/projected/690e82f5-4a59-4926-95b5-b13362cb1a97-kube-api-access-5sqs4\") pod \"keystone-operator-index-cnq8m\" (UID: \"690e82f5-4a59-4926-95b5-b13362cb1a97\") " pod="openstack-operators/keystone-operator-index-cnq8m" Mar 07 14:54:43 crc kubenswrapper[4943]: I0307 14:54:43.722431 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5sqs4\" (UniqueName: \"kubernetes.io/projected/690e82f5-4a59-4926-95b5-b13362cb1a97-kube-api-access-5sqs4\") pod \"keystone-operator-index-cnq8m\" (UID: \"690e82f5-4a59-4926-95b5-b13362cb1a97\") " pod="openstack-operators/keystone-operator-index-cnq8m" Mar 07 14:54:43 crc kubenswrapper[4943]: I0307 14:54:43.808160 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-cnq8m" Mar 07 14:54:45 crc kubenswrapper[4943]: I0307 14:54:45.242557 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8nwmm" Mar 07 14:54:45 crc kubenswrapper[4943]: I0307 14:54:45.243030 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8nwmm" Mar 07 14:54:45 crc kubenswrapper[4943]: I0307 14:54:45.295774 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8nwmm" Mar 07 14:54:45 crc kubenswrapper[4943]: I0307 14:54:45.948635 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8nwmm" Mar 07 14:54:46 crc kubenswrapper[4943]: I0307 14:54:46.382064 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-cnq8m"] Mar 07 14:54:46 crc kubenswrapper[4943]: I0307 14:54:46.886624 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-cnq8m" event={"ID":"690e82f5-4a59-4926-95b5-b13362cb1a97","Type":"ContainerStarted","Data":"e280153bbe62840bd9e8dfa597199c307452f55ce623823e763b55872d83b263"} Mar 07 14:54:47 crc kubenswrapper[4943]: I0307 14:54:47.895220 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-cnq8m" event={"ID":"690e82f5-4a59-4926-95b5-b13362cb1a97","Type":"ContainerStarted","Data":"429d19d2f5dabb51961d83e1f71e20b2dff6522096c6daa706f668323d0960a6"} Mar 07 14:54:47 crc kubenswrapper[4943]: I0307 14:54:47.913563 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-index-cnq8m" podStartSLOduration=4.011115146 podStartE2EDuration="4.913538214s" podCreationTimestamp="2026-03-07 14:54:43 +0000 UTC" firstStartedPulling="2026-03-07 14:54:46.680458689 +0000 UTC m=+928.632595187" lastFinishedPulling="2026-03-07 14:54:47.582881717 +0000 UTC m=+929.535018255" observedRunningTime="2026-03-07 14:54:47.909380101 +0000 UTC m=+929.861516609" watchObservedRunningTime="2026-03-07 14:54:47.913538214 +0000 UTC m=+929.865674712" Mar 07 14:54:48 crc kubenswrapper[4943]: I0307 14:54:48.920636 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/rabbitmq-server-0" event={"ID":"f73c32b4-3904-4bf3-9bb4-13750cb45e88","Type":"ContainerStarted","Data":"789f2f6e1ba35f3c76fb7b16e3606a4d28e4b8aa980148e08397fb5067ab1532"} Mar 07 14:54:51 crc kubenswrapper[4943]: I0307 14:54:51.486714 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mxdxf"] Mar 07 14:54:51 crc kubenswrapper[4943]: I0307 14:54:51.489399 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mxdxf" Mar 07 14:54:51 crc kubenswrapper[4943]: I0307 14:54:51.499558 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mxdxf"] Mar 07 14:54:51 crc kubenswrapper[4943]: I0307 14:54:51.603524 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eed0d6a6-a45f-4e6a-85f1-9c7157055498-catalog-content\") pod \"certified-operators-mxdxf\" (UID: \"eed0d6a6-a45f-4e6a-85f1-9c7157055498\") " pod="openshift-marketplace/certified-operators-mxdxf" Mar 07 14:54:51 crc kubenswrapper[4943]: I0307 14:54:51.603725 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eed0d6a6-a45f-4e6a-85f1-9c7157055498-utilities\") pod \"certified-operators-mxdxf\" (UID: \"eed0d6a6-a45f-4e6a-85f1-9c7157055498\") " pod="openshift-marketplace/certified-operators-mxdxf" Mar 07 14:54:51 crc kubenswrapper[4943]: I0307 14:54:51.603993 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-th2gh\" (UniqueName: \"kubernetes.io/projected/eed0d6a6-a45f-4e6a-85f1-9c7157055498-kube-api-access-th2gh\") pod \"certified-operators-mxdxf\" (UID: \"eed0d6a6-a45f-4e6a-85f1-9c7157055498\") " pod="openshift-marketplace/certified-operators-mxdxf" Mar 07 14:54:51 crc kubenswrapper[4943]: I0307 14:54:51.705623 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eed0d6a6-a45f-4e6a-85f1-9c7157055498-catalog-content\") pod \"certified-operators-mxdxf\" (UID: \"eed0d6a6-a45f-4e6a-85f1-9c7157055498\") " pod="openshift-marketplace/certified-operators-mxdxf" Mar 07 14:54:51 crc kubenswrapper[4943]: I0307 14:54:51.705696 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eed0d6a6-a45f-4e6a-85f1-9c7157055498-utilities\") pod \"certified-operators-mxdxf\" (UID: \"eed0d6a6-a45f-4e6a-85f1-9c7157055498\") " pod="openshift-marketplace/certified-operators-mxdxf" Mar 07 14:54:51 crc kubenswrapper[4943]: I0307 14:54:51.705756 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-th2gh\" (UniqueName: \"kubernetes.io/projected/eed0d6a6-a45f-4e6a-85f1-9c7157055498-kube-api-access-th2gh\") pod \"certified-operators-mxdxf\" (UID: \"eed0d6a6-a45f-4e6a-85f1-9c7157055498\") " pod="openshift-marketplace/certified-operators-mxdxf" Mar 07 14:54:51 crc kubenswrapper[4943]: I0307 14:54:51.706235 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eed0d6a6-a45f-4e6a-85f1-9c7157055498-catalog-content\") pod \"certified-operators-mxdxf\" (UID: \"eed0d6a6-a45f-4e6a-85f1-9c7157055498\") " pod="openshift-marketplace/certified-operators-mxdxf" Mar 07 14:54:51 crc kubenswrapper[4943]: I0307 14:54:51.706305 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eed0d6a6-a45f-4e6a-85f1-9c7157055498-utilities\") pod \"certified-operators-mxdxf\" (UID: \"eed0d6a6-a45f-4e6a-85f1-9c7157055498\") " pod="openshift-marketplace/certified-operators-mxdxf" Mar 07 14:54:51 crc kubenswrapper[4943]: I0307 14:54:51.741794 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-th2gh\" (UniqueName: \"kubernetes.io/projected/eed0d6a6-a45f-4e6a-85f1-9c7157055498-kube-api-access-th2gh\") pod \"certified-operators-mxdxf\" (UID: \"eed0d6a6-a45f-4e6a-85f1-9c7157055498\") " pod="openshift-marketplace/certified-operators-mxdxf" Mar 07 14:54:51 crc kubenswrapper[4943]: I0307 14:54:51.861823 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mxdxf" Mar 07 14:54:52 crc kubenswrapper[4943]: I0307 14:54:52.320699 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mxdxf"] Mar 07 14:54:52 crc kubenswrapper[4943]: W0307 14:54:52.329670 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeed0d6a6_a45f_4e6a_85f1_9c7157055498.slice/crio-0335d5b206c8841c4d083ff77acddc44ed59dc43ef15c601af9fba723489b384 WatchSource:0}: Error finding container 0335d5b206c8841c4d083ff77acddc44ed59dc43ef15c601af9fba723489b384: Status 404 returned error can't find the container with id 0335d5b206c8841c4d083ff77acddc44ed59dc43ef15c601af9fba723489b384 Mar 07 14:54:52 crc kubenswrapper[4943]: I0307 14:54:52.666158 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8nwmm"] Mar 07 14:54:52 crc kubenswrapper[4943]: I0307 14:54:52.666796 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8nwmm" podUID="e18c052b-1baa-4e77-8554-48bc2f05679b" containerName="registry-server" containerID="cri-o://54d7879c92deadbd4a46ddb9d96375df62a16c3d8292afd296acff3f8b59b0bd" gracePeriod=2 Mar 07 14:54:52 crc kubenswrapper[4943]: I0307 14:54:52.962481 4943 generic.go:334] "Generic (PLEG): container finished" podID="eed0d6a6-a45f-4e6a-85f1-9c7157055498" containerID="01145689e18e84dab85ace54fc50e03257460ad9e7ca59d531c189d8b5260dd7" exitCode=0 Mar 07 14:54:52 crc kubenswrapper[4943]: I0307 14:54:52.962592 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mxdxf" event={"ID":"eed0d6a6-a45f-4e6a-85f1-9c7157055498","Type":"ContainerDied","Data":"01145689e18e84dab85ace54fc50e03257460ad9e7ca59d531c189d8b5260dd7"} Mar 07 14:54:52 crc kubenswrapper[4943]: I0307 14:54:52.962979 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mxdxf" event={"ID":"eed0d6a6-a45f-4e6a-85f1-9c7157055498","Type":"ContainerStarted","Data":"0335d5b206c8841c4d083ff77acddc44ed59dc43ef15c601af9fba723489b384"} Mar 07 14:54:52 crc kubenswrapper[4943]: I0307 14:54:52.971238 4943 generic.go:334] "Generic (PLEG): container finished" podID="e18c052b-1baa-4e77-8554-48bc2f05679b" containerID="54d7879c92deadbd4a46ddb9d96375df62a16c3d8292afd296acff3f8b59b0bd" exitCode=0 Mar 07 14:54:52 crc kubenswrapper[4943]: I0307 14:54:52.971291 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8nwmm" event={"ID":"e18c052b-1baa-4e77-8554-48bc2f05679b","Type":"ContainerDied","Data":"54d7879c92deadbd4a46ddb9d96375df62a16c3d8292afd296acff3f8b59b0bd"} Mar 07 14:54:53 crc kubenswrapper[4943]: I0307 14:54:53.088688 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8nwmm" Mar 07 14:54:53 crc kubenswrapper[4943]: I0307 14:54:53.125180 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e18c052b-1baa-4e77-8554-48bc2f05679b-utilities\") pod \"e18c052b-1baa-4e77-8554-48bc2f05679b\" (UID: \"e18c052b-1baa-4e77-8554-48bc2f05679b\") " Mar 07 14:54:53 crc kubenswrapper[4943]: I0307 14:54:53.125264 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-48hqx\" (UniqueName: \"kubernetes.io/projected/e18c052b-1baa-4e77-8554-48bc2f05679b-kube-api-access-48hqx\") pod \"e18c052b-1baa-4e77-8554-48bc2f05679b\" (UID: \"e18c052b-1baa-4e77-8554-48bc2f05679b\") " Mar 07 14:54:53 crc kubenswrapper[4943]: I0307 14:54:53.125326 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e18c052b-1baa-4e77-8554-48bc2f05679b-catalog-content\") pod \"e18c052b-1baa-4e77-8554-48bc2f05679b\" (UID: \"e18c052b-1baa-4e77-8554-48bc2f05679b\") " Mar 07 14:54:53 crc kubenswrapper[4943]: I0307 14:54:53.129508 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e18c052b-1baa-4e77-8554-48bc2f05679b-utilities" (OuterVolumeSpecName: "utilities") pod "e18c052b-1baa-4e77-8554-48bc2f05679b" (UID: "e18c052b-1baa-4e77-8554-48bc2f05679b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:54:53 crc kubenswrapper[4943]: I0307 14:54:53.131651 4943 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e18c052b-1baa-4e77-8554-48bc2f05679b-utilities\") on node \"crc\" DevicePath \"\"" Mar 07 14:54:53 crc kubenswrapper[4943]: I0307 14:54:53.146270 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e18c052b-1baa-4e77-8554-48bc2f05679b-kube-api-access-48hqx" (OuterVolumeSpecName: "kube-api-access-48hqx") pod "e18c052b-1baa-4e77-8554-48bc2f05679b" (UID: "e18c052b-1baa-4e77-8554-48bc2f05679b"). InnerVolumeSpecName "kube-api-access-48hqx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:54:53 crc kubenswrapper[4943]: I0307 14:54:53.192678 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e18c052b-1baa-4e77-8554-48bc2f05679b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e18c052b-1baa-4e77-8554-48bc2f05679b" (UID: "e18c052b-1baa-4e77-8554-48bc2f05679b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:54:53 crc kubenswrapper[4943]: I0307 14:54:53.232533 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-48hqx\" (UniqueName: \"kubernetes.io/projected/e18c052b-1baa-4e77-8554-48bc2f05679b-kube-api-access-48hqx\") on node \"crc\" DevicePath \"\"" Mar 07 14:54:53 crc kubenswrapper[4943]: I0307 14:54:53.232581 4943 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e18c052b-1baa-4e77-8554-48bc2f05679b-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 07 14:54:53 crc kubenswrapper[4943]: I0307 14:54:53.808617 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/keystone-operator-index-cnq8m" Mar 07 14:54:53 crc kubenswrapper[4943]: I0307 14:54:53.809189 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-index-cnq8m" Mar 07 14:54:53 crc kubenswrapper[4943]: I0307 14:54:53.852538 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/keystone-operator-index-cnq8m" Mar 07 14:54:53 crc kubenswrapper[4943]: I0307 14:54:53.978837 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mxdxf" event={"ID":"eed0d6a6-a45f-4e6a-85f1-9c7157055498","Type":"ContainerStarted","Data":"07b61ad3582c5dfa1fd3d97b58b3b7fadba821e754a0e083515e5e4aed44c1b3"} Mar 07 14:54:53 crc kubenswrapper[4943]: I0307 14:54:53.981764 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8nwmm" Mar 07 14:54:53 crc kubenswrapper[4943]: I0307 14:54:53.982350 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8nwmm" event={"ID":"e18c052b-1baa-4e77-8554-48bc2f05679b","Type":"ContainerDied","Data":"31b06e05201d28281ac46e6af54ba7a4db2d46a920944b3d8b86060b656f44ea"} Mar 07 14:54:53 crc kubenswrapper[4943]: I0307 14:54:53.982401 4943 scope.go:117] "RemoveContainer" containerID="54d7879c92deadbd4a46ddb9d96375df62a16c3d8292afd296acff3f8b59b0bd" Mar 07 14:54:54 crc kubenswrapper[4943]: I0307 14:54:54.016435 4943 scope.go:117] "RemoveContainer" containerID="e0af7b9c1b8fc768adda7d70a85ab0e011adca80fed19d67c27a14ca1d91a6d2" Mar 07 14:54:54 crc kubenswrapper[4943]: I0307 14:54:54.020119 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8nwmm"] Mar 07 14:54:54 crc kubenswrapper[4943]: I0307 14:54:54.023896 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8nwmm"] Mar 07 14:54:54 crc kubenswrapper[4943]: I0307 14:54:54.034618 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-index-cnq8m" Mar 07 14:54:54 crc kubenswrapper[4943]: I0307 14:54:54.043721 4943 scope.go:117] "RemoveContainer" containerID="8b3a7cf00e74c1c923ca3f8a5143d2848f929e6afac683e626afc0df4428053a" Mar 07 14:54:54 crc kubenswrapper[4943]: I0307 14:54:54.768859 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e18c052b-1baa-4e77-8554-48bc2f05679b" path="/var/lib/kubelet/pods/e18c052b-1baa-4e77-8554-48bc2f05679b/volumes" Mar 07 14:54:54 crc kubenswrapper[4943]: I0307 14:54:54.994749 4943 generic.go:334] "Generic (PLEG): container finished" podID="eed0d6a6-a45f-4e6a-85f1-9c7157055498" containerID="07b61ad3582c5dfa1fd3d97b58b3b7fadba821e754a0e083515e5e4aed44c1b3" exitCode=0 Mar 07 14:54:54 crc kubenswrapper[4943]: I0307 14:54:54.994881 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mxdxf" event={"ID":"eed0d6a6-a45f-4e6a-85f1-9c7157055498","Type":"ContainerDied","Data":"07b61ad3582c5dfa1fd3d97b58b3b7fadba821e754a0e083515e5e4aed44c1b3"} Mar 07 14:54:56 crc kubenswrapper[4943]: I0307 14:54:56.006169 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mxdxf" event={"ID":"eed0d6a6-a45f-4e6a-85f1-9c7157055498","Type":"ContainerStarted","Data":"25d9b42d87f972ad4f1b77373d5655da4ea30456d7ead6ee627e91705934cbf1"} Mar 07 14:54:56 crc kubenswrapper[4943]: I0307 14:54:56.035270 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mxdxf" podStartSLOduration=2.601838858 podStartE2EDuration="5.035246532s" podCreationTimestamp="2026-03-07 14:54:51 +0000 UTC" firstStartedPulling="2026-03-07 14:54:52.96508992 +0000 UTC m=+934.917226428" lastFinishedPulling="2026-03-07 14:54:55.398497574 +0000 UTC m=+937.350634102" observedRunningTime="2026-03-07 14:54:56.030457614 +0000 UTC m=+937.982594152" watchObservedRunningTime="2026-03-07 14:54:56.035246532 +0000 UTC m=+937.987383030" Mar 07 14:54:56 crc kubenswrapper[4943]: I0307 14:54:56.139050 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40frz9l"] Mar 07 14:54:56 crc kubenswrapper[4943]: E0307 14:54:56.139445 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e18c052b-1baa-4e77-8554-48bc2f05679b" containerName="extract-utilities" Mar 07 14:54:56 crc kubenswrapper[4943]: I0307 14:54:56.139475 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="e18c052b-1baa-4e77-8554-48bc2f05679b" containerName="extract-utilities" Mar 07 14:54:56 crc kubenswrapper[4943]: E0307 14:54:56.139500 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e18c052b-1baa-4e77-8554-48bc2f05679b" containerName="registry-server" Mar 07 14:54:56 crc kubenswrapper[4943]: I0307 14:54:56.139513 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="e18c052b-1baa-4e77-8554-48bc2f05679b" containerName="registry-server" Mar 07 14:54:56 crc kubenswrapper[4943]: E0307 14:54:56.139532 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e18c052b-1baa-4e77-8554-48bc2f05679b" containerName="extract-content" Mar 07 14:54:56 crc kubenswrapper[4943]: I0307 14:54:56.139575 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="e18c052b-1baa-4e77-8554-48bc2f05679b" containerName="extract-content" Mar 07 14:54:56 crc kubenswrapper[4943]: I0307 14:54:56.139826 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="e18c052b-1baa-4e77-8554-48bc2f05679b" containerName="registry-server" Mar 07 14:54:56 crc kubenswrapper[4943]: I0307 14:54:56.141432 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40frz9l" Mar 07 14:54:56 crc kubenswrapper[4943]: I0307 14:54:56.143531 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-k6gp2" Mar 07 14:54:56 crc kubenswrapper[4943]: I0307 14:54:56.191321 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fb540778-9fd9-48c8-b3f8-45327871b647-bundle\") pod \"a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40frz9l\" (UID: \"fb540778-9fd9-48c8-b3f8-45327871b647\") " pod="openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40frz9l" Mar 07 14:54:56 crc kubenswrapper[4943]: I0307 14:54:56.191429 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4gfk4\" (UniqueName: \"kubernetes.io/projected/fb540778-9fd9-48c8-b3f8-45327871b647-kube-api-access-4gfk4\") pod \"a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40frz9l\" (UID: \"fb540778-9fd9-48c8-b3f8-45327871b647\") " pod="openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40frz9l" Mar 07 14:54:56 crc kubenswrapper[4943]: I0307 14:54:56.191477 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fb540778-9fd9-48c8-b3f8-45327871b647-util\") pod \"a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40frz9l\" (UID: \"fb540778-9fd9-48c8-b3f8-45327871b647\") " pod="openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40frz9l" Mar 07 14:54:56 crc kubenswrapper[4943]: I0307 14:54:56.199200 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40frz9l"] Mar 07 14:54:56 crc kubenswrapper[4943]: I0307 14:54:56.292567 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fb540778-9fd9-48c8-b3f8-45327871b647-bundle\") pod \"a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40frz9l\" (UID: \"fb540778-9fd9-48c8-b3f8-45327871b647\") " pod="openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40frz9l" Mar 07 14:54:56 crc kubenswrapper[4943]: I0307 14:54:56.292644 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4gfk4\" (UniqueName: \"kubernetes.io/projected/fb540778-9fd9-48c8-b3f8-45327871b647-kube-api-access-4gfk4\") pod \"a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40frz9l\" (UID: \"fb540778-9fd9-48c8-b3f8-45327871b647\") " pod="openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40frz9l" Mar 07 14:54:56 crc kubenswrapper[4943]: I0307 14:54:56.292681 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fb540778-9fd9-48c8-b3f8-45327871b647-util\") pod \"a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40frz9l\" (UID: \"fb540778-9fd9-48c8-b3f8-45327871b647\") " pod="openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40frz9l" Mar 07 14:54:56 crc kubenswrapper[4943]: I0307 14:54:56.293396 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fb540778-9fd9-48c8-b3f8-45327871b647-util\") pod \"a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40frz9l\" (UID: \"fb540778-9fd9-48c8-b3f8-45327871b647\") " pod="openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40frz9l" Mar 07 14:54:56 crc kubenswrapper[4943]: I0307 14:54:56.293808 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fb540778-9fd9-48c8-b3f8-45327871b647-bundle\") pod \"a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40frz9l\" (UID: \"fb540778-9fd9-48c8-b3f8-45327871b647\") " pod="openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40frz9l" Mar 07 14:54:56 crc kubenswrapper[4943]: I0307 14:54:56.323700 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4gfk4\" (UniqueName: \"kubernetes.io/projected/fb540778-9fd9-48c8-b3f8-45327871b647-kube-api-access-4gfk4\") pod \"a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40frz9l\" (UID: \"fb540778-9fd9-48c8-b3f8-45327871b647\") " pod="openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40frz9l" Mar 07 14:54:56 crc kubenswrapper[4943]: I0307 14:54:56.496234 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40frz9l" Mar 07 14:54:56 crc kubenswrapper[4943]: I0307 14:54:56.973877 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40frz9l"] Mar 07 14:54:56 crc kubenswrapper[4943]: W0307 14:54:56.977487 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfb540778_9fd9_48c8_b3f8_45327871b647.slice/crio-c747fc18557bf7f4d1e1ad0dd1b4053367143cea9471a55efdcb4b1af182c948 WatchSource:0}: Error finding container c747fc18557bf7f4d1e1ad0dd1b4053367143cea9471a55efdcb4b1af182c948: Status 404 returned error can't find the container with id c747fc18557bf7f4d1e1ad0dd1b4053367143cea9471a55efdcb4b1af182c948 Mar 07 14:54:57 crc kubenswrapper[4943]: I0307 14:54:57.032112 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40frz9l" event={"ID":"fb540778-9fd9-48c8-b3f8-45327871b647","Type":"ContainerStarted","Data":"c747fc18557bf7f4d1e1ad0dd1b4053367143cea9471a55efdcb4b1af182c948"} Mar 07 14:54:58 crc kubenswrapper[4943]: I0307 14:54:58.040429 4943 generic.go:334] "Generic (PLEG): container finished" podID="fb540778-9fd9-48c8-b3f8-45327871b647" containerID="c968eb7c44577e079dfc6d159494a0989583e1227a82f1e8b840142f63a13d57" exitCode=0 Mar 07 14:54:58 crc kubenswrapper[4943]: I0307 14:54:58.040555 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40frz9l" event={"ID":"fb540778-9fd9-48c8-b3f8-45327871b647","Type":"ContainerDied","Data":"c968eb7c44577e079dfc6d159494a0989583e1227a82f1e8b840142f63a13d57"} Mar 07 14:54:59 crc kubenswrapper[4943]: I0307 14:54:59.049884 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40frz9l" event={"ID":"fb540778-9fd9-48c8-b3f8-45327871b647","Type":"ContainerStarted","Data":"88fc8d52192e72bff63c151aa734d89fede28b60a8649e9e6b82ef6e7e4c5f60"} Mar 07 14:55:00 crc kubenswrapper[4943]: I0307 14:55:00.064093 4943 generic.go:334] "Generic (PLEG): container finished" podID="fb540778-9fd9-48c8-b3f8-45327871b647" containerID="88fc8d52192e72bff63c151aa734d89fede28b60a8649e9e6b82ef6e7e4c5f60" exitCode=0 Mar 07 14:55:00 crc kubenswrapper[4943]: I0307 14:55:00.064163 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40frz9l" event={"ID":"fb540778-9fd9-48c8-b3f8-45327871b647","Type":"ContainerDied","Data":"88fc8d52192e72bff63c151aa734d89fede28b60a8649e9e6b82ef6e7e4c5f60"} Mar 07 14:55:01 crc kubenswrapper[4943]: I0307 14:55:01.075453 4943 generic.go:334] "Generic (PLEG): container finished" podID="fb540778-9fd9-48c8-b3f8-45327871b647" containerID="fb367cbeb1ae25325c78afd31956284e784a918547aadac07a187ecdece54a76" exitCode=0 Mar 07 14:55:01 crc kubenswrapper[4943]: I0307 14:55:01.075547 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40frz9l" event={"ID":"fb540778-9fd9-48c8-b3f8-45327871b647","Type":"ContainerDied","Data":"fb367cbeb1ae25325c78afd31956284e784a918547aadac07a187ecdece54a76"} Mar 07 14:55:01 crc kubenswrapper[4943]: I0307 14:55:01.862368 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mxdxf" Mar 07 14:55:01 crc kubenswrapper[4943]: I0307 14:55:01.862453 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mxdxf" Mar 07 14:55:01 crc kubenswrapper[4943]: I0307 14:55:01.926695 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mxdxf" Mar 07 14:55:02 crc kubenswrapper[4943]: I0307 14:55:02.157291 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mxdxf" Mar 07 14:55:02 crc kubenswrapper[4943]: I0307 14:55:02.488198 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40frz9l" Mar 07 14:55:02 crc kubenswrapper[4943]: I0307 14:55:02.601502 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fb540778-9fd9-48c8-b3f8-45327871b647-util\") pod \"fb540778-9fd9-48c8-b3f8-45327871b647\" (UID: \"fb540778-9fd9-48c8-b3f8-45327871b647\") " Mar 07 14:55:02 crc kubenswrapper[4943]: I0307 14:55:02.601725 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fb540778-9fd9-48c8-b3f8-45327871b647-bundle\") pod \"fb540778-9fd9-48c8-b3f8-45327871b647\" (UID: \"fb540778-9fd9-48c8-b3f8-45327871b647\") " Mar 07 14:55:02 crc kubenswrapper[4943]: I0307 14:55:02.601791 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4gfk4\" (UniqueName: \"kubernetes.io/projected/fb540778-9fd9-48c8-b3f8-45327871b647-kube-api-access-4gfk4\") pod \"fb540778-9fd9-48c8-b3f8-45327871b647\" (UID: \"fb540778-9fd9-48c8-b3f8-45327871b647\") " Mar 07 14:55:02 crc kubenswrapper[4943]: I0307 14:55:02.603892 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb540778-9fd9-48c8-b3f8-45327871b647-bundle" (OuterVolumeSpecName: "bundle") pod "fb540778-9fd9-48c8-b3f8-45327871b647" (UID: "fb540778-9fd9-48c8-b3f8-45327871b647"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:55:02 crc kubenswrapper[4943]: I0307 14:55:02.608482 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb540778-9fd9-48c8-b3f8-45327871b647-kube-api-access-4gfk4" (OuterVolumeSpecName: "kube-api-access-4gfk4") pod "fb540778-9fd9-48c8-b3f8-45327871b647" (UID: "fb540778-9fd9-48c8-b3f8-45327871b647"). InnerVolumeSpecName "kube-api-access-4gfk4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:55:02 crc kubenswrapper[4943]: I0307 14:55:02.703969 4943 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fb540778-9fd9-48c8-b3f8-45327871b647-bundle\") on node \"crc\" DevicePath \"\"" Mar 07 14:55:02 crc kubenswrapper[4943]: I0307 14:55:02.704026 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4gfk4\" (UniqueName: \"kubernetes.io/projected/fb540778-9fd9-48c8-b3f8-45327871b647-kube-api-access-4gfk4\") on node \"crc\" DevicePath \"\"" Mar 07 14:55:02 crc kubenswrapper[4943]: I0307 14:55:02.857231 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb540778-9fd9-48c8-b3f8-45327871b647-util" (OuterVolumeSpecName: "util") pod "fb540778-9fd9-48c8-b3f8-45327871b647" (UID: "fb540778-9fd9-48c8-b3f8-45327871b647"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:55:02 crc kubenswrapper[4943]: I0307 14:55:02.907431 4943 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fb540778-9fd9-48c8-b3f8-45327871b647-util\") on node \"crc\" DevicePath \"\"" Mar 07 14:55:03 crc kubenswrapper[4943]: I0307 14:55:03.094178 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40frz9l" Mar 07 14:55:03 crc kubenswrapper[4943]: I0307 14:55:03.094247 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40frz9l" event={"ID":"fb540778-9fd9-48c8-b3f8-45327871b647","Type":"ContainerDied","Data":"c747fc18557bf7f4d1e1ad0dd1b4053367143cea9471a55efdcb4b1af182c948"} Mar 07 14:55:03 crc kubenswrapper[4943]: I0307 14:55:03.094285 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c747fc18557bf7f4d1e1ad0dd1b4053367143cea9471a55efdcb4b1af182c948" Mar 07 14:55:05 crc kubenswrapper[4943]: I0307 14:55:05.673306 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mxdxf"] Mar 07 14:55:05 crc kubenswrapper[4943]: I0307 14:55:05.674032 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mxdxf" podUID="eed0d6a6-a45f-4e6a-85f1-9c7157055498" containerName="registry-server" containerID="cri-o://25d9b42d87f972ad4f1b77373d5655da4ea30456d7ead6ee627e91705934cbf1" gracePeriod=2 Mar 07 14:55:06 crc kubenswrapper[4943]: I0307 14:55:06.074707 4943 patch_prober.go:28] interesting pod/machine-config-daemon-cpdmm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 07 14:55:06 crc kubenswrapper[4943]: I0307 14:55:06.074801 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 07 14:55:06 crc kubenswrapper[4943]: I0307 14:55:06.123213 4943 generic.go:334] "Generic (PLEG): container finished" podID="eed0d6a6-a45f-4e6a-85f1-9c7157055498" containerID="25d9b42d87f972ad4f1b77373d5655da4ea30456d7ead6ee627e91705934cbf1" exitCode=0 Mar 07 14:55:06 crc kubenswrapper[4943]: I0307 14:55:06.123274 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mxdxf" event={"ID":"eed0d6a6-a45f-4e6a-85f1-9c7157055498","Type":"ContainerDied","Data":"25d9b42d87f972ad4f1b77373d5655da4ea30456d7ead6ee627e91705934cbf1"} Mar 07 14:55:06 crc kubenswrapper[4943]: I0307 14:55:06.752109 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mxdxf" Mar 07 14:55:06 crc kubenswrapper[4943]: I0307 14:55:06.868027 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eed0d6a6-a45f-4e6a-85f1-9c7157055498-utilities\") pod \"eed0d6a6-a45f-4e6a-85f1-9c7157055498\" (UID: \"eed0d6a6-a45f-4e6a-85f1-9c7157055498\") " Mar 07 14:55:06 crc kubenswrapper[4943]: I0307 14:55:06.868110 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eed0d6a6-a45f-4e6a-85f1-9c7157055498-catalog-content\") pod \"eed0d6a6-a45f-4e6a-85f1-9c7157055498\" (UID: \"eed0d6a6-a45f-4e6a-85f1-9c7157055498\") " Mar 07 14:55:06 crc kubenswrapper[4943]: I0307 14:55:06.868184 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-th2gh\" (UniqueName: \"kubernetes.io/projected/eed0d6a6-a45f-4e6a-85f1-9c7157055498-kube-api-access-th2gh\") pod \"eed0d6a6-a45f-4e6a-85f1-9c7157055498\" (UID: \"eed0d6a6-a45f-4e6a-85f1-9c7157055498\") " Mar 07 14:55:06 crc kubenswrapper[4943]: I0307 14:55:06.869567 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eed0d6a6-a45f-4e6a-85f1-9c7157055498-utilities" (OuterVolumeSpecName: "utilities") pod "eed0d6a6-a45f-4e6a-85f1-9c7157055498" (UID: "eed0d6a6-a45f-4e6a-85f1-9c7157055498"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:55:06 crc kubenswrapper[4943]: I0307 14:55:06.875508 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eed0d6a6-a45f-4e6a-85f1-9c7157055498-kube-api-access-th2gh" (OuterVolumeSpecName: "kube-api-access-th2gh") pod "eed0d6a6-a45f-4e6a-85f1-9c7157055498" (UID: "eed0d6a6-a45f-4e6a-85f1-9c7157055498"). InnerVolumeSpecName "kube-api-access-th2gh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:55:06 crc kubenswrapper[4943]: I0307 14:55:06.931452 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eed0d6a6-a45f-4e6a-85f1-9c7157055498-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eed0d6a6-a45f-4e6a-85f1-9c7157055498" (UID: "eed0d6a6-a45f-4e6a-85f1-9c7157055498"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:55:06 crc kubenswrapper[4943]: I0307 14:55:06.970133 4943 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eed0d6a6-a45f-4e6a-85f1-9c7157055498-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 07 14:55:06 crc kubenswrapper[4943]: I0307 14:55:06.970174 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-th2gh\" (UniqueName: \"kubernetes.io/projected/eed0d6a6-a45f-4e6a-85f1-9c7157055498-kube-api-access-th2gh\") on node \"crc\" DevicePath \"\"" Mar 07 14:55:06 crc kubenswrapper[4943]: I0307 14:55:06.970192 4943 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eed0d6a6-a45f-4e6a-85f1-9c7157055498-utilities\") on node \"crc\" DevicePath \"\"" Mar 07 14:55:07 crc kubenswrapper[4943]: I0307 14:55:07.135527 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mxdxf" event={"ID":"eed0d6a6-a45f-4e6a-85f1-9c7157055498","Type":"ContainerDied","Data":"0335d5b206c8841c4d083ff77acddc44ed59dc43ef15c601af9fba723489b384"} Mar 07 14:55:07 crc kubenswrapper[4943]: I0307 14:55:07.135594 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mxdxf" Mar 07 14:55:07 crc kubenswrapper[4943]: I0307 14:55:07.135621 4943 scope.go:117] "RemoveContainer" containerID="25d9b42d87f972ad4f1b77373d5655da4ea30456d7ead6ee627e91705934cbf1" Mar 07 14:55:07 crc kubenswrapper[4943]: I0307 14:55:07.163808 4943 scope.go:117] "RemoveContainer" containerID="07b61ad3582c5dfa1fd3d97b58b3b7fadba821e754a0e083515e5e4aed44c1b3" Mar 07 14:55:07 crc kubenswrapper[4943]: I0307 14:55:07.177444 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mxdxf"] Mar 07 14:55:07 crc kubenswrapper[4943]: I0307 14:55:07.183856 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mxdxf"] Mar 07 14:55:07 crc kubenswrapper[4943]: I0307 14:55:07.211047 4943 scope.go:117] "RemoveContainer" containerID="01145689e18e84dab85ace54fc50e03257460ad9e7ca59d531c189d8b5260dd7" Mar 07 14:55:08 crc kubenswrapper[4943]: I0307 14:55:08.762141 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eed0d6a6-a45f-4e6a-85f1-9c7157055498" path="/var/lib/kubelet/pods/eed0d6a6-a45f-4e6a-85f1-9c7157055498/volumes" Mar 07 14:55:13 crc kubenswrapper[4943]: I0307 14:55:13.659233 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7d8dd5b6b9-q7chd"] Mar 07 14:55:13 crc kubenswrapper[4943]: E0307 14:55:13.660062 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eed0d6a6-a45f-4e6a-85f1-9c7157055498" containerName="extract-utilities" Mar 07 14:55:13 crc kubenswrapper[4943]: I0307 14:55:13.660074 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="eed0d6a6-a45f-4e6a-85f1-9c7157055498" containerName="extract-utilities" Mar 07 14:55:13 crc kubenswrapper[4943]: E0307 14:55:13.660082 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eed0d6a6-a45f-4e6a-85f1-9c7157055498" containerName="registry-server" Mar 07 14:55:13 crc kubenswrapper[4943]: I0307 14:55:13.660087 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="eed0d6a6-a45f-4e6a-85f1-9c7157055498" containerName="registry-server" Mar 07 14:55:13 crc kubenswrapper[4943]: E0307 14:55:13.660096 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb540778-9fd9-48c8-b3f8-45327871b647" containerName="util" Mar 07 14:55:13 crc kubenswrapper[4943]: I0307 14:55:13.660102 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb540778-9fd9-48c8-b3f8-45327871b647" containerName="util" Mar 07 14:55:13 crc kubenswrapper[4943]: E0307 14:55:13.660118 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb540778-9fd9-48c8-b3f8-45327871b647" containerName="pull" Mar 07 14:55:13 crc kubenswrapper[4943]: I0307 14:55:13.660124 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb540778-9fd9-48c8-b3f8-45327871b647" containerName="pull" Mar 07 14:55:13 crc kubenswrapper[4943]: E0307 14:55:13.660133 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb540778-9fd9-48c8-b3f8-45327871b647" containerName="extract" Mar 07 14:55:13 crc kubenswrapper[4943]: I0307 14:55:13.660139 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb540778-9fd9-48c8-b3f8-45327871b647" containerName="extract" Mar 07 14:55:13 crc kubenswrapper[4943]: E0307 14:55:13.660146 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eed0d6a6-a45f-4e6a-85f1-9c7157055498" containerName="extract-content" Mar 07 14:55:13 crc kubenswrapper[4943]: I0307 14:55:13.660151 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="eed0d6a6-a45f-4e6a-85f1-9c7157055498" containerName="extract-content" Mar 07 14:55:13 crc kubenswrapper[4943]: I0307 14:55:13.660260 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb540778-9fd9-48c8-b3f8-45327871b647" containerName="extract" Mar 07 14:55:13 crc kubenswrapper[4943]: I0307 14:55:13.660269 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="eed0d6a6-a45f-4e6a-85f1-9c7157055498" containerName="registry-server" Mar 07 14:55:13 crc kubenswrapper[4943]: I0307 14:55:13.660689 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7d8dd5b6b9-q7chd" Mar 07 14:55:13 crc kubenswrapper[4943]: I0307 14:55:13.664625 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-service-cert" Mar 07 14:55:13 crc kubenswrapper[4943]: I0307 14:55:13.666180 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-vwrkf" Mar 07 14:55:13 crc kubenswrapper[4943]: I0307 14:55:13.679908 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7d8dd5b6b9-q7chd"] Mar 07 14:55:13 crc kubenswrapper[4943]: I0307 14:55:13.772115 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/83c1e172-0efd-48eb-bfe3-b35302e82114-apiservice-cert\") pod \"keystone-operator-controller-manager-7d8dd5b6b9-q7chd\" (UID: \"83c1e172-0efd-48eb-bfe3-b35302e82114\") " pod="openstack-operators/keystone-operator-controller-manager-7d8dd5b6b9-q7chd" Mar 07 14:55:13 crc kubenswrapper[4943]: I0307 14:55:13.772174 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mb9qc\" (UniqueName: \"kubernetes.io/projected/83c1e172-0efd-48eb-bfe3-b35302e82114-kube-api-access-mb9qc\") pod \"keystone-operator-controller-manager-7d8dd5b6b9-q7chd\" (UID: \"83c1e172-0efd-48eb-bfe3-b35302e82114\") " pod="openstack-operators/keystone-operator-controller-manager-7d8dd5b6b9-q7chd" Mar 07 14:55:13 crc kubenswrapper[4943]: I0307 14:55:13.772373 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/83c1e172-0efd-48eb-bfe3-b35302e82114-webhook-cert\") pod \"keystone-operator-controller-manager-7d8dd5b6b9-q7chd\" (UID: \"83c1e172-0efd-48eb-bfe3-b35302e82114\") " pod="openstack-operators/keystone-operator-controller-manager-7d8dd5b6b9-q7chd" Mar 07 14:55:13 crc kubenswrapper[4943]: I0307 14:55:13.874120 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/83c1e172-0efd-48eb-bfe3-b35302e82114-apiservice-cert\") pod \"keystone-operator-controller-manager-7d8dd5b6b9-q7chd\" (UID: \"83c1e172-0efd-48eb-bfe3-b35302e82114\") " pod="openstack-operators/keystone-operator-controller-manager-7d8dd5b6b9-q7chd" Mar 07 14:55:13 crc kubenswrapper[4943]: I0307 14:55:13.874218 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mb9qc\" (UniqueName: \"kubernetes.io/projected/83c1e172-0efd-48eb-bfe3-b35302e82114-kube-api-access-mb9qc\") pod \"keystone-operator-controller-manager-7d8dd5b6b9-q7chd\" (UID: \"83c1e172-0efd-48eb-bfe3-b35302e82114\") " pod="openstack-operators/keystone-operator-controller-manager-7d8dd5b6b9-q7chd" Mar 07 14:55:13 crc kubenswrapper[4943]: I0307 14:55:13.874297 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/83c1e172-0efd-48eb-bfe3-b35302e82114-webhook-cert\") pod \"keystone-operator-controller-manager-7d8dd5b6b9-q7chd\" (UID: \"83c1e172-0efd-48eb-bfe3-b35302e82114\") " pod="openstack-operators/keystone-operator-controller-manager-7d8dd5b6b9-q7chd" Mar 07 14:55:13 crc kubenswrapper[4943]: I0307 14:55:13.880849 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/83c1e172-0efd-48eb-bfe3-b35302e82114-apiservice-cert\") pod \"keystone-operator-controller-manager-7d8dd5b6b9-q7chd\" (UID: \"83c1e172-0efd-48eb-bfe3-b35302e82114\") " pod="openstack-operators/keystone-operator-controller-manager-7d8dd5b6b9-q7chd" Mar 07 14:55:13 crc kubenswrapper[4943]: I0307 14:55:13.890606 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/83c1e172-0efd-48eb-bfe3-b35302e82114-webhook-cert\") pod \"keystone-operator-controller-manager-7d8dd5b6b9-q7chd\" (UID: \"83c1e172-0efd-48eb-bfe3-b35302e82114\") " pod="openstack-operators/keystone-operator-controller-manager-7d8dd5b6b9-q7chd" Mar 07 14:55:13 crc kubenswrapper[4943]: I0307 14:55:13.892766 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mb9qc\" (UniqueName: \"kubernetes.io/projected/83c1e172-0efd-48eb-bfe3-b35302e82114-kube-api-access-mb9qc\") pod \"keystone-operator-controller-manager-7d8dd5b6b9-q7chd\" (UID: \"83c1e172-0efd-48eb-bfe3-b35302e82114\") " pod="openstack-operators/keystone-operator-controller-manager-7d8dd5b6b9-q7chd" Mar 07 14:55:13 crc kubenswrapper[4943]: I0307 14:55:13.979069 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7d8dd5b6b9-q7chd" Mar 07 14:55:14 crc kubenswrapper[4943]: I0307 14:55:14.462851 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7d8dd5b6b9-q7chd"] Mar 07 14:55:15 crc kubenswrapper[4943]: I0307 14:55:15.205764 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7d8dd5b6b9-q7chd" event={"ID":"83c1e172-0efd-48eb-bfe3-b35302e82114","Type":"ContainerStarted","Data":"ff981591ad14681a547dd240f30a8a49d7f83075135b6dbe5dd7c4efc263b252"} Mar 07 14:55:18 crc kubenswrapper[4943]: I0307 14:55:18.252577 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7d8dd5b6b9-q7chd" event={"ID":"83c1e172-0efd-48eb-bfe3-b35302e82114","Type":"ContainerStarted","Data":"05ca08699001e381eda0ca00df6ed726f8a814c48b8bd8c05162f014fdb41551"} Mar 07 14:55:18 crc kubenswrapper[4943]: I0307 14:55:18.253622 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7d8dd5b6b9-q7chd" Mar 07 14:55:18 crc kubenswrapper[4943]: I0307 14:55:18.283572 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7d8dd5b6b9-q7chd" podStartSLOduration=1.817903983 podStartE2EDuration="5.283553755s" podCreationTimestamp="2026-03-07 14:55:13 +0000 UTC" firstStartedPulling="2026-03-07 14:55:14.48004828 +0000 UTC m=+956.432184778" lastFinishedPulling="2026-03-07 14:55:17.945698012 +0000 UTC m=+959.897834550" observedRunningTime="2026-03-07 14:55:18.283099224 +0000 UTC m=+960.235235782" watchObservedRunningTime="2026-03-07 14:55:18.283553755 +0000 UTC m=+960.235690273" Mar 07 14:55:20 crc kubenswrapper[4943]: I0307 14:55:20.281721 4943 generic.go:334] "Generic (PLEG): container finished" podID="f73c32b4-3904-4bf3-9bb4-13750cb45e88" containerID="789f2f6e1ba35f3c76fb7b16e3606a4d28e4b8aa980148e08397fb5067ab1532" exitCode=0 Mar 07 14:55:20 crc kubenswrapper[4943]: I0307 14:55:20.281853 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/rabbitmq-server-0" event={"ID":"f73c32b4-3904-4bf3-9bb4-13750cb45e88","Type":"ContainerDied","Data":"789f2f6e1ba35f3c76fb7b16e3606a4d28e4b8aa980148e08397fb5067ab1532"} Mar 07 14:55:21 crc kubenswrapper[4943]: I0307 14:55:21.291475 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/rabbitmq-server-0" event={"ID":"f73c32b4-3904-4bf3-9bb4-13750cb45e88","Type":"ContainerStarted","Data":"7116297f07239e80d7c5a301752da51c12e70aeb39c2620a5899d4decfed165f"} Mar 07 14:55:21 crc kubenswrapper[4943]: I0307 14:55:21.292032 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="swift-kuttl-tests/rabbitmq-server-0" Mar 07 14:55:21 crc kubenswrapper[4943]: I0307 14:55:21.322894 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/rabbitmq-server-0" podStartSLOduration=37.19163011 podStartE2EDuration="44.322864089s" podCreationTimestamp="2026-03-07 14:54:37 +0000 UTC" firstStartedPulling="2026-03-07 14:54:39.636414343 +0000 UTC m=+921.588550851" lastFinishedPulling="2026-03-07 14:54:46.767648342 +0000 UTC m=+928.719784830" observedRunningTime="2026-03-07 14:55:21.318145883 +0000 UTC m=+963.270282451" watchObservedRunningTime="2026-03-07 14:55:21.322864089 +0000 UTC m=+963.275000617" Mar 07 14:55:23 crc kubenswrapper[4943]: I0307 14:55:23.986905 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7d8dd5b6b9-q7chd" Mar 07 14:55:30 crc kubenswrapper[4943]: I0307 14:55:30.886594 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-index-z6wwq"] Mar 07 14:55:30 crc kubenswrapper[4943]: I0307 14:55:30.888692 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-index-z6wwq" Mar 07 14:55:30 crc kubenswrapper[4943]: I0307 14:55:30.892220 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-index-dockercfg-7bft7" Mar 07 14:55:30 crc kubenswrapper[4943]: I0307 14:55:30.896973 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-index-z6wwq"] Mar 07 14:55:30 crc kubenswrapper[4943]: I0307 14:55:30.968846 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxlsw\" (UniqueName: \"kubernetes.io/projected/08b0d828-0a74-4630-bffb-815ec50fda48-kube-api-access-sxlsw\") pod \"barbican-operator-index-z6wwq\" (UID: \"08b0d828-0a74-4630-bffb-815ec50fda48\") " pod="openstack-operators/barbican-operator-index-z6wwq" Mar 07 14:55:31 crc kubenswrapper[4943]: I0307 14:55:31.070660 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxlsw\" (UniqueName: \"kubernetes.io/projected/08b0d828-0a74-4630-bffb-815ec50fda48-kube-api-access-sxlsw\") pod \"barbican-operator-index-z6wwq\" (UID: \"08b0d828-0a74-4630-bffb-815ec50fda48\") " pod="openstack-operators/barbican-operator-index-z6wwq" Mar 07 14:55:31 crc kubenswrapper[4943]: I0307 14:55:31.093511 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxlsw\" (UniqueName: \"kubernetes.io/projected/08b0d828-0a74-4630-bffb-815ec50fda48-kube-api-access-sxlsw\") pod \"barbican-operator-index-z6wwq\" (UID: \"08b0d828-0a74-4630-bffb-815ec50fda48\") " pod="openstack-operators/barbican-operator-index-z6wwq" Mar 07 14:55:31 crc kubenswrapper[4943]: I0307 14:55:31.219242 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-index-z6wwq" Mar 07 14:55:31 crc kubenswrapper[4943]: W0307 14:55:31.730628 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod08b0d828_0a74_4630_bffb_815ec50fda48.slice/crio-963255c2b001e6955c673e4a98a54e9b705ed174b9a2b1709c2a1a061d430263 WatchSource:0}: Error finding container 963255c2b001e6955c673e4a98a54e9b705ed174b9a2b1709c2a1a061d430263: Status 404 returned error can't find the container with id 963255c2b001e6955c673e4a98a54e9b705ed174b9a2b1709c2a1a061d430263 Mar 07 14:55:31 crc kubenswrapper[4943]: I0307 14:55:31.732405 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-index-z6wwq"] Mar 07 14:55:32 crc kubenswrapper[4943]: I0307 14:55:32.393121 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-index-z6wwq" event={"ID":"08b0d828-0a74-4630-bffb-815ec50fda48","Type":"ContainerStarted","Data":"963255c2b001e6955c673e4a98a54e9b705ed174b9a2b1709c2a1a061d430263"} Mar 07 14:55:34 crc kubenswrapper[4943]: I0307 14:55:34.411791 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-index-z6wwq" event={"ID":"08b0d828-0a74-4630-bffb-815ec50fda48","Type":"ContainerStarted","Data":"975c86a627c499cf6663307077ebd05c17fd94f95318b04f4a4b7399ad5f81f4"} Mar 07 14:55:34 crc kubenswrapper[4943]: I0307 14:55:34.434642 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-index-z6wwq" podStartSLOduration=2.812709428 podStartE2EDuration="4.434612579s" podCreationTimestamp="2026-03-07 14:55:30 +0000 UTC" firstStartedPulling="2026-03-07 14:55:31.733355976 +0000 UTC m=+973.685492484" lastFinishedPulling="2026-03-07 14:55:33.355259107 +0000 UTC m=+975.307395635" observedRunningTime="2026-03-07 14:55:34.42937091 +0000 UTC m=+976.381507438" watchObservedRunningTime="2026-03-07 14:55:34.434612579 +0000 UTC m=+976.386749117" Mar 07 14:55:36 crc kubenswrapper[4943]: I0307 14:55:36.074819 4943 patch_prober.go:28] interesting pod/machine-config-daemon-cpdmm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 07 14:55:36 crc kubenswrapper[4943]: I0307 14:55:36.075273 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 07 14:55:38 crc kubenswrapper[4943]: I0307 14:55:38.433786 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/keystone-db-create-r2wkc"] Mar 07 14:55:38 crc kubenswrapper[4943]: I0307 14:55:38.435412 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/keystone-db-create-r2wkc" Mar 07 14:55:38 crc kubenswrapper[4943]: I0307 14:55:38.448296 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/keystone-c24b-account-create-update-jj69g"] Mar 07 14:55:38 crc kubenswrapper[4943]: I0307 14:55:38.449656 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/keystone-c24b-account-create-update-jj69g" Mar 07 14:55:38 crc kubenswrapper[4943]: I0307 14:55:38.451552 4943 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"keystone-db-secret" Mar 07 14:55:38 crc kubenswrapper[4943]: I0307 14:55:38.463221 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/keystone-db-create-r2wkc"] Mar 07 14:55:38 crc kubenswrapper[4943]: I0307 14:55:38.474455 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/keystone-c24b-account-create-update-jj69g"] Mar 07 14:55:38 crc kubenswrapper[4943]: I0307 14:55:38.492211 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d94e16f-d333-4d99-b7e8-441cfb8d7252-operator-scripts\") pod \"keystone-db-create-r2wkc\" (UID: \"2d94e16f-d333-4d99-b7e8-441cfb8d7252\") " pod="swift-kuttl-tests/keystone-db-create-r2wkc" Mar 07 14:55:38 crc kubenswrapper[4943]: I0307 14:55:38.492477 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8e5f2081-60f4-4af6-ab3f-50ba7ba11c6c-operator-scripts\") pod \"keystone-c24b-account-create-update-jj69g\" (UID: \"8e5f2081-60f4-4af6-ab3f-50ba7ba11c6c\") " pod="swift-kuttl-tests/keystone-c24b-account-create-update-jj69g" Mar 07 14:55:38 crc kubenswrapper[4943]: I0307 14:55:38.492585 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tq5l\" (UniqueName: \"kubernetes.io/projected/2d94e16f-d333-4d99-b7e8-441cfb8d7252-kube-api-access-2tq5l\") pod \"keystone-db-create-r2wkc\" (UID: \"2d94e16f-d333-4d99-b7e8-441cfb8d7252\") " pod="swift-kuttl-tests/keystone-db-create-r2wkc" Mar 07 14:55:38 crc kubenswrapper[4943]: I0307 14:55:38.500124 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzqcm\" (UniqueName: \"kubernetes.io/projected/8e5f2081-60f4-4af6-ab3f-50ba7ba11c6c-kube-api-access-tzqcm\") pod \"keystone-c24b-account-create-update-jj69g\" (UID: \"8e5f2081-60f4-4af6-ab3f-50ba7ba11c6c\") " pod="swift-kuttl-tests/keystone-c24b-account-create-update-jj69g" Mar 07 14:55:38 crc kubenswrapper[4943]: I0307 14:55:38.601773 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tq5l\" (UniqueName: \"kubernetes.io/projected/2d94e16f-d333-4d99-b7e8-441cfb8d7252-kube-api-access-2tq5l\") pod \"keystone-db-create-r2wkc\" (UID: \"2d94e16f-d333-4d99-b7e8-441cfb8d7252\") " pod="swift-kuttl-tests/keystone-db-create-r2wkc" Mar 07 14:55:38 crc kubenswrapper[4943]: I0307 14:55:38.601816 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzqcm\" (UniqueName: \"kubernetes.io/projected/8e5f2081-60f4-4af6-ab3f-50ba7ba11c6c-kube-api-access-tzqcm\") pod \"keystone-c24b-account-create-update-jj69g\" (UID: \"8e5f2081-60f4-4af6-ab3f-50ba7ba11c6c\") " pod="swift-kuttl-tests/keystone-c24b-account-create-update-jj69g" Mar 07 14:55:38 crc kubenswrapper[4943]: I0307 14:55:38.602115 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d94e16f-d333-4d99-b7e8-441cfb8d7252-operator-scripts\") pod \"keystone-db-create-r2wkc\" (UID: \"2d94e16f-d333-4d99-b7e8-441cfb8d7252\") " pod="swift-kuttl-tests/keystone-db-create-r2wkc" Mar 07 14:55:38 crc kubenswrapper[4943]: I0307 14:55:38.602141 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8e5f2081-60f4-4af6-ab3f-50ba7ba11c6c-operator-scripts\") pod \"keystone-c24b-account-create-update-jj69g\" (UID: \"8e5f2081-60f4-4af6-ab3f-50ba7ba11c6c\") " pod="swift-kuttl-tests/keystone-c24b-account-create-update-jj69g" Mar 07 14:55:38 crc kubenswrapper[4943]: I0307 14:55:38.602823 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8e5f2081-60f4-4af6-ab3f-50ba7ba11c6c-operator-scripts\") pod \"keystone-c24b-account-create-update-jj69g\" (UID: \"8e5f2081-60f4-4af6-ab3f-50ba7ba11c6c\") " pod="swift-kuttl-tests/keystone-c24b-account-create-update-jj69g" Mar 07 14:55:38 crc kubenswrapper[4943]: I0307 14:55:38.602825 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d94e16f-d333-4d99-b7e8-441cfb8d7252-operator-scripts\") pod \"keystone-db-create-r2wkc\" (UID: \"2d94e16f-d333-4d99-b7e8-441cfb8d7252\") " pod="swift-kuttl-tests/keystone-db-create-r2wkc" Mar 07 14:55:38 crc kubenswrapper[4943]: I0307 14:55:38.621907 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzqcm\" (UniqueName: \"kubernetes.io/projected/8e5f2081-60f4-4af6-ab3f-50ba7ba11c6c-kube-api-access-tzqcm\") pod \"keystone-c24b-account-create-update-jj69g\" (UID: \"8e5f2081-60f4-4af6-ab3f-50ba7ba11c6c\") " pod="swift-kuttl-tests/keystone-c24b-account-create-update-jj69g" Mar 07 14:55:38 crc kubenswrapper[4943]: I0307 14:55:38.625954 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tq5l\" (UniqueName: \"kubernetes.io/projected/2d94e16f-d333-4d99-b7e8-441cfb8d7252-kube-api-access-2tq5l\") pod \"keystone-db-create-r2wkc\" (UID: \"2d94e16f-d333-4d99-b7e8-441cfb8d7252\") " pod="swift-kuttl-tests/keystone-db-create-r2wkc" Mar 07 14:55:38 crc kubenswrapper[4943]: I0307 14:55:38.753133 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/keystone-db-create-r2wkc" Mar 07 14:55:38 crc kubenswrapper[4943]: I0307 14:55:38.766238 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/keystone-c24b-account-create-update-jj69g" Mar 07 14:55:39 crc kubenswrapper[4943]: I0307 14:55:39.083162 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/keystone-c24b-account-create-update-jj69g"] Mar 07 14:55:39 crc kubenswrapper[4943]: I0307 14:55:39.141340 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="swift-kuttl-tests/rabbitmq-server-0" Mar 07 14:55:39 crc kubenswrapper[4943]: I0307 14:55:39.233273 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/keystone-db-create-r2wkc"] Mar 07 14:55:39 crc kubenswrapper[4943]: W0307 14:55:39.234819 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2d94e16f_d333_4d99_b7e8_441cfb8d7252.slice/crio-2ac944c30d68f44c5a785e64e6776dfbb8f5fba6eb911d3c84277e0f05304a75 WatchSource:0}: Error finding container 2ac944c30d68f44c5a785e64e6776dfbb8f5fba6eb911d3c84277e0f05304a75: Status 404 returned error can't find the container with id 2ac944c30d68f44c5a785e64e6776dfbb8f5fba6eb911d3c84277e0f05304a75 Mar 07 14:55:39 crc kubenswrapper[4943]: I0307 14:55:39.448922 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/keystone-db-create-r2wkc" event={"ID":"2d94e16f-d333-4d99-b7e8-441cfb8d7252","Type":"ContainerStarted","Data":"7868657e9c49897c842757eadb44bc3eac4bcdd4e427e2a8d8b43b9a3362e40c"} Mar 07 14:55:39 crc kubenswrapper[4943]: I0307 14:55:39.450474 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/keystone-db-create-r2wkc" event={"ID":"2d94e16f-d333-4d99-b7e8-441cfb8d7252","Type":"ContainerStarted","Data":"2ac944c30d68f44c5a785e64e6776dfbb8f5fba6eb911d3c84277e0f05304a75"} Mar 07 14:55:39 crc kubenswrapper[4943]: I0307 14:55:39.450630 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/keystone-c24b-account-create-update-jj69g" event={"ID":"8e5f2081-60f4-4af6-ab3f-50ba7ba11c6c","Type":"ContainerStarted","Data":"cb632251025997ee50958ee0c15255ead5450b4159b77c62918a7990f1ad1002"} Mar 07 14:55:39 crc kubenswrapper[4943]: I0307 14:55:39.450769 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/keystone-c24b-account-create-update-jj69g" event={"ID":"8e5f2081-60f4-4af6-ab3f-50ba7ba11c6c","Type":"ContainerStarted","Data":"ffdab450b363575efd54565b51263ba2d7f90b755caac2ec6e4c1fa0c1efc946"} Mar 07 14:55:39 crc kubenswrapper[4943]: I0307 14:55:39.472917 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/keystone-db-create-r2wkc" podStartSLOduration=1.472893403 podStartE2EDuration="1.472893403s" podCreationTimestamp="2026-03-07 14:55:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:55:39.466159557 +0000 UTC m=+981.418296055" watchObservedRunningTime="2026-03-07 14:55:39.472893403 +0000 UTC m=+981.425029891" Mar 07 14:55:39 crc kubenswrapper[4943]: I0307 14:55:39.497032 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/keystone-c24b-account-create-update-jj69g" podStartSLOduration=1.497006756 podStartE2EDuration="1.497006756s" podCreationTimestamp="2026-03-07 14:55:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:55:39.490467815 +0000 UTC m=+981.442604313" watchObservedRunningTime="2026-03-07 14:55:39.497006756 +0000 UTC m=+981.449143254" Mar 07 14:55:40 crc kubenswrapper[4943]: I0307 14:55:40.463067 4943 generic.go:334] "Generic (PLEG): container finished" podID="8e5f2081-60f4-4af6-ab3f-50ba7ba11c6c" containerID="cb632251025997ee50958ee0c15255ead5450b4159b77c62918a7990f1ad1002" exitCode=0 Mar 07 14:55:40 crc kubenswrapper[4943]: I0307 14:55:40.463174 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/keystone-c24b-account-create-update-jj69g" event={"ID":"8e5f2081-60f4-4af6-ab3f-50ba7ba11c6c","Type":"ContainerDied","Data":"cb632251025997ee50958ee0c15255ead5450b4159b77c62918a7990f1ad1002"} Mar 07 14:55:40 crc kubenswrapper[4943]: I0307 14:55:40.465793 4943 generic.go:334] "Generic (PLEG): container finished" podID="2d94e16f-d333-4d99-b7e8-441cfb8d7252" containerID="7868657e9c49897c842757eadb44bc3eac4bcdd4e427e2a8d8b43b9a3362e40c" exitCode=0 Mar 07 14:55:40 crc kubenswrapper[4943]: I0307 14:55:40.465865 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/keystone-db-create-r2wkc" event={"ID":"2d94e16f-d333-4d99-b7e8-441cfb8d7252","Type":"ContainerDied","Data":"7868657e9c49897c842757eadb44bc3eac4bcdd4e427e2a8d8b43b9a3362e40c"} Mar 07 14:55:41 crc kubenswrapper[4943]: I0307 14:55:41.220025 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-index-z6wwq" Mar 07 14:55:41 crc kubenswrapper[4943]: I0307 14:55:41.220580 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/barbican-operator-index-z6wwq" Mar 07 14:55:41 crc kubenswrapper[4943]: I0307 14:55:41.262262 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/barbican-operator-index-z6wwq" Mar 07 14:55:41 crc kubenswrapper[4943]: I0307 14:55:41.503838 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-index-z6wwq" Mar 07 14:55:41 crc kubenswrapper[4943]: I0307 14:55:41.940698 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/keystone-db-create-r2wkc" Mar 07 14:55:41 crc kubenswrapper[4943]: I0307 14:55:41.947677 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/keystone-c24b-account-create-update-jj69g" Mar 07 14:55:42 crc kubenswrapper[4943]: I0307 14:55:42.055707 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8e5f2081-60f4-4af6-ab3f-50ba7ba11c6c-operator-scripts\") pod \"8e5f2081-60f4-4af6-ab3f-50ba7ba11c6c\" (UID: \"8e5f2081-60f4-4af6-ab3f-50ba7ba11c6c\") " Mar 07 14:55:42 crc kubenswrapper[4943]: I0307 14:55:42.055980 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2tq5l\" (UniqueName: \"kubernetes.io/projected/2d94e16f-d333-4d99-b7e8-441cfb8d7252-kube-api-access-2tq5l\") pod \"2d94e16f-d333-4d99-b7e8-441cfb8d7252\" (UID: \"2d94e16f-d333-4d99-b7e8-441cfb8d7252\") " Mar 07 14:55:42 crc kubenswrapper[4943]: I0307 14:55:42.056151 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d94e16f-d333-4d99-b7e8-441cfb8d7252-operator-scripts\") pod \"2d94e16f-d333-4d99-b7e8-441cfb8d7252\" (UID: \"2d94e16f-d333-4d99-b7e8-441cfb8d7252\") " Mar 07 14:55:42 crc kubenswrapper[4943]: I0307 14:55:42.056199 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tzqcm\" (UniqueName: \"kubernetes.io/projected/8e5f2081-60f4-4af6-ab3f-50ba7ba11c6c-kube-api-access-tzqcm\") pod \"8e5f2081-60f4-4af6-ab3f-50ba7ba11c6c\" (UID: \"8e5f2081-60f4-4af6-ab3f-50ba7ba11c6c\") " Mar 07 14:55:42 crc kubenswrapper[4943]: I0307 14:55:42.057114 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e5f2081-60f4-4af6-ab3f-50ba7ba11c6c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8e5f2081-60f4-4af6-ab3f-50ba7ba11c6c" (UID: "8e5f2081-60f4-4af6-ab3f-50ba7ba11c6c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:55:42 crc kubenswrapper[4943]: I0307 14:55:42.057277 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d94e16f-d333-4d99-b7e8-441cfb8d7252-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2d94e16f-d333-4d99-b7e8-441cfb8d7252" (UID: "2d94e16f-d333-4d99-b7e8-441cfb8d7252"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:55:42 crc kubenswrapper[4943]: I0307 14:55:42.068224 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d94e16f-d333-4d99-b7e8-441cfb8d7252-kube-api-access-2tq5l" (OuterVolumeSpecName: "kube-api-access-2tq5l") pod "2d94e16f-d333-4d99-b7e8-441cfb8d7252" (UID: "2d94e16f-d333-4d99-b7e8-441cfb8d7252"). InnerVolumeSpecName "kube-api-access-2tq5l". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:55:42 crc kubenswrapper[4943]: I0307 14:55:42.070239 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e5f2081-60f4-4af6-ab3f-50ba7ba11c6c-kube-api-access-tzqcm" (OuterVolumeSpecName: "kube-api-access-tzqcm") pod "8e5f2081-60f4-4af6-ab3f-50ba7ba11c6c" (UID: "8e5f2081-60f4-4af6-ab3f-50ba7ba11c6c"). InnerVolumeSpecName "kube-api-access-tzqcm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:55:42 crc kubenswrapper[4943]: I0307 14:55:42.158848 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2tq5l\" (UniqueName: \"kubernetes.io/projected/2d94e16f-d333-4d99-b7e8-441cfb8d7252-kube-api-access-2tq5l\") on node \"crc\" DevicePath \"\"" Mar 07 14:55:42 crc kubenswrapper[4943]: I0307 14:55:42.158901 4943 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d94e16f-d333-4d99-b7e8-441cfb8d7252-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 14:55:42 crc kubenswrapper[4943]: I0307 14:55:42.158922 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tzqcm\" (UniqueName: \"kubernetes.io/projected/8e5f2081-60f4-4af6-ab3f-50ba7ba11c6c-kube-api-access-tzqcm\") on node \"crc\" DevicePath \"\"" Mar 07 14:55:42 crc kubenswrapper[4943]: I0307 14:55:42.158969 4943 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8e5f2081-60f4-4af6-ab3f-50ba7ba11c6c-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 14:55:42 crc kubenswrapper[4943]: I0307 14:55:42.483344 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/keystone-c24b-account-create-update-jj69g" event={"ID":"8e5f2081-60f4-4af6-ab3f-50ba7ba11c6c","Type":"ContainerDied","Data":"ffdab450b363575efd54565b51263ba2d7f90b755caac2ec6e4c1fa0c1efc946"} Mar 07 14:55:42 crc kubenswrapper[4943]: I0307 14:55:42.483413 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ffdab450b363575efd54565b51263ba2d7f90b755caac2ec6e4c1fa0c1efc946" Mar 07 14:55:42 crc kubenswrapper[4943]: I0307 14:55:42.483899 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/keystone-c24b-account-create-update-jj69g" Mar 07 14:55:42 crc kubenswrapper[4943]: I0307 14:55:42.486186 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/keystone-db-create-r2wkc" Mar 07 14:55:42 crc kubenswrapper[4943]: I0307 14:55:42.486181 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/keystone-db-create-r2wkc" event={"ID":"2d94e16f-d333-4d99-b7e8-441cfb8d7252","Type":"ContainerDied","Data":"2ac944c30d68f44c5a785e64e6776dfbb8f5fba6eb911d3c84277e0f05304a75"} Mar 07 14:55:42 crc kubenswrapper[4943]: I0307 14:55:42.486406 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2ac944c30d68f44c5a785e64e6776dfbb8f5fba6eb911d3c84277e0f05304a75" Mar 07 14:55:43 crc kubenswrapper[4943]: I0307 14:55:43.738607 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar4jd2"] Mar 07 14:55:43 crc kubenswrapper[4943]: E0307 14:55:43.741572 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d94e16f-d333-4d99-b7e8-441cfb8d7252" containerName="mariadb-database-create" Mar 07 14:55:43 crc kubenswrapper[4943]: I0307 14:55:43.741737 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d94e16f-d333-4d99-b7e8-441cfb8d7252" containerName="mariadb-database-create" Mar 07 14:55:43 crc kubenswrapper[4943]: E0307 14:55:43.741887 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e5f2081-60f4-4af6-ab3f-50ba7ba11c6c" containerName="mariadb-account-create-update" Mar 07 14:55:43 crc kubenswrapper[4943]: I0307 14:55:43.742349 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e5f2081-60f4-4af6-ab3f-50ba7ba11c6c" containerName="mariadb-account-create-update" Mar 07 14:55:43 crc kubenswrapper[4943]: I0307 14:55:43.742706 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d94e16f-d333-4d99-b7e8-441cfb8d7252" containerName="mariadb-database-create" Mar 07 14:55:43 crc kubenswrapper[4943]: I0307 14:55:43.742850 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e5f2081-60f4-4af6-ab3f-50ba7ba11c6c" containerName="mariadb-account-create-update" Mar 07 14:55:43 crc kubenswrapper[4943]: I0307 14:55:43.745902 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar4jd2" Mar 07 14:55:43 crc kubenswrapper[4943]: I0307 14:55:43.749837 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-k6gp2" Mar 07 14:55:43 crc kubenswrapper[4943]: I0307 14:55:43.759062 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar4jd2"] Mar 07 14:55:43 crc kubenswrapper[4943]: I0307 14:55:43.781177 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1-bundle\") pod \"30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar4jd2\" (UID: \"3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1\") " pod="openstack-operators/30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar4jd2" Mar 07 14:55:43 crc kubenswrapper[4943]: I0307 14:55:43.781266 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmhpv\" (UniqueName: \"kubernetes.io/projected/3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1-kube-api-access-lmhpv\") pod \"30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar4jd2\" (UID: \"3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1\") " pod="openstack-operators/30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar4jd2" Mar 07 14:55:43 crc kubenswrapper[4943]: I0307 14:55:43.781348 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1-util\") pod \"30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar4jd2\" (UID: \"3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1\") " pod="openstack-operators/30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar4jd2" Mar 07 14:55:43 crc kubenswrapper[4943]: I0307 14:55:43.883367 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1-bundle\") pod \"30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar4jd2\" (UID: \"3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1\") " pod="openstack-operators/30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar4jd2" Mar 07 14:55:43 crc kubenswrapper[4943]: I0307 14:55:43.883439 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmhpv\" (UniqueName: \"kubernetes.io/projected/3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1-kube-api-access-lmhpv\") pod \"30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar4jd2\" (UID: \"3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1\") " pod="openstack-operators/30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar4jd2" Mar 07 14:55:43 crc kubenswrapper[4943]: I0307 14:55:43.883487 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1-util\") pod \"30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar4jd2\" (UID: \"3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1\") " pod="openstack-operators/30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar4jd2" Mar 07 14:55:43 crc kubenswrapper[4943]: I0307 14:55:43.884635 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1-bundle\") pod \"30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar4jd2\" (UID: \"3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1\") " pod="openstack-operators/30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar4jd2" Mar 07 14:55:43 crc kubenswrapper[4943]: I0307 14:55:43.884832 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1-util\") pod \"30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar4jd2\" (UID: \"3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1\") " pod="openstack-operators/30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar4jd2" Mar 07 14:55:43 crc kubenswrapper[4943]: I0307 14:55:43.920201 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmhpv\" (UniqueName: \"kubernetes.io/projected/3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1-kube-api-access-lmhpv\") pod \"30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar4jd2\" (UID: \"3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1\") " pod="openstack-operators/30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar4jd2" Mar 07 14:55:44 crc kubenswrapper[4943]: I0307 14:55:44.023832 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/keystone-db-sync-nq8hl"] Mar 07 14:55:44 crc kubenswrapper[4943]: I0307 14:55:44.026039 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/keystone-db-sync-nq8hl" Mar 07 14:55:44 crc kubenswrapper[4943]: I0307 14:55:44.029408 4943 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"keystone-keystone-dockercfg-vk5hc" Mar 07 14:55:44 crc kubenswrapper[4943]: I0307 14:55:44.029424 4943 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"keystone-scripts" Mar 07 14:55:44 crc kubenswrapper[4943]: I0307 14:55:44.029969 4943 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"keystone" Mar 07 14:55:44 crc kubenswrapper[4943]: I0307 14:55:44.030084 4943 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"keystone-config-data" Mar 07 14:55:44 crc kubenswrapper[4943]: I0307 14:55:44.043413 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/keystone-db-sync-nq8hl"] Mar 07 14:55:44 crc kubenswrapper[4943]: I0307 14:55:44.077341 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar4jd2" Mar 07 14:55:44 crc kubenswrapper[4943]: I0307 14:55:44.085923 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea23fd1e-8c76-4483-a624-be13a1715513-config-data\") pod \"keystone-db-sync-nq8hl\" (UID: \"ea23fd1e-8c76-4483-a624-be13a1715513\") " pod="swift-kuttl-tests/keystone-db-sync-nq8hl" Mar 07 14:55:44 crc kubenswrapper[4943]: I0307 14:55:44.086030 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7q5c\" (UniqueName: \"kubernetes.io/projected/ea23fd1e-8c76-4483-a624-be13a1715513-kube-api-access-j7q5c\") pod \"keystone-db-sync-nq8hl\" (UID: \"ea23fd1e-8c76-4483-a624-be13a1715513\") " pod="swift-kuttl-tests/keystone-db-sync-nq8hl" Mar 07 14:55:44 crc kubenswrapper[4943]: I0307 14:55:44.187179 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea23fd1e-8c76-4483-a624-be13a1715513-config-data\") pod \"keystone-db-sync-nq8hl\" (UID: \"ea23fd1e-8c76-4483-a624-be13a1715513\") " pod="swift-kuttl-tests/keystone-db-sync-nq8hl" Mar 07 14:55:44 crc kubenswrapper[4943]: I0307 14:55:44.187225 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7q5c\" (UniqueName: \"kubernetes.io/projected/ea23fd1e-8c76-4483-a624-be13a1715513-kube-api-access-j7q5c\") pod \"keystone-db-sync-nq8hl\" (UID: \"ea23fd1e-8c76-4483-a624-be13a1715513\") " pod="swift-kuttl-tests/keystone-db-sync-nq8hl" Mar 07 14:55:44 crc kubenswrapper[4943]: I0307 14:55:44.205458 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea23fd1e-8c76-4483-a624-be13a1715513-config-data\") pod \"keystone-db-sync-nq8hl\" (UID: \"ea23fd1e-8c76-4483-a624-be13a1715513\") " pod="swift-kuttl-tests/keystone-db-sync-nq8hl" Mar 07 14:55:44 crc kubenswrapper[4943]: I0307 14:55:44.214073 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7q5c\" (UniqueName: \"kubernetes.io/projected/ea23fd1e-8c76-4483-a624-be13a1715513-kube-api-access-j7q5c\") pod \"keystone-db-sync-nq8hl\" (UID: \"ea23fd1e-8c76-4483-a624-be13a1715513\") " pod="swift-kuttl-tests/keystone-db-sync-nq8hl" Mar 07 14:55:44 crc kubenswrapper[4943]: I0307 14:55:44.397989 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/keystone-db-sync-nq8hl" Mar 07 14:55:44 crc kubenswrapper[4943]: I0307 14:55:44.446489 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar4jd2"] Mar 07 14:55:44 crc kubenswrapper[4943]: I0307 14:55:44.502462 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar4jd2" event={"ID":"3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1","Type":"ContainerStarted","Data":"2a2beb0c51c482052788860fce14c2b983fd5f44ca3d13d21c1bdf0d991c44e4"} Mar 07 14:55:44 crc kubenswrapper[4943]: I0307 14:55:44.652919 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/keystone-db-sync-nq8hl"] Mar 07 14:55:44 crc kubenswrapper[4943]: W0307 14:55:44.666897 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podea23fd1e_8c76_4483_a624_be13a1715513.slice/crio-eb9c856d03d588c99cf1bf6811e82e3a02fe319f2cdbc13f2ef9d5948f2d4ca7 WatchSource:0}: Error finding container eb9c856d03d588c99cf1bf6811e82e3a02fe319f2cdbc13f2ef9d5948f2d4ca7: Status 404 returned error can't find the container with id eb9c856d03d588c99cf1bf6811e82e3a02fe319f2cdbc13f2ef9d5948f2d4ca7 Mar 07 14:55:45 crc kubenswrapper[4943]: I0307 14:55:45.513489 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/keystone-db-sync-nq8hl" event={"ID":"ea23fd1e-8c76-4483-a624-be13a1715513","Type":"ContainerStarted","Data":"eb9c856d03d588c99cf1bf6811e82e3a02fe319f2cdbc13f2ef9d5948f2d4ca7"} Mar 07 14:55:45 crc kubenswrapper[4943]: I0307 14:55:45.516683 4943 generic.go:334] "Generic (PLEG): container finished" podID="3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1" containerID="89ca8ae9b9a519386b98a327fa8b581a220f9dabd7369fb4987c1c51e710734e" exitCode=0 Mar 07 14:55:45 crc kubenswrapper[4943]: I0307 14:55:45.516731 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar4jd2" event={"ID":"3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1","Type":"ContainerDied","Data":"89ca8ae9b9a519386b98a327fa8b581a220f9dabd7369fb4987c1c51e710734e"} Mar 07 14:55:53 crc kubenswrapper[4943]: I0307 14:55:53.581085 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/keystone-db-sync-nq8hl" event={"ID":"ea23fd1e-8c76-4483-a624-be13a1715513","Type":"ContainerStarted","Data":"0f58432047df35f921d354c24be9618c04c8812420d1943928b1bb6ef760be7a"} Mar 07 14:55:53 crc kubenswrapper[4943]: I0307 14:55:53.583263 4943 generic.go:334] "Generic (PLEG): container finished" podID="3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1" containerID="9660122aed8f49fe2ed08f11a2c8bf842794c5feb6e9e0003733d0bbfbbf7c25" exitCode=0 Mar 07 14:55:53 crc kubenswrapper[4943]: I0307 14:55:53.583303 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar4jd2" event={"ID":"3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1","Type":"ContainerDied","Data":"9660122aed8f49fe2ed08f11a2c8bf842794c5feb6e9e0003733d0bbfbbf7c25"} Mar 07 14:55:53 crc kubenswrapper[4943]: I0307 14:55:53.616187 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/keystone-db-sync-nq8hl" podStartSLOduration=1.813988464 podStartE2EDuration="9.616159564s" podCreationTimestamp="2026-03-07 14:55:44 +0000 UTC" firstStartedPulling="2026-03-07 14:55:44.671395008 +0000 UTC m=+986.623531506" lastFinishedPulling="2026-03-07 14:55:52.473566108 +0000 UTC m=+994.425702606" observedRunningTime="2026-03-07 14:55:53.602206281 +0000 UTC m=+995.554342809" watchObservedRunningTime="2026-03-07 14:55:53.616159564 +0000 UTC m=+995.568296102" Mar 07 14:55:54 crc kubenswrapper[4943]: I0307 14:55:54.594482 4943 generic.go:334] "Generic (PLEG): container finished" podID="3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1" containerID="3301e9ebadd8288fea29b15b63c11bc3252df2a3fbd5bf963ccfd18d072af883" exitCode=0 Mar 07 14:55:54 crc kubenswrapper[4943]: I0307 14:55:54.594589 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar4jd2" event={"ID":"3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1","Type":"ContainerDied","Data":"3301e9ebadd8288fea29b15b63c11bc3252df2a3fbd5bf963ccfd18d072af883"} Mar 07 14:55:55 crc kubenswrapper[4943]: I0307 14:55:55.907225 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar4jd2" Mar 07 14:55:56 crc kubenswrapper[4943]: I0307 14:55:56.001627 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lmhpv\" (UniqueName: \"kubernetes.io/projected/3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1-kube-api-access-lmhpv\") pod \"3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1\" (UID: \"3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1\") " Mar 07 14:55:56 crc kubenswrapper[4943]: I0307 14:55:56.001738 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1-util\") pod \"3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1\" (UID: \"3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1\") " Mar 07 14:55:56 crc kubenswrapper[4943]: I0307 14:55:56.001758 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1-bundle\") pod \"3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1\" (UID: \"3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1\") " Mar 07 14:55:56 crc kubenswrapper[4943]: I0307 14:55:56.002971 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1-bundle" (OuterVolumeSpecName: "bundle") pod "3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1" (UID: "3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:55:56 crc kubenswrapper[4943]: I0307 14:55:56.008634 4943 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1-bundle\") on node \"crc\" DevicePath \"\"" Mar 07 14:55:56 crc kubenswrapper[4943]: I0307 14:55:56.008936 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1-kube-api-access-lmhpv" (OuterVolumeSpecName: "kube-api-access-lmhpv") pod "3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1" (UID: "3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1"). InnerVolumeSpecName "kube-api-access-lmhpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:55:56 crc kubenswrapper[4943]: I0307 14:55:56.012241 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1-util" (OuterVolumeSpecName: "util") pod "3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1" (UID: "3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:55:56 crc kubenswrapper[4943]: I0307 14:55:56.110341 4943 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1-util\") on node \"crc\" DevicePath \"\"" Mar 07 14:55:56 crc kubenswrapper[4943]: I0307 14:55:56.110378 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lmhpv\" (UniqueName: \"kubernetes.io/projected/3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1-kube-api-access-lmhpv\") on node \"crc\" DevicePath \"\"" Mar 07 14:55:56 crc kubenswrapper[4943]: I0307 14:55:56.610791 4943 generic.go:334] "Generic (PLEG): container finished" podID="ea23fd1e-8c76-4483-a624-be13a1715513" containerID="0f58432047df35f921d354c24be9618c04c8812420d1943928b1bb6ef760be7a" exitCode=0 Mar 07 14:55:56 crc kubenswrapper[4943]: I0307 14:55:56.610880 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/keystone-db-sync-nq8hl" event={"ID":"ea23fd1e-8c76-4483-a624-be13a1715513","Type":"ContainerDied","Data":"0f58432047df35f921d354c24be9618c04c8812420d1943928b1bb6ef760be7a"} Mar 07 14:55:56 crc kubenswrapper[4943]: I0307 14:55:56.612904 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar4jd2" event={"ID":"3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1","Type":"ContainerDied","Data":"2a2beb0c51c482052788860fce14c2b983fd5f44ca3d13d21c1bdf0d991c44e4"} Mar 07 14:55:56 crc kubenswrapper[4943]: I0307 14:55:56.612967 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2a2beb0c51c482052788860fce14c2b983fd5f44ca3d13d21c1bdf0d991c44e4" Mar 07 14:55:56 crc kubenswrapper[4943]: I0307 14:55:56.612990 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar4jd2" Mar 07 14:55:58 crc kubenswrapper[4943]: I0307 14:55:58.331657 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/keystone-db-sync-nq8hl" Mar 07 14:55:58 crc kubenswrapper[4943]: I0307 14:55:58.442198 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j7q5c\" (UniqueName: \"kubernetes.io/projected/ea23fd1e-8c76-4483-a624-be13a1715513-kube-api-access-j7q5c\") pod \"ea23fd1e-8c76-4483-a624-be13a1715513\" (UID: \"ea23fd1e-8c76-4483-a624-be13a1715513\") " Mar 07 14:55:58 crc kubenswrapper[4943]: I0307 14:55:58.442381 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea23fd1e-8c76-4483-a624-be13a1715513-config-data\") pod \"ea23fd1e-8c76-4483-a624-be13a1715513\" (UID: \"ea23fd1e-8c76-4483-a624-be13a1715513\") " Mar 07 14:55:58 crc kubenswrapper[4943]: I0307 14:55:58.454187 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea23fd1e-8c76-4483-a624-be13a1715513-kube-api-access-j7q5c" (OuterVolumeSpecName: "kube-api-access-j7q5c") pod "ea23fd1e-8c76-4483-a624-be13a1715513" (UID: "ea23fd1e-8c76-4483-a624-be13a1715513"). InnerVolumeSpecName "kube-api-access-j7q5c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:55:58 crc kubenswrapper[4943]: I0307 14:55:58.493236 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea23fd1e-8c76-4483-a624-be13a1715513-config-data" (OuterVolumeSpecName: "config-data") pod "ea23fd1e-8c76-4483-a624-be13a1715513" (UID: "ea23fd1e-8c76-4483-a624-be13a1715513"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:55:58 crc kubenswrapper[4943]: I0307 14:55:58.545840 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j7q5c\" (UniqueName: \"kubernetes.io/projected/ea23fd1e-8c76-4483-a624-be13a1715513-kube-api-access-j7q5c\") on node \"crc\" DevicePath \"\"" Mar 07 14:55:58 crc kubenswrapper[4943]: I0307 14:55:58.545908 4943 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea23fd1e-8c76-4483-a624-be13a1715513-config-data\") on node \"crc\" DevicePath \"\"" Mar 07 14:55:58 crc kubenswrapper[4943]: I0307 14:55:58.632807 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/keystone-db-sync-nq8hl" event={"ID":"ea23fd1e-8c76-4483-a624-be13a1715513","Type":"ContainerDied","Data":"eb9c856d03d588c99cf1bf6811e82e3a02fe319f2cdbc13f2ef9d5948f2d4ca7"} Mar 07 14:55:58 crc kubenswrapper[4943]: I0307 14:55:58.632854 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eb9c856d03d588c99cf1bf6811e82e3a02fe319f2cdbc13f2ef9d5948f2d4ca7" Mar 07 14:55:58 crc kubenswrapper[4943]: I0307 14:55:58.632950 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/keystone-db-sync-nq8hl" Mar 07 14:55:58 crc kubenswrapper[4943]: I0307 14:55:58.830853 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/keystone-bootstrap-fz942"] Mar 07 14:55:58 crc kubenswrapper[4943]: E0307 14:55:58.831493 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1" containerName="pull" Mar 07 14:55:58 crc kubenswrapper[4943]: I0307 14:55:58.831514 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1" containerName="pull" Mar 07 14:55:58 crc kubenswrapper[4943]: E0307 14:55:58.831541 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea23fd1e-8c76-4483-a624-be13a1715513" containerName="keystone-db-sync" Mar 07 14:55:58 crc kubenswrapper[4943]: I0307 14:55:58.831550 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea23fd1e-8c76-4483-a624-be13a1715513" containerName="keystone-db-sync" Mar 07 14:55:58 crc kubenswrapper[4943]: E0307 14:55:58.831564 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1" containerName="extract" Mar 07 14:55:58 crc kubenswrapper[4943]: I0307 14:55:58.831573 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1" containerName="extract" Mar 07 14:55:58 crc kubenswrapper[4943]: E0307 14:55:58.831594 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1" containerName="util" Mar 07 14:55:58 crc kubenswrapper[4943]: I0307 14:55:58.831603 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1" containerName="util" Mar 07 14:55:58 crc kubenswrapper[4943]: I0307 14:55:58.831748 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1" containerName="extract" Mar 07 14:55:58 crc kubenswrapper[4943]: I0307 14:55:58.831765 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea23fd1e-8c76-4483-a624-be13a1715513" containerName="keystone-db-sync" Mar 07 14:55:58 crc kubenswrapper[4943]: I0307 14:55:58.832293 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/keystone-bootstrap-fz942" Mar 07 14:55:58 crc kubenswrapper[4943]: I0307 14:55:58.837798 4943 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"keystone" Mar 07 14:55:58 crc kubenswrapper[4943]: I0307 14:55:58.837901 4943 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"osp-secret" Mar 07 14:55:58 crc kubenswrapper[4943]: I0307 14:55:58.838001 4943 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"keystone-config-data" Mar 07 14:55:58 crc kubenswrapper[4943]: I0307 14:55:58.838141 4943 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"keystone-scripts" Mar 07 14:55:58 crc kubenswrapper[4943]: I0307 14:55:58.838305 4943 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"keystone-keystone-dockercfg-vk5hc" Mar 07 14:55:58 crc kubenswrapper[4943]: I0307 14:55:58.850251 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2bbf9dc7-77bd-470f-8f33-3b62d014c239-credential-keys\") pod \"keystone-bootstrap-fz942\" (UID: \"2bbf9dc7-77bd-470f-8f33-3b62d014c239\") " pod="swift-kuttl-tests/keystone-bootstrap-fz942" Mar 07 14:55:58 crc kubenswrapper[4943]: I0307 14:55:58.850291 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrs29\" (UniqueName: \"kubernetes.io/projected/2bbf9dc7-77bd-470f-8f33-3b62d014c239-kube-api-access-nrs29\") pod \"keystone-bootstrap-fz942\" (UID: \"2bbf9dc7-77bd-470f-8f33-3b62d014c239\") " pod="swift-kuttl-tests/keystone-bootstrap-fz942" Mar 07 14:55:58 crc kubenswrapper[4943]: I0307 14:55:58.850383 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bbf9dc7-77bd-470f-8f33-3b62d014c239-scripts\") pod \"keystone-bootstrap-fz942\" (UID: \"2bbf9dc7-77bd-470f-8f33-3b62d014c239\") " pod="swift-kuttl-tests/keystone-bootstrap-fz942" Mar 07 14:55:58 crc kubenswrapper[4943]: I0307 14:55:58.850425 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2bbf9dc7-77bd-470f-8f33-3b62d014c239-fernet-keys\") pod \"keystone-bootstrap-fz942\" (UID: \"2bbf9dc7-77bd-470f-8f33-3b62d014c239\") " pod="swift-kuttl-tests/keystone-bootstrap-fz942" Mar 07 14:55:58 crc kubenswrapper[4943]: I0307 14:55:58.850450 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bbf9dc7-77bd-470f-8f33-3b62d014c239-config-data\") pod \"keystone-bootstrap-fz942\" (UID: \"2bbf9dc7-77bd-470f-8f33-3b62d014c239\") " pod="swift-kuttl-tests/keystone-bootstrap-fz942" Mar 07 14:55:58 crc kubenswrapper[4943]: I0307 14:55:58.854643 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/keystone-bootstrap-fz942"] Mar 07 14:55:58 crc kubenswrapper[4943]: I0307 14:55:58.951239 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2bbf9dc7-77bd-470f-8f33-3b62d014c239-fernet-keys\") pod \"keystone-bootstrap-fz942\" (UID: \"2bbf9dc7-77bd-470f-8f33-3b62d014c239\") " pod="swift-kuttl-tests/keystone-bootstrap-fz942" Mar 07 14:55:58 crc kubenswrapper[4943]: I0307 14:55:58.951505 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bbf9dc7-77bd-470f-8f33-3b62d014c239-config-data\") pod \"keystone-bootstrap-fz942\" (UID: \"2bbf9dc7-77bd-470f-8f33-3b62d014c239\") " pod="swift-kuttl-tests/keystone-bootstrap-fz942" Mar 07 14:55:58 crc kubenswrapper[4943]: I0307 14:55:58.951587 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2bbf9dc7-77bd-470f-8f33-3b62d014c239-credential-keys\") pod \"keystone-bootstrap-fz942\" (UID: \"2bbf9dc7-77bd-470f-8f33-3b62d014c239\") " pod="swift-kuttl-tests/keystone-bootstrap-fz942" Mar 07 14:55:58 crc kubenswrapper[4943]: I0307 14:55:58.951622 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrs29\" (UniqueName: \"kubernetes.io/projected/2bbf9dc7-77bd-470f-8f33-3b62d014c239-kube-api-access-nrs29\") pod \"keystone-bootstrap-fz942\" (UID: \"2bbf9dc7-77bd-470f-8f33-3b62d014c239\") " pod="swift-kuttl-tests/keystone-bootstrap-fz942" Mar 07 14:55:58 crc kubenswrapper[4943]: I0307 14:55:58.951760 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bbf9dc7-77bd-470f-8f33-3b62d014c239-scripts\") pod \"keystone-bootstrap-fz942\" (UID: \"2bbf9dc7-77bd-470f-8f33-3b62d014c239\") " pod="swift-kuttl-tests/keystone-bootstrap-fz942" Mar 07 14:55:58 crc kubenswrapper[4943]: I0307 14:55:58.956151 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bbf9dc7-77bd-470f-8f33-3b62d014c239-scripts\") pod \"keystone-bootstrap-fz942\" (UID: \"2bbf9dc7-77bd-470f-8f33-3b62d014c239\") " pod="swift-kuttl-tests/keystone-bootstrap-fz942" Mar 07 14:55:58 crc kubenswrapper[4943]: I0307 14:55:58.956583 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bbf9dc7-77bd-470f-8f33-3b62d014c239-config-data\") pod \"keystone-bootstrap-fz942\" (UID: \"2bbf9dc7-77bd-470f-8f33-3b62d014c239\") " pod="swift-kuttl-tests/keystone-bootstrap-fz942" Mar 07 14:55:58 crc kubenswrapper[4943]: I0307 14:55:58.956704 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2bbf9dc7-77bd-470f-8f33-3b62d014c239-credential-keys\") pod \"keystone-bootstrap-fz942\" (UID: \"2bbf9dc7-77bd-470f-8f33-3b62d014c239\") " pod="swift-kuttl-tests/keystone-bootstrap-fz942" Mar 07 14:55:58 crc kubenswrapper[4943]: I0307 14:55:58.956861 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2bbf9dc7-77bd-470f-8f33-3b62d014c239-fernet-keys\") pod \"keystone-bootstrap-fz942\" (UID: \"2bbf9dc7-77bd-470f-8f33-3b62d014c239\") " pod="swift-kuttl-tests/keystone-bootstrap-fz942" Mar 07 14:55:58 crc kubenswrapper[4943]: I0307 14:55:58.979728 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrs29\" (UniqueName: \"kubernetes.io/projected/2bbf9dc7-77bd-470f-8f33-3b62d014c239-kube-api-access-nrs29\") pod \"keystone-bootstrap-fz942\" (UID: \"2bbf9dc7-77bd-470f-8f33-3b62d014c239\") " pod="swift-kuttl-tests/keystone-bootstrap-fz942" Mar 07 14:55:59 crc kubenswrapper[4943]: I0307 14:55:59.155450 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/keystone-bootstrap-fz942" Mar 07 14:55:59 crc kubenswrapper[4943]: I0307 14:55:59.774213 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/keystone-bootstrap-fz942"] Mar 07 14:56:00 crc kubenswrapper[4943]: I0307 14:56:00.123246 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29548256-q9db4"] Mar 07 14:56:00 crc kubenswrapper[4943]: I0307 14:56:00.124598 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548256-q9db4" Mar 07 14:56:00 crc kubenswrapper[4943]: I0307 14:56:00.127004 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 07 14:56:00 crc kubenswrapper[4943]: I0307 14:56:00.127204 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 07 14:56:00 crc kubenswrapper[4943]: I0307 14:56:00.128053 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-qqlk5" Mar 07 14:56:00 crc kubenswrapper[4943]: I0307 14:56:00.133594 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29548256-q9db4"] Mar 07 14:56:00 crc kubenswrapper[4943]: I0307 14:56:00.280776 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99gq6\" (UniqueName: \"kubernetes.io/projected/8a5f15db-2baf-441d-b3c7-d93e0779d268-kube-api-access-99gq6\") pod \"auto-csr-approver-29548256-q9db4\" (UID: \"8a5f15db-2baf-441d-b3c7-d93e0779d268\") " pod="openshift-infra/auto-csr-approver-29548256-q9db4" Mar 07 14:56:00 crc kubenswrapper[4943]: I0307 14:56:00.381640 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99gq6\" (UniqueName: \"kubernetes.io/projected/8a5f15db-2baf-441d-b3c7-d93e0779d268-kube-api-access-99gq6\") pod \"auto-csr-approver-29548256-q9db4\" (UID: \"8a5f15db-2baf-441d-b3c7-d93e0779d268\") " pod="openshift-infra/auto-csr-approver-29548256-q9db4" Mar 07 14:56:00 crc kubenswrapper[4943]: I0307 14:56:00.415633 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99gq6\" (UniqueName: \"kubernetes.io/projected/8a5f15db-2baf-441d-b3c7-d93e0779d268-kube-api-access-99gq6\") pod \"auto-csr-approver-29548256-q9db4\" (UID: \"8a5f15db-2baf-441d-b3c7-d93e0779d268\") " pod="openshift-infra/auto-csr-approver-29548256-q9db4" Mar 07 14:56:00 crc kubenswrapper[4943]: I0307 14:56:00.447556 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548256-q9db4" Mar 07 14:56:00 crc kubenswrapper[4943]: I0307 14:56:00.647171 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/keystone-bootstrap-fz942" event={"ID":"2bbf9dc7-77bd-470f-8f33-3b62d014c239","Type":"ContainerStarted","Data":"2d4a24d7450c798d91a834f2b32ef0ddd47a854c5e4a00328222985152149372"} Mar 07 14:56:00 crc kubenswrapper[4943]: I0307 14:56:00.647647 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/keystone-bootstrap-fz942" event={"ID":"2bbf9dc7-77bd-470f-8f33-3b62d014c239","Type":"ContainerStarted","Data":"482440c100d261caf68e5e97a2f30ce51e4e2d1f22ce58d5c1681b85cb2c1a2a"} Mar 07 14:56:00 crc kubenswrapper[4943]: I0307 14:56:00.666495 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/keystone-bootstrap-fz942" podStartSLOduration=2.666465256 podStartE2EDuration="2.666465256s" podCreationTimestamp="2026-03-07 14:55:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:56:00.661442852 +0000 UTC m=+1002.613579390" watchObservedRunningTime="2026-03-07 14:56:00.666465256 +0000 UTC m=+1002.618601794" Mar 07 14:56:00 crc kubenswrapper[4943]: I0307 14:56:00.830587 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29548256-q9db4"] Mar 07 14:56:01 crc kubenswrapper[4943]: I0307 14:56:01.657191 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548256-q9db4" event={"ID":"8a5f15db-2baf-441d-b3c7-d93e0779d268","Type":"ContainerStarted","Data":"ef7616fb3ccbc2ed6597b83d0c9fff8cb621fd34ea7d7bde0c34ac37fbe09366"} Mar 07 14:56:01 crc kubenswrapper[4943]: I0307 14:56:01.884153 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-whqgv"] Mar 07 14:56:01 crc kubenswrapper[4943]: I0307 14:56:01.885795 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-whqgv" Mar 07 14:56:01 crc kubenswrapper[4943]: I0307 14:56:01.901149 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-whqgv"] Mar 07 14:56:01 crc kubenswrapper[4943]: I0307 14:56:01.938844 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ea7d2c3-7a0f-48dd-826f-5b4806612e66-catalog-content\") pod \"redhat-operators-whqgv\" (UID: \"8ea7d2c3-7a0f-48dd-826f-5b4806612e66\") " pod="openshift-marketplace/redhat-operators-whqgv" Mar 07 14:56:01 crc kubenswrapper[4943]: I0307 14:56:01.938990 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbv6t\" (UniqueName: \"kubernetes.io/projected/8ea7d2c3-7a0f-48dd-826f-5b4806612e66-kube-api-access-wbv6t\") pod \"redhat-operators-whqgv\" (UID: \"8ea7d2c3-7a0f-48dd-826f-5b4806612e66\") " pod="openshift-marketplace/redhat-operators-whqgv" Mar 07 14:56:01 crc kubenswrapper[4943]: I0307 14:56:01.939108 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ea7d2c3-7a0f-48dd-826f-5b4806612e66-utilities\") pod \"redhat-operators-whqgv\" (UID: \"8ea7d2c3-7a0f-48dd-826f-5b4806612e66\") " pod="openshift-marketplace/redhat-operators-whqgv" Mar 07 14:56:02 crc kubenswrapper[4943]: I0307 14:56:02.040410 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ea7d2c3-7a0f-48dd-826f-5b4806612e66-catalog-content\") pod \"redhat-operators-whqgv\" (UID: \"8ea7d2c3-7a0f-48dd-826f-5b4806612e66\") " pod="openshift-marketplace/redhat-operators-whqgv" Mar 07 14:56:02 crc kubenswrapper[4943]: I0307 14:56:02.040501 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbv6t\" (UniqueName: \"kubernetes.io/projected/8ea7d2c3-7a0f-48dd-826f-5b4806612e66-kube-api-access-wbv6t\") pod \"redhat-operators-whqgv\" (UID: \"8ea7d2c3-7a0f-48dd-826f-5b4806612e66\") " pod="openshift-marketplace/redhat-operators-whqgv" Mar 07 14:56:02 crc kubenswrapper[4943]: I0307 14:56:02.040545 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ea7d2c3-7a0f-48dd-826f-5b4806612e66-utilities\") pod \"redhat-operators-whqgv\" (UID: \"8ea7d2c3-7a0f-48dd-826f-5b4806612e66\") " pod="openshift-marketplace/redhat-operators-whqgv" Mar 07 14:56:02 crc kubenswrapper[4943]: I0307 14:56:02.041074 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ea7d2c3-7a0f-48dd-826f-5b4806612e66-catalog-content\") pod \"redhat-operators-whqgv\" (UID: \"8ea7d2c3-7a0f-48dd-826f-5b4806612e66\") " pod="openshift-marketplace/redhat-operators-whqgv" Mar 07 14:56:02 crc kubenswrapper[4943]: I0307 14:56:02.041109 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ea7d2c3-7a0f-48dd-826f-5b4806612e66-utilities\") pod \"redhat-operators-whqgv\" (UID: \"8ea7d2c3-7a0f-48dd-826f-5b4806612e66\") " pod="openshift-marketplace/redhat-operators-whqgv" Mar 07 14:56:02 crc kubenswrapper[4943]: I0307 14:56:02.087301 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbv6t\" (UniqueName: \"kubernetes.io/projected/8ea7d2c3-7a0f-48dd-826f-5b4806612e66-kube-api-access-wbv6t\") pod \"redhat-operators-whqgv\" (UID: \"8ea7d2c3-7a0f-48dd-826f-5b4806612e66\") " pod="openshift-marketplace/redhat-operators-whqgv" Mar 07 14:56:02 crc kubenswrapper[4943]: I0307 14:56:02.265751 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-whqgv" Mar 07 14:56:02 crc kubenswrapper[4943]: I0307 14:56:02.665145 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548256-q9db4" event={"ID":"8a5f15db-2baf-441d-b3c7-d93e0779d268","Type":"ContainerStarted","Data":"950a2ae469245e87ccf033b6bcaab52a5596412a5f98a2ad425c32993ae4aec0"} Mar 07 14:56:02 crc kubenswrapper[4943]: I0307 14:56:02.677520 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29548256-q9db4" podStartSLOduration=1.3722627219999999 podStartE2EDuration="2.677502719s" podCreationTimestamp="2026-03-07 14:56:00 +0000 UTC" firstStartedPulling="2026-03-07 14:56:00.845126789 +0000 UTC m=+1002.797263297" lastFinishedPulling="2026-03-07 14:56:02.150366786 +0000 UTC m=+1004.102503294" observedRunningTime="2026-03-07 14:56:02.677025067 +0000 UTC m=+1004.629161555" watchObservedRunningTime="2026-03-07 14:56:02.677502719 +0000 UTC m=+1004.629639217" Mar 07 14:56:02 crc kubenswrapper[4943]: I0307 14:56:02.845554 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-whqgv"] Mar 07 14:56:02 crc kubenswrapper[4943]: W0307 14:56:02.857036 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8ea7d2c3_7a0f_48dd_826f_5b4806612e66.slice/crio-cbd345bd6691301795a78042ca1943a3745dd7faef6fcb6dd1b14d1d37552603 WatchSource:0}: Error finding container cbd345bd6691301795a78042ca1943a3745dd7faef6fcb6dd1b14d1d37552603: Status 404 returned error can't find the container with id cbd345bd6691301795a78042ca1943a3745dd7faef6fcb6dd1b14d1d37552603 Mar 07 14:56:03 crc kubenswrapper[4943]: E0307 14:56:03.376167 4943 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2bbf9dc7_77bd_470f_8f33_3b62d014c239.slice/crio-conmon-2d4a24d7450c798d91a834f2b32ef0ddd47a854c5e4a00328222985152149372.scope\": RecentStats: unable to find data in memory cache]" Mar 07 14:56:03 crc kubenswrapper[4943]: I0307 14:56:03.679957 4943 generic.go:334] "Generic (PLEG): container finished" podID="2bbf9dc7-77bd-470f-8f33-3b62d014c239" containerID="2d4a24d7450c798d91a834f2b32ef0ddd47a854c5e4a00328222985152149372" exitCode=0 Mar 07 14:56:03 crc kubenswrapper[4943]: I0307 14:56:03.680223 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/keystone-bootstrap-fz942" event={"ID":"2bbf9dc7-77bd-470f-8f33-3b62d014c239","Type":"ContainerDied","Data":"2d4a24d7450c798d91a834f2b32ef0ddd47a854c5e4a00328222985152149372"} Mar 07 14:56:03 crc kubenswrapper[4943]: I0307 14:56:03.686300 4943 generic.go:334] "Generic (PLEG): container finished" podID="8a5f15db-2baf-441d-b3c7-d93e0779d268" containerID="950a2ae469245e87ccf033b6bcaab52a5596412a5f98a2ad425c32993ae4aec0" exitCode=0 Mar 07 14:56:03 crc kubenswrapper[4943]: I0307 14:56:03.686360 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548256-q9db4" event={"ID":"8a5f15db-2baf-441d-b3c7-d93e0779d268","Type":"ContainerDied","Data":"950a2ae469245e87ccf033b6bcaab52a5596412a5f98a2ad425c32993ae4aec0"} Mar 07 14:56:03 crc kubenswrapper[4943]: I0307 14:56:03.687921 4943 generic.go:334] "Generic (PLEG): container finished" podID="8ea7d2c3-7a0f-48dd-826f-5b4806612e66" containerID="3b8dd0f0cf7529b2c288ce9dd6453698eefb85758f94635f6e781656f18bee4a" exitCode=0 Mar 07 14:56:03 crc kubenswrapper[4943]: I0307 14:56:03.687976 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-whqgv" event={"ID":"8ea7d2c3-7a0f-48dd-826f-5b4806612e66","Type":"ContainerDied","Data":"3b8dd0f0cf7529b2c288ce9dd6453698eefb85758f94635f6e781656f18bee4a"} Mar 07 14:56:03 crc kubenswrapper[4943]: I0307 14:56:03.688006 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-whqgv" event={"ID":"8ea7d2c3-7a0f-48dd-826f-5b4806612e66","Type":"ContainerStarted","Data":"cbd345bd6691301795a78042ca1943a3745dd7faef6fcb6dd1b14d1d37552603"} Mar 07 14:56:04 crc kubenswrapper[4943]: I0307 14:56:04.698191 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-whqgv" event={"ID":"8ea7d2c3-7a0f-48dd-826f-5b4806612e66","Type":"ContainerStarted","Data":"457d722faf5a6dba954ea95543a456183feba10955952d148d3b54a614fcd140"} Mar 07 14:56:05 crc kubenswrapper[4943]: I0307 14:56:05.057461 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548256-q9db4" Mar 07 14:56:05 crc kubenswrapper[4943]: I0307 14:56:05.062909 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/keystone-bootstrap-fz942" Mar 07 14:56:05 crc kubenswrapper[4943]: I0307 14:56:05.135520 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nrs29\" (UniqueName: \"kubernetes.io/projected/2bbf9dc7-77bd-470f-8f33-3b62d014c239-kube-api-access-nrs29\") pod \"2bbf9dc7-77bd-470f-8f33-3b62d014c239\" (UID: \"2bbf9dc7-77bd-470f-8f33-3b62d014c239\") " Mar 07 14:56:05 crc kubenswrapper[4943]: I0307 14:56:05.135590 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bbf9dc7-77bd-470f-8f33-3b62d014c239-config-data\") pod \"2bbf9dc7-77bd-470f-8f33-3b62d014c239\" (UID: \"2bbf9dc7-77bd-470f-8f33-3b62d014c239\") " Mar 07 14:56:05 crc kubenswrapper[4943]: I0307 14:56:05.135692 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2bbf9dc7-77bd-470f-8f33-3b62d014c239-fernet-keys\") pod \"2bbf9dc7-77bd-470f-8f33-3b62d014c239\" (UID: \"2bbf9dc7-77bd-470f-8f33-3b62d014c239\") " Mar 07 14:56:05 crc kubenswrapper[4943]: I0307 14:56:05.135740 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2bbf9dc7-77bd-470f-8f33-3b62d014c239-credential-keys\") pod \"2bbf9dc7-77bd-470f-8f33-3b62d014c239\" (UID: \"2bbf9dc7-77bd-470f-8f33-3b62d014c239\") " Mar 07 14:56:05 crc kubenswrapper[4943]: I0307 14:56:05.135799 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-99gq6\" (UniqueName: \"kubernetes.io/projected/8a5f15db-2baf-441d-b3c7-d93e0779d268-kube-api-access-99gq6\") pod \"8a5f15db-2baf-441d-b3c7-d93e0779d268\" (UID: \"8a5f15db-2baf-441d-b3c7-d93e0779d268\") " Mar 07 14:56:05 crc kubenswrapper[4943]: I0307 14:56:05.135828 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bbf9dc7-77bd-470f-8f33-3b62d014c239-scripts\") pod \"2bbf9dc7-77bd-470f-8f33-3b62d014c239\" (UID: \"2bbf9dc7-77bd-470f-8f33-3b62d014c239\") " Mar 07 14:56:05 crc kubenswrapper[4943]: I0307 14:56:05.142116 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bbf9dc7-77bd-470f-8f33-3b62d014c239-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "2bbf9dc7-77bd-470f-8f33-3b62d014c239" (UID: "2bbf9dc7-77bd-470f-8f33-3b62d014c239"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:56:05 crc kubenswrapper[4943]: I0307 14:56:05.142302 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bbf9dc7-77bd-470f-8f33-3b62d014c239-scripts" (OuterVolumeSpecName: "scripts") pod "2bbf9dc7-77bd-470f-8f33-3b62d014c239" (UID: "2bbf9dc7-77bd-470f-8f33-3b62d014c239"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:56:05 crc kubenswrapper[4943]: I0307 14:56:05.142698 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bbf9dc7-77bd-470f-8f33-3b62d014c239-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "2bbf9dc7-77bd-470f-8f33-3b62d014c239" (UID: "2bbf9dc7-77bd-470f-8f33-3b62d014c239"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:56:05 crc kubenswrapper[4943]: I0307 14:56:05.143231 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a5f15db-2baf-441d-b3c7-d93e0779d268-kube-api-access-99gq6" (OuterVolumeSpecName: "kube-api-access-99gq6") pod "8a5f15db-2baf-441d-b3c7-d93e0779d268" (UID: "8a5f15db-2baf-441d-b3c7-d93e0779d268"). InnerVolumeSpecName "kube-api-access-99gq6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:56:05 crc kubenswrapper[4943]: I0307 14:56:05.143264 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bbf9dc7-77bd-470f-8f33-3b62d014c239-kube-api-access-nrs29" (OuterVolumeSpecName: "kube-api-access-nrs29") pod "2bbf9dc7-77bd-470f-8f33-3b62d014c239" (UID: "2bbf9dc7-77bd-470f-8f33-3b62d014c239"). InnerVolumeSpecName "kube-api-access-nrs29". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:56:05 crc kubenswrapper[4943]: I0307 14:56:05.158457 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bbf9dc7-77bd-470f-8f33-3b62d014c239-config-data" (OuterVolumeSpecName: "config-data") pod "2bbf9dc7-77bd-470f-8f33-3b62d014c239" (UID: "2bbf9dc7-77bd-470f-8f33-3b62d014c239"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:56:05 crc kubenswrapper[4943]: I0307 14:56:05.243301 4943 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2bbf9dc7-77bd-470f-8f33-3b62d014c239-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 07 14:56:05 crc kubenswrapper[4943]: I0307 14:56:05.243343 4943 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2bbf9dc7-77bd-470f-8f33-3b62d014c239-credential-keys\") on node \"crc\" DevicePath \"\"" Mar 07 14:56:05 crc kubenswrapper[4943]: I0307 14:56:05.243361 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-99gq6\" (UniqueName: \"kubernetes.io/projected/8a5f15db-2baf-441d-b3c7-d93e0779d268-kube-api-access-99gq6\") on node \"crc\" DevicePath \"\"" Mar 07 14:56:05 crc kubenswrapper[4943]: I0307 14:56:05.243376 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bbf9dc7-77bd-470f-8f33-3b62d014c239-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 14:56:05 crc kubenswrapper[4943]: I0307 14:56:05.243389 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nrs29\" (UniqueName: \"kubernetes.io/projected/2bbf9dc7-77bd-470f-8f33-3b62d014c239-kube-api-access-nrs29\") on node \"crc\" DevicePath \"\"" Mar 07 14:56:05 crc kubenswrapper[4943]: I0307 14:56:05.243400 4943 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bbf9dc7-77bd-470f-8f33-3b62d014c239-config-data\") on node \"crc\" DevicePath \"\"" Mar 07 14:56:05 crc kubenswrapper[4943]: I0307 14:56:05.705665 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548256-q9db4" Mar 07 14:56:05 crc kubenswrapper[4943]: I0307 14:56:05.705660 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548256-q9db4" event={"ID":"8a5f15db-2baf-441d-b3c7-d93e0779d268","Type":"ContainerDied","Data":"ef7616fb3ccbc2ed6597b83d0c9fff8cb621fd34ea7d7bde0c34ac37fbe09366"} Mar 07 14:56:05 crc kubenswrapper[4943]: I0307 14:56:05.705825 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ef7616fb3ccbc2ed6597b83d0c9fff8cb621fd34ea7d7bde0c34ac37fbe09366" Mar 07 14:56:05 crc kubenswrapper[4943]: I0307 14:56:05.707971 4943 generic.go:334] "Generic (PLEG): container finished" podID="8ea7d2c3-7a0f-48dd-826f-5b4806612e66" containerID="457d722faf5a6dba954ea95543a456183feba10955952d148d3b54a614fcd140" exitCode=0 Mar 07 14:56:05 crc kubenswrapper[4943]: I0307 14:56:05.708030 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-whqgv" event={"ID":"8ea7d2c3-7a0f-48dd-826f-5b4806612e66","Type":"ContainerDied","Data":"457d722faf5a6dba954ea95543a456183feba10955952d148d3b54a614fcd140"} Mar 07 14:56:05 crc kubenswrapper[4943]: I0307 14:56:05.710758 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/keystone-bootstrap-fz942" event={"ID":"2bbf9dc7-77bd-470f-8f33-3b62d014c239","Type":"ContainerDied","Data":"482440c100d261caf68e5e97a2f30ce51e4e2d1f22ce58d5c1681b85cb2c1a2a"} Mar 07 14:56:05 crc kubenswrapper[4943]: I0307 14:56:05.710783 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="482440c100d261caf68e5e97a2f30ce51e4e2d1f22ce58d5c1681b85cb2c1a2a" Mar 07 14:56:05 crc kubenswrapper[4943]: I0307 14:56:05.710842 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/keystone-bootstrap-fz942" Mar 07 14:56:05 crc kubenswrapper[4943]: I0307 14:56:05.751660 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29548250-cgbqg"] Mar 07 14:56:05 crc kubenswrapper[4943]: I0307 14:56:05.760627 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29548250-cgbqg"] Mar 07 14:56:05 crc kubenswrapper[4943]: I0307 14:56:05.814499 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/keystone-5c4bccddcc-5vjzw"] Mar 07 14:56:05 crc kubenswrapper[4943]: E0307 14:56:05.814827 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bbf9dc7-77bd-470f-8f33-3b62d014c239" containerName="keystone-bootstrap" Mar 07 14:56:05 crc kubenswrapper[4943]: I0307 14:56:05.814847 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bbf9dc7-77bd-470f-8f33-3b62d014c239" containerName="keystone-bootstrap" Mar 07 14:56:05 crc kubenswrapper[4943]: E0307 14:56:05.814877 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a5f15db-2baf-441d-b3c7-d93e0779d268" containerName="oc" Mar 07 14:56:05 crc kubenswrapper[4943]: I0307 14:56:05.814885 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a5f15db-2baf-441d-b3c7-d93e0779d268" containerName="oc" Mar 07 14:56:05 crc kubenswrapper[4943]: I0307 14:56:05.815035 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a5f15db-2baf-441d-b3c7-d93e0779d268" containerName="oc" Mar 07 14:56:05 crc kubenswrapper[4943]: I0307 14:56:05.815055 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bbf9dc7-77bd-470f-8f33-3b62d014c239" containerName="keystone-bootstrap" Mar 07 14:56:05 crc kubenswrapper[4943]: I0307 14:56:05.815570 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/keystone-5c4bccddcc-5vjzw" Mar 07 14:56:05 crc kubenswrapper[4943]: I0307 14:56:05.819991 4943 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"keystone-keystone-dockercfg-vk5hc" Mar 07 14:56:05 crc kubenswrapper[4943]: I0307 14:56:05.819995 4943 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"keystone-config-data" Mar 07 14:56:05 crc kubenswrapper[4943]: I0307 14:56:05.820030 4943 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"keystone-scripts" Mar 07 14:56:05 crc kubenswrapper[4943]: I0307 14:56:05.820857 4943 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"keystone" Mar 07 14:56:05 crc kubenswrapper[4943]: I0307 14:56:05.827642 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/keystone-5c4bccddcc-5vjzw"] Mar 07 14:56:05 crc kubenswrapper[4943]: I0307 14:56:05.953572 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4c58fc8e-3b29-49a6-b592-9455d49f89cd-credential-keys\") pod \"keystone-5c4bccddcc-5vjzw\" (UID: \"4c58fc8e-3b29-49a6-b592-9455d49f89cd\") " pod="swift-kuttl-tests/keystone-5c4bccddcc-5vjzw" Mar 07 14:56:05 crc kubenswrapper[4943]: I0307 14:56:05.953993 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4c58fc8e-3b29-49a6-b592-9455d49f89cd-fernet-keys\") pod \"keystone-5c4bccddcc-5vjzw\" (UID: \"4c58fc8e-3b29-49a6-b592-9455d49f89cd\") " pod="swift-kuttl-tests/keystone-5c4bccddcc-5vjzw" Mar 07 14:56:05 crc kubenswrapper[4943]: I0307 14:56:05.954050 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c58fc8e-3b29-49a6-b592-9455d49f89cd-scripts\") pod \"keystone-5c4bccddcc-5vjzw\" (UID: \"4c58fc8e-3b29-49a6-b592-9455d49f89cd\") " pod="swift-kuttl-tests/keystone-5c4bccddcc-5vjzw" Mar 07 14:56:05 crc kubenswrapper[4943]: I0307 14:56:05.954130 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvjdp\" (UniqueName: \"kubernetes.io/projected/4c58fc8e-3b29-49a6-b592-9455d49f89cd-kube-api-access-fvjdp\") pod \"keystone-5c4bccddcc-5vjzw\" (UID: \"4c58fc8e-3b29-49a6-b592-9455d49f89cd\") " pod="swift-kuttl-tests/keystone-5c4bccddcc-5vjzw" Mar 07 14:56:05 crc kubenswrapper[4943]: I0307 14:56:05.954164 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c58fc8e-3b29-49a6-b592-9455d49f89cd-config-data\") pod \"keystone-5c4bccddcc-5vjzw\" (UID: \"4c58fc8e-3b29-49a6-b592-9455d49f89cd\") " pod="swift-kuttl-tests/keystone-5c4bccddcc-5vjzw" Mar 07 14:56:06 crc kubenswrapper[4943]: I0307 14:56:06.055308 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c58fc8e-3b29-49a6-b592-9455d49f89cd-config-data\") pod \"keystone-5c4bccddcc-5vjzw\" (UID: \"4c58fc8e-3b29-49a6-b592-9455d49f89cd\") " pod="swift-kuttl-tests/keystone-5c4bccddcc-5vjzw" Mar 07 14:56:06 crc kubenswrapper[4943]: I0307 14:56:06.055610 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4c58fc8e-3b29-49a6-b592-9455d49f89cd-credential-keys\") pod \"keystone-5c4bccddcc-5vjzw\" (UID: \"4c58fc8e-3b29-49a6-b592-9455d49f89cd\") " pod="swift-kuttl-tests/keystone-5c4bccddcc-5vjzw" Mar 07 14:56:06 crc kubenswrapper[4943]: I0307 14:56:06.055705 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4c58fc8e-3b29-49a6-b592-9455d49f89cd-fernet-keys\") pod \"keystone-5c4bccddcc-5vjzw\" (UID: \"4c58fc8e-3b29-49a6-b592-9455d49f89cd\") " pod="swift-kuttl-tests/keystone-5c4bccddcc-5vjzw" Mar 07 14:56:06 crc kubenswrapper[4943]: I0307 14:56:06.055800 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c58fc8e-3b29-49a6-b592-9455d49f89cd-scripts\") pod \"keystone-5c4bccddcc-5vjzw\" (UID: \"4c58fc8e-3b29-49a6-b592-9455d49f89cd\") " pod="swift-kuttl-tests/keystone-5c4bccddcc-5vjzw" Mar 07 14:56:06 crc kubenswrapper[4943]: I0307 14:56:06.055881 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvjdp\" (UniqueName: \"kubernetes.io/projected/4c58fc8e-3b29-49a6-b592-9455d49f89cd-kube-api-access-fvjdp\") pod \"keystone-5c4bccddcc-5vjzw\" (UID: \"4c58fc8e-3b29-49a6-b592-9455d49f89cd\") " pod="swift-kuttl-tests/keystone-5c4bccddcc-5vjzw" Mar 07 14:56:06 crc kubenswrapper[4943]: I0307 14:56:06.062962 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4c58fc8e-3b29-49a6-b592-9455d49f89cd-fernet-keys\") pod \"keystone-5c4bccddcc-5vjzw\" (UID: \"4c58fc8e-3b29-49a6-b592-9455d49f89cd\") " pod="swift-kuttl-tests/keystone-5c4bccddcc-5vjzw" Mar 07 14:56:06 crc kubenswrapper[4943]: I0307 14:56:06.064580 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c58fc8e-3b29-49a6-b592-9455d49f89cd-config-data\") pod \"keystone-5c4bccddcc-5vjzw\" (UID: \"4c58fc8e-3b29-49a6-b592-9455d49f89cd\") " pod="swift-kuttl-tests/keystone-5c4bccddcc-5vjzw" Mar 07 14:56:06 crc kubenswrapper[4943]: I0307 14:56:06.066789 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4c58fc8e-3b29-49a6-b592-9455d49f89cd-credential-keys\") pod \"keystone-5c4bccddcc-5vjzw\" (UID: \"4c58fc8e-3b29-49a6-b592-9455d49f89cd\") " pod="swift-kuttl-tests/keystone-5c4bccddcc-5vjzw" Mar 07 14:56:06 crc kubenswrapper[4943]: I0307 14:56:06.073242 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c58fc8e-3b29-49a6-b592-9455d49f89cd-scripts\") pod \"keystone-5c4bccddcc-5vjzw\" (UID: \"4c58fc8e-3b29-49a6-b592-9455d49f89cd\") " pod="swift-kuttl-tests/keystone-5c4bccddcc-5vjzw" Mar 07 14:56:06 crc kubenswrapper[4943]: I0307 14:56:06.074038 4943 patch_prober.go:28] interesting pod/machine-config-daemon-cpdmm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 07 14:56:06 crc kubenswrapper[4943]: I0307 14:56:06.074393 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 07 14:56:06 crc kubenswrapper[4943]: I0307 14:56:06.074640 4943 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" Mar 07 14:56:06 crc kubenswrapper[4943]: I0307 14:56:06.076061 4943 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fe852e1f10f4ca72f749789ca1e6ab73542eaedb5349732682970babc3f1afd1"} pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 07 14:56:06 crc kubenswrapper[4943]: I0307 14:56:06.076477 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerName="machine-config-daemon" containerID="cri-o://fe852e1f10f4ca72f749789ca1e6ab73542eaedb5349732682970babc3f1afd1" gracePeriod=600 Mar 07 14:56:06 crc kubenswrapper[4943]: I0307 14:56:06.094008 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvjdp\" (UniqueName: \"kubernetes.io/projected/4c58fc8e-3b29-49a6-b592-9455d49f89cd-kube-api-access-fvjdp\") pod \"keystone-5c4bccddcc-5vjzw\" (UID: \"4c58fc8e-3b29-49a6-b592-9455d49f89cd\") " pod="swift-kuttl-tests/keystone-5c4bccddcc-5vjzw" Mar 07 14:56:06 crc kubenswrapper[4943]: I0307 14:56:06.135774 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/keystone-5c4bccddcc-5vjzw" Mar 07 14:56:06 crc kubenswrapper[4943]: I0307 14:56:06.410707 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/keystone-5c4bccddcc-5vjzw"] Mar 07 14:56:06 crc kubenswrapper[4943]: W0307 14:56:06.417739 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4c58fc8e_3b29_49a6_b592_9455d49f89cd.slice/crio-767a11d0fec43df8ba815934fcb9639252b7fa3022d8b2a15e57eac96f44c1d7 WatchSource:0}: Error finding container 767a11d0fec43df8ba815934fcb9639252b7fa3022d8b2a15e57eac96f44c1d7: Status 404 returned error can't find the container with id 767a11d0fec43df8ba815934fcb9639252b7fa3022d8b2a15e57eac96f44c1d7 Mar 07 14:56:06 crc kubenswrapper[4943]: I0307 14:56:06.720422 4943 generic.go:334] "Generic (PLEG): container finished" podID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerID="fe852e1f10f4ca72f749789ca1e6ab73542eaedb5349732682970babc3f1afd1" exitCode=0 Mar 07 14:56:06 crc kubenswrapper[4943]: I0307 14:56:06.720510 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" event={"ID":"4c1f18ea-e557-41fc-95dd-ce4e1722a61e","Type":"ContainerDied","Data":"fe852e1f10f4ca72f749789ca1e6ab73542eaedb5349732682970babc3f1afd1"} Mar 07 14:56:06 crc kubenswrapper[4943]: I0307 14:56:06.720963 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" event={"ID":"4c1f18ea-e557-41fc-95dd-ce4e1722a61e","Type":"ContainerStarted","Data":"808321a0c8648ed7264aba7412d4317b4e5e5302eaf1e96b5220f23b29527cec"} Mar 07 14:56:06 crc kubenswrapper[4943]: I0307 14:56:06.720997 4943 scope.go:117] "RemoveContainer" containerID="8ffd8d858c956cfb1e523c263c554f1c6d621212d5b2c18a1bc047ffbcf47fee" Mar 07 14:56:06 crc kubenswrapper[4943]: I0307 14:56:06.724894 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-whqgv" event={"ID":"8ea7d2c3-7a0f-48dd-826f-5b4806612e66","Type":"ContainerStarted","Data":"e14311cea59eaafea745941fcaf7dbbaa13565700d7ecc2d05890c621bdb566f"} Mar 07 14:56:06 crc kubenswrapper[4943]: I0307 14:56:06.727097 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/keystone-5c4bccddcc-5vjzw" event={"ID":"4c58fc8e-3b29-49a6-b592-9455d49f89cd","Type":"ContainerStarted","Data":"229533d340da739263161b9a4acf5dc5e66e323d2d582190030347f6dc39d14c"} Mar 07 14:56:06 crc kubenswrapper[4943]: I0307 14:56:06.727140 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/keystone-5c4bccddcc-5vjzw" event={"ID":"4c58fc8e-3b29-49a6-b592-9455d49f89cd","Type":"ContainerStarted","Data":"767a11d0fec43df8ba815934fcb9639252b7fa3022d8b2a15e57eac96f44c1d7"} Mar 07 14:56:06 crc kubenswrapper[4943]: I0307 14:56:06.727229 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="swift-kuttl-tests/keystone-5c4bccddcc-5vjzw" Mar 07 14:56:06 crc kubenswrapper[4943]: I0307 14:56:06.764877 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce0f9eb4-53fa-4a58-b68f-e4765e56739d" path="/var/lib/kubelet/pods/ce0f9eb4-53fa-4a58-b68f-e4765e56739d/volumes" Mar 07 14:56:06 crc kubenswrapper[4943]: I0307 14:56:06.778852 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-whqgv" podStartSLOduration=3.305503432 podStartE2EDuration="5.778833523s" podCreationTimestamp="2026-03-07 14:56:01 +0000 UTC" firstStartedPulling="2026-03-07 14:56:03.690409706 +0000 UTC m=+1005.642546204" lastFinishedPulling="2026-03-07 14:56:06.163739767 +0000 UTC m=+1008.115876295" observedRunningTime="2026-03-07 14:56:06.774130387 +0000 UTC m=+1008.726266905" watchObservedRunningTime="2026-03-07 14:56:06.778833523 +0000 UTC m=+1008.730970041" Mar 07 14:56:06 crc kubenswrapper[4943]: I0307 14:56:06.788221 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/keystone-5c4bccddcc-5vjzw" podStartSLOduration=1.788203414 podStartE2EDuration="1.788203414s" podCreationTimestamp="2026-03-07 14:56:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:56:06.786563803 +0000 UTC m=+1008.738700321" watchObservedRunningTime="2026-03-07 14:56:06.788203414 +0000 UTC m=+1008.740339912" Mar 07 14:56:08 crc kubenswrapper[4943]: I0307 14:56:08.072132 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-675b7b5b56-95zhm"] Mar 07 14:56:08 crc kubenswrapper[4943]: I0307 14:56:08.073611 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-675b7b5b56-95zhm" Mar 07 14:56:08 crc kubenswrapper[4943]: I0307 14:56:08.075348 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-fc9x7" Mar 07 14:56:08 crc kubenswrapper[4943]: I0307 14:56:08.075723 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-service-cert" Mar 07 14:56:08 crc kubenswrapper[4943]: I0307 14:56:08.085500 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-675b7b5b56-95zhm"] Mar 07 14:56:08 crc kubenswrapper[4943]: I0307 14:56:08.188534 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ebf90d0c-1550-42d8-96b6-695f161d5ace-apiservice-cert\") pod \"barbican-operator-controller-manager-675b7b5b56-95zhm\" (UID: \"ebf90d0c-1550-42d8-96b6-695f161d5ace\") " pod="openstack-operators/barbican-operator-controller-manager-675b7b5b56-95zhm" Mar 07 14:56:08 crc kubenswrapper[4943]: I0307 14:56:08.189028 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ebf90d0c-1550-42d8-96b6-695f161d5ace-webhook-cert\") pod \"barbican-operator-controller-manager-675b7b5b56-95zhm\" (UID: \"ebf90d0c-1550-42d8-96b6-695f161d5ace\") " pod="openstack-operators/barbican-operator-controller-manager-675b7b5b56-95zhm" Mar 07 14:56:08 crc kubenswrapper[4943]: I0307 14:56:08.189073 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhphs\" (UniqueName: \"kubernetes.io/projected/ebf90d0c-1550-42d8-96b6-695f161d5ace-kube-api-access-rhphs\") pod \"barbican-operator-controller-manager-675b7b5b56-95zhm\" (UID: \"ebf90d0c-1550-42d8-96b6-695f161d5ace\") " pod="openstack-operators/barbican-operator-controller-manager-675b7b5b56-95zhm" Mar 07 14:56:08 crc kubenswrapper[4943]: I0307 14:56:08.290384 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ebf90d0c-1550-42d8-96b6-695f161d5ace-apiservice-cert\") pod \"barbican-operator-controller-manager-675b7b5b56-95zhm\" (UID: \"ebf90d0c-1550-42d8-96b6-695f161d5ace\") " pod="openstack-operators/barbican-operator-controller-manager-675b7b5b56-95zhm" Mar 07 14:56:08 crc kubenswrapper[4943]: I0307 14:56:08.290562 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ebf90d0c-1550-42d8-96b6-695f161d5ace-webhook-cert\") pod \"barbican-operator-controller-manager-675b7b5b56-95zhm\" (UID: \"ebf90d0c-1550-42d8-96b6-695f161d5ace\") " pod="openstack-operators/barbican-operator-controller-manager-675b7b5b56-95zhm" Mar 07 14:56:08 crc kubenswrapper[4943]: I0307 14:56:08.290652 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhphs\" (UniqueName: \"kubernetes.io/projected/ebf90d0c-1550-42d8-96b6-695f161d5ace-kube-api-access-rhphs\") pod \"barbican-operator-controller-manager-675b7b5b56-95zhm\" (UID: \"ebf90d0c-1550-42d8-96b6-695f161d5ace\") " pod="openstack-operators/barbican-operator-controller-manager-675b7b5b56-95zhm" Mar 07 14:56:08 crc kubenswrapper[4943]: I0307 14:56:08.297723 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ebf90d0c-1550-42d8-96b6-695f161d5ace-apiservice-cert\") pod \"barbican-operator-controller-manager-675b7b5b56-95zhm\" (UID: \"ebf90d0c-1550-42d8-96b6-695f161d5ace\") " pod="openstack-operators/barbican-operator-controller-manager-675b7b5b56-95zhm" Mar 07 14:56:08 crc kubenswrapper[4943]: I0307 14:56:08.302713 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ebf90d0c-1550-42d8-96b6-695f161d5ace-webhook-cert\") pod \"barbican-operator-controller-manager-675b7b5b56-95zhm\" (UID: \"ebf90d0c-1550-42d8-96b6-695f161d5ace\") " pod="openstack-operators/barbican-operator-controller-manager-675b7b5b56-95zhm" Mar 07 14:56:08 crc kubenswrapper[4943]: I0307 14:56:08.309329 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhphs\" (UniqueName: \"kubernetes.io/projected/ebf90d0c-1550-42d8-96b6-695f161d5ace-kube-api-access-rhphs\") pod \"barbican-operator-controller-manager-675b7b5b56-95zhm\" (UID: \"ebf90d0c-1550-42d8-96b6-695f161d5ace\") " pod="openstack-operators/barbican-operator-controller-manager-675b7b5b56-95zhm" Mar 07 14:56:08 crc kubenswrapper[4943]: I0307 14:56:08.394240 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-675b7b5b56-95zhm" Mar 07 14:56:08 crc kubenswrapper[4943]: I0307 14:56:08.886983 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-675b7b5b56-95zhm"] Mar 07 14:56:09 crc kubenswrapper[4943]: I0307 14:56:09.759407 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-675b7b5b56-95zhm" event={"ID":"ebf90d0c-1550-42d8-96b6-695f161d5ace","Type":"ContainerStarted","Data":"457255f4cefb6dfa4111233731002acba04c8ce42e5a5703b2db70c20f7b2e36"} Mar 07 14:56:11 crc kubenswrapper[4943]: I0307 14:56:11.783593 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-675b7b5b56-95zhm" event={"ID":"ebf90d0c-1550-42d8-96b6-695f161d5ace","Type":"ContainerStarted","Data":"f9d539d56da21d68e85dd06309e64e213d188b760a7ecf727d16fcbbbf2f802a"} Mar 07 14:56:11 crc kubenswrapper[4943]: I0307 14:56:11.784319 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-675b7b5b56-95zhm" Mar 07 14:56:11 crc kubenswrapper[4943]: I0307 14:56:11.817708 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-675b7b5b56-95zhm" podStartSLOduration=1.861755854 podStartE2EDuration="3.817680941s" podCreationTimestamp="2026-03-07 14:56:08 +0000 UTC" firstStartedPulling="2026-03-07 14:56:08.898336362 +0000 UTC m=+1010.850472860" lastFinishedPulling="2026-03-07 14:56:10.854261409 +0000 UTC m=+1012.806397947" observedRunningTime="2026-03-07 14:56:11.80991645 +0000 UTC m=+1013.762052988" watchObservedRunningTime="2026-03-07 14:56:11.817680941 +0000 UTC m=+1013.769817489" Mar 07 14:56:12 crc kubenswrapper[4943]: I0307 14:56:12.266493 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-whqgv" Mar 07 14:56:12 crc kubenswrapper[4943]: I0307 14:56:12.266892 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-whqgv" Mar 07 14:56:13 crc kubenswrapper[4943]: I0307 14:56:13.321575 4943 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-whqgv" podUID="8ea7d2c3-7a0f-48dd-826f-5b4806612e66" containerName="registry-server" probeResult="failure" output=< Mar 07 14:56:13 crc kubenswrapper[4943]: timeout: failed to connect service ":50051" within 1s Mar 07 14:56:13 crc kubenswrapper[4943]: > Mar 07 14:56:18 crc kubenswrapper[4943]: I0307 14:56:18.400645 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-675b7b5b56-95zhm" Mar 07 14:56:19 crc kubenswrapper[4943]: I0307 14:56:19.699900 4943 scope.go:117] "RemoveContainer" containerID="47a10ef328ff258dcc2bbdee9f96df4df3cb5742ca81b9d42a9a55dcb2c42fb4" Mar 07 14:56:20 crc kubenswrapper[4943]: I0307 14:56:20.364269 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/barbican-db-create-52f2b"] Mar 07 14:56:20 crc kubenswrapper[4943]: I0307 14:56:20.365242 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/barbican-db-create-52f2b" Mar 07 14:56:20 crc kubenswrapper[4943]: I0307 14:56:20.380605 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/barbican-f3ae-account-create-update-5pbbk"] Mar 07 14:56:20 crc kubenswrapper[4943]: I0307 14:56:20.381976 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/barbican-f3ae-account-create-update-5pbbk" Mar 07 14:56:20 crc kubenswrapper[4943]: I0307 14:56:20.384311 4943 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"barbican-db-secret" Mar 07 14:56:20 crc kubenswrapper[4943]: I0307 14:56:20.396236 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/barbican-db-create-52f2b"] Mar 07 14:56:20 crc kubenswrapper[4943]: I0307 14:56:20.403701 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/barbican-f3ae-account-create-update-5pbbk"] Mar 07 14:56:20 crc kubenswrapper[4943]: I0307 14:56:20.483916 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f2bd919a-7322-446b-82e3-dab28d834f96-operator-scripts\") pod \"barbican-f3ae-account-create-update-5pbbk\" (UID: \"f2bd919a-7322-446b-82e3-dab28d834f96\") " pod="swift-kuttl-tests/barbican-f3ae-account-create-update-5pbbk" Mar 07 14:56:20 crc kubenswrapper[4943]: I0307 14:56:20.484049 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r77km\" (UniqueName: \"kubernetes.io/projected/f2bd919a-7322-446b-82e3-dab28d834f96-kube-api-access-r77km\") pod \"barbican-f3ae-account-create-update-5pbbk\" (UID: \"f2bd919a-7322-446b-82e3-dab28d834f96\") " pod="swift-kuttl-tests/barbican-f3ae-account-create-update-5pbbk" Mar 07 14:56:20 crc kubenswrapper[4943]: I0307 14:56:20.484130 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzccq\" (UniqueName: \"kubernetes.io/projected/dc429dfa-6f9e-4715-9eb7-daccbbad62b6-kube-api-access-hzccq\") pod \"barbican-db-create-52f2b\" (UID: \"dc429dfa-6f9e-4715-9eb7-daccbbad62b6\") " pod="swift-kuttl-tests/barbican-db-create-52f2b" Mar 07 14:56:20 crc kubenswrapper[4943]: I0307 14:56:20.484198 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc429dfa-6f9e-4715-9eb7-daccbbad62b6-operator-scripts\") pod \"barbican-db-create-52f2b\" (UID: \"dc429dfa-6f9e-4715-9eb7-daccbbad62b6\") " pod="swift-kuttl-tests/barbican-db-create-52f2b" Mar 07 14:56:20 crc kubenswrapper[4943]: I0307 14:56:20.585800 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f2bd919a-7322-446b-82e3-dab28d834f96-operator-scripts\") pod \"barbican-f3ae-account-create-update-5pbbk\" (UID: \"f2bd919a-7322-446b-82e3-dab28d834f96\") " pod="swift-kuttl-tests/barbican-f3ae-account-create-update-5pbbk" Mar 07 14:56:20 crc kubenswrapper[4943]: I0307 14:56:20.585861 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r77km\" (UniqueName: \"kubernetes.io/projected/f2bd919a-7322-446b-82e3-dab28d834f96-kube-api-access-r77km\") pod \"barbican-f3ae-account-create-update-5pbbk\" (UID: \"f2bd919a-7322-446b-82e3-dab28d834f96\") " pod="swift-kuttl-tests/barbican-f3ae-account-create-update-5pbbk" Mar 07 14:56:20 crc kubenswrapper[4943]: I0307 14:56:20.585908 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzccq\" (UniqueName: \"kubernetes.io/projected/dc429dfa-6f9e-4715-9eb7-daccbbad62b6-kube-api-access-hzccq\") pod \"barbican-db-create-52f2b\" (UID: \"dc429dfa-6f9e-4715-9eb7-daccbbad62b6\") " pod="swift-kuttl-tests/barbican-db-create-52f2b" Mar 07 14:56:20 crc kubenswrapper[4943]: I0307 14:56:20.585961 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc429dfa-6f9e-4715-9eb7-daccbbad62b6-operator-scripts\") pod \"barbican-db-create-52f2b\" (UID: \"dc429dfa-6f9e-4715-9eb7-daccbbad62b6\") " pod="swift-kuttl-tests/barbican-db-create-52f2b" Mar 07 14:56:20 crc kubenswrapper[4943]: I0307 14:56:20.586857 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc429dfa-6f9e-4715-9eb7-daccbbad62b6-operator-scripts\") pod \"barbican-db-create-52f2b\" (UID: \"dc429dfa-6f9e-4715-9eb7-daccbbad62b6\") " pod="swift-kuttl-tests/barbican-db-create-52f2b" Mar 07 14:56:20 crc kubenswrapper[4943]: I0307 14:56:20.587157 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f2bd919a-7322-446b-82e3-dab28d834f96-operator-scripts\") pod \"barbican-f3ae-account-create-update-5pbbk\" (UID: \"f2bd919a-7322-446b-82e3-dab28d834f96\") " pod="swift-kuttl-tests/barbican-f3ae-account-create-update-5pbbk" Mar 07 14:56:20 crc kubenswrapper[4943]: I0307 14:56:20.613228 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzccq\" (UniqueName: \"kubernetes.io/projected/dc429dfa-6f9e-4715-9eb7-daccbbad62b6-kube-api-access-hzccq\") pod \"barbican-db-create-52f2b\" (UID: \"dc429dfa-6f9e-4715-9eb7-daccbbad62b6\") " pod="swift-kuttl-tests/barbican-db-create-52f2b" Mar 07 14:56:20 crc kubenswrapper[4943]: I0307 14:56:20.617546 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r77km\" (UniqueName: \"kubernetes.io/projected/f2bd919a-7322-446b-82e3-dab28d834f96-kube-api-access-r77km\") pod \"barbican-f3ae-account-create-update-5pbbk\" (UID: \"f2bd919a-7322-446b-82e3-dab28d834f96\") " pod="swift-kuttl-tests/barbican-f3ae-account-create-update-5pbbk" Mar 07 14:56:20 crc kubenswrapper[4943]: I0307 14:56:20.680752 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/barbican-db-create-52f2b" Mar 07 14:56:20 crc kubenswrapper[4943]: I0307 14:56:20.697465 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/barbican-f3ae-account-create-update-5pbbk" Mar 07 14:56:21 crc kubenswrapper[4943]: I0307 14:56:21.334544 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/barbican-f3ae-account-create-update-5pbbk"] Mar 07 14:56:21 crc kubenswrapper[4943]: W0307 14:56:21.340261 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf2bd919a_7322_446b_82e3_dab28d834f96.slice/crio-90ccbd80dda73a57676e62f1b613f8264efeb7e2fe26b8753b378ada2cac4790 WatchSource:0}: Error finding container 90ccbd80dda73a57676e62f1b613f8264efeb7e2fe26b8753b378ada2cac4790: Status 404 returned error can't find the container with id 90ccbd80dda73a57676e62f1b613f8264efeb7e2fe26b8753b378ada2cac4790 Mar 07 14:56:21 crc kubenswrapper[4943]: I0307 14:56:21.413460 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/barbican-db-create-52f2b"] Mar 07 14:56:21 crc kubenswrapper[4943]: I0307 14:56:21.901207 4943 generic.go:334] "Generic (PLEG): container finished" podID="dc429dfa-6f9e-4715-9eb7-daccbbad62b6" containerID="4e0a94c932a04db92eef6513ced511a8e2e1584cab874c3afd4d3cb291b46377" exitCode=0 Mar 07 14:56:21 crc kubenswrapper[4943]: I0307 14:56:21.901323 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/barbican-db-create-52f2b" event={"ID":"dc429dfa-6f9e-4715-9eb7-daccbbad62b6","Type":"ContainerDied","Data":"4e0a94c932a04db92eef6513ced511a8e2e1584cab874c3afd4d3cb291b46377"} Mar 07 14:56:21 crc kubenswrapper[4943]: I0307 14:56:21.901683 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/barbican-db-create-52f2b" event={"ID":"dc429dfa-6f9e-4715-9eb7-daccbbad62b6","Type":"ContainerStarted","Data":"fb1914139db9bb18e4bd1b13a496937a5bd3e073f7091c69320ef096f0c12249"} Mar 07 14:56:21 crc kubenswrapper[4943]: I0307 14:56:21.905054 4943 generic.go:334] "Generic (PLEG): container finished" podID="f2bd919a-7322-446b-82e3-dab28d834f96" containerID="a4e6945468fd8ddc3d4d40a6d039ba2e245da2a8ca630f738f186b6895066a42" exitCode=0 Mar 07 14:56:21 crc kubenswrapper[4943]: I0307 14:56:21.905108 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/barbican-f3ae-account-create-update-5pbbk" event={"ID":"f2bd919a-7322-446b-82e3-dab28d834f96","Type":"ContainerDied","Data":"a4e6945468fd8ddc3d4d40a6d039ba2e245da2a8ca630f738f186b6895066a42"} Mar 07 14:56:21 crc kubenswrapper[4943]: I0307 14:56:21.905420 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/barbican-f3ae-account-create-update-5pbbk" event={"ID":"f2bd919a-7322-446b-82e3-dab28d834f96","Type":"ContainerStarted","Data":"90ccbd80dda73a57676e62f1b613f8264efeb7e2fe26b8753b378ada2cac4790"} Mar 07 14:56:22 crc kubenswrapper[4943]: I0307 14:56:22.337135 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-whqgv" Mar 07 14:56:22 crc kubenswrapper[4943]: I0307 14:56:22.399599 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-whqgv" Mar 07 14:56:23 crc kubenswrapper[4943]: I0307 14:56:23.276393 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-index-zltk7"] Mar 07 14:56:23 crc kubenswrapper[4943]: I0307 14:56:23.277681 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-zltk7" Mar 07 14:56:23 crc kubenswrapper[4943]: I0307 14:56:23.279447 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/barbican-db-create-52f2b" Mar 07 14:56:23 crc kubenswrapper[4943]: I0307 14:56:23.280776 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-index-dockercfg-hth9h" Mar 07 14:56:23 crc kubenswrapper[4943]: I0307 14:56:23.286522 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-index-zltk7"] Mar 07 14:56:23 crc kubenswrapper[4943]: I0307 14:56:23.324698 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc429dfa-6f9e-4715-9eb7-daccbbad62b6-operator-scripts\") pod \"dc429dfa-6f9e-4715-9eb7-daccbbad62b6\" (UID: \"dc429dfa-6f9e-4715-9eb7-daccbbad62b6\") " Mar 07 14:56:23 crc kubenswrapper[4943]: I0307 14:56:23.324801 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hzccq\" (UniqueName: \"kubernetes.io/projected/dc429dfa-6f9e-4715-9eb7-daccbbad62b6-kube-api-access-hzccq\") pod \"dc429dfa-6f9e-4715-9eb7-daccbbad62b6\" (UID: \"dc429dfa-6f9e-4715-9eb7-daccbbad62b6\") " Mar 07 14:56:23 crc kubenswrapper[4943]: I0307 14:56:23.325073 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5rms\" (UniqueName: \"kubernetes.io/projected/89b40fec-692e-44d3-9290-66c7a2931e40-kube-api-access-x5rms\") pod \"swift-operator-index-zltk7\" (UID: \"89b40fec-692e-44d3-9290-66c7a2931e40\") " pod="openstack-operators/swift-operator-index-zltk7" Mar 07 14:56:23 crc kubenswrapper[4943]: I0307 14:56:23.325486 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc429dfa-6f9e-4715-9eb7-daccbbad62b6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "dc429dfa-6f9e-4715-9eb7-daccbbad62b6" (UID: "dc429dfa-6f9e-4715-9eb7-daccbbad62b6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:56:23 crc kubenswrapper[4943]: I0307 14:56:23.335195 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc429dfa-6f9e-4715-9eb7-daccbbad62b6-kube-api-access-hzccq" (OuterVolumeSpecName: "kube-api-access-hzccq") pod "dc429dfa-6f9e-4715-9eb7-daccbbad62b6" (UID: "dc429dfa-6f9e-4715-9eb7-daccbbad62b6"). InnerVolumeSpecName "kube-api-access-hzccq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:56:23 crc kubenswrapper[4943]: I0307 14:56:23.370541 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/barbican-f3ae-account-create-update-5pbbk" Mar 07 14:56:23 crc kubenswrapper[4943]: I0307 14:56:23.425993 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r77km\" (UniqueName: \"kubernetes.io/projected/f2bd919a-7322-446b-82e3-dab28d834f96-kube-api-access-r77km\") pod \"f2bd919a-7322-446b-82e3-dab28d834f96\" (UID: \"f2bd919a-7322-446b-82e3-dab28d834f96\") " Mar 07 14:56:23 crc kubenswrapper[4943]: I0307 14:56:23.426079 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f2bd919a-7322-446b-82e3-dab28d834f96-operator-scripts\") pod \"f2bd919a-7322-446b-82e3-dab28d834f96\" (UID: \"f2bd919a-7322-446b-82e3-dab28d834f96\") " Mar 07 14:56:23 crc kubenswrapper[4943]: I0307 14:56:23.426285 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5rms\" (UniqueName: \"kubernetes.io/projected/89b40fec-692e-44d3-9290-66c7a2931e40-kube-api-access-x5rms\") pod \"swift-operator-index-zltk7\" (UID: \"89b40fec-692e-44d3-9290-66c7a2931e40\") " pod="openstack-operators/swift-operator-index-zltk7" Mar 07 14:56:23 crc kubenswrapper[4943]: I0307 14:56:23.426390 4943 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc429dfa-6f9e-4715-9eb7-daccbbad62b6-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 14:56:23 crc kubenswrapper[4943]: I0307 14:56:23.426401 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hzccq\" (UniqueName: \"kubernetes.io/projected/dc429dfa-6f9e-4715-9eb7-daccbbad62b6-kube-api-access-hzccq\") on node \"crc\" DevicePath \"\"" Mar 07 14:56:23 crc kubenswrapper[4943]: I0307 14:56:23.426590 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2bd919a-7322-446b-82e3-dab28d834f96-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f2bd919a-7322-446b-82e3-dab28d834f96" (UID: "f2bd919a-7322-446b-82e3-dab28d834f96"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:56:23 crc kubenswrapper[4943]: I0307 14:56:23.430869 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2bd919a-7322-446b-82e3-dab28d834f96-kube-api-access-r77km" (OuterVolumeSpecName: "kube-api-access-r77km") pod "f2bd919a-7322-446b-82e3-dab28d834f96" (UID: "f2bd919a-7322-446b-82e3-dab28d834f96"). InnerVolumeSpecName "kube-api-access-r77km". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:56:23 crc kubenswrapper[4943]: I0307 14:56:23.443371 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5rms\" (UniqueName: \"kubernetes.io/projected/89b40fec-692e-44d3-9290-66c7a2931e40-kube-api-access-x5rms\") pod \"swift-operator-index-zltk7\" (UID: \"89b40fec-692e-44d3-9290-66c7a2931e40\") " pod="openstack-operators/swift-operator-index-zltk7" Mar 07 14:56:23 crc kubenswrapper[4943]: I0307 14:56:23.528043 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r77km\" (UniqueName: \"kubernetes.io/projected/f2bd919a-7322-446b-82e3-dab28d834f96-kube-api-access-r77km\") on node \"crc\" DevicePath \"\"" Mar 07 14:56:23 crc kubenswrapper[4943]: I0307 14:56:23.528087 4943 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f2bd919a-7322-446b-82e3-dab28d834f96-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 14:56:23 crc kubenswrapper[4943]: I0307 14:56:23.593205 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-zltk7" Mar 07 14:56:23 crc kubenswrapper[4943]: I0307 14:56:23.816179 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-index-zltk7"] Mar 07 14:56:23 crc kubenswrapper[4943]: I0307 14:56:23.931386 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/barbican-db-create-52f2b" event={"ID":"dc429dfa-6f9e-4715-9eb7-daccbbad62b6","Type":"ContainerDied","Data":"fb1914139db9bb18e4bd1b13a496937a5bd3e073f7091c69320ef096f0c12249"} Mar 07 14:56:23 crc kubenswrapper[4943]: I0307 14:56:23.931439 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fb1914139db9bb18e4bd1b13a496937a5bd3e073f7091c69320ef096f0c12249" Mar 07 14:56:23 crc kubenswrapper[4943]: I0307 14:56:23.931519 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/barbican-db-create-52f2b" Mar 07 14:56:23 crc kubenswrapper[4943]: I0307 14:56:23.938714 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-zltk7" event={"ID":"89b40fec-692e-44d3-9290-66c7a2931e40","Type":"ContainerStarted","Data":"db62541a58abe78c547fc5a06d37f697d8dd59ec1ec5554ada0197d832912fba"} Mar 07 14:56:23 crc kubenswrapper[4943]: I0307 14:56:23.943582 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/barbican-f3ae-account-create-update-5pbbk" event={"ID":"f2bd919a-7322-446b-82e3-dab28d834f96","Type":"ContainerDied","Data":"90ccbd80dda73a57676e62f1b613f8264efeb7e2fe26b8753b378ada2cac4790"} Mar 07 14:56:23 crc kubenswrapper[4943]: I0307 14:56:23.943611 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="90ccbd80dda73a57676e62f1b613f8264efeb7e2fe26b8753b378ada2cac4790" Mar 07 14:56:23 crc kubenswrapper[4943]: I0307 14:56:23.943663 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/barbican-f3ae-account-create-update-5pbbk" Mar 07 14:56:25 crc kubenswrapper[4943]: I0307 14:56:25.661998 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/barbican-db-sync-xtjgw"] Mar 07 14:56:25 crc kubenswrapper[4943]: E0307 14:56:25.662983 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2bd919a-7322-446b-82e3-dab28d834f96" containerName="mariadb-account-create-update" Mar 07 14:56:25 crc kubenswrapper[4943]: I0307 14:56:25.663000 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2bd919a-7322-446b-82e3-dab28d834f96" containerName="mariadb-account-create-update" Mar 07 14:56:25 crc kubenswrapper[4943]: E0307 14:56:25.663031 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc429dfa-6f9e-4715-9eb7-daccbbad62b6" containerName="mariadb-database-create" Mar 07 14:56:25 crc kubenswrapper[4943]: I0307 14:56:25.663041 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc429dfa-6f9e-4715-9eb7-daccbbad62b6" containerName="mariadb-database-create" Mar 07 14:56:25 crc kubenswrapper[4943]: I0307 14:56:25.663195 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc429dfa-6f9e-4715-9eb7-daccbbad62b6" containerName="mariadb-database-create" Mar 07 14:56:25 crc kubenswrapper[4943]: I0307 14:56:25.663218 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2bd919a-7322-446b-82e3-dab28d834f96" containerName="mariadb-account-create-update" Mar 07 14:56:25 crc kubenswrapper[4943]: I0307 14:56:25.663762 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/barbican-db-sync-xtjgw" Mar 07 14:56:25 crc kubenswrapper[4943]: I0307 14:56:25.666854 4943 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"barbican-config-data" Mar 07 14:56:25 crc kubenswrapper[4943]: I0307 14:56:25.667650 4943 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"barbican-barbican-dockercfg-7cj95" Mar 07 14:56:25 crc kubenswrapper[4943]: I0307 14:56:25.671885 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/barbican-db-sync-xtjgw"] Mar 07 14:56:25 crc kubenswrapper[4943]: I0307 14:56:25.764904 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/76563709-9e41-463e-a16f-d30e9cd4abf8-db-sync-config-data\") pod \"barbican-db-sync-xtjgw\" (UID: \"76563709-9e41-463e-a16f-d30e9cd4abf8\") " pod="swift-kuttl-tests/barbican-db-sync-xtjgw" Mar 07 14:56:25 crc kubenswrapper[4943]: I0307 14:56:25.765331 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtbjm\" (UniqueName: \"kubernetes.io/projected/76563709-9e41-463e-a16f-d30e9cd4abf8-kube-api-access-gtbjm\") pod \"barbican-db-sync-xtjgw\" (UID: \"76563709-9e41-463e-a16f-d30e9cd4abf8\") " pod="swift-kuttl-tests/barbican-db-sync-xtjgw" Mar 07 14:56:25 crc kubenswrapper[4943]: I0307 14:56:25.866974 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtbjm\" (UniqueName: \"kubernetes.io/projected/76563709-9e41-463e-a16f-d30e9cd4abf8-kube-api-access-gtbjm\") pod \"barbican-db-sync-xtjgw\" (UID: \"76563709-9e41-463e-a16f-d30e9cd4abf8\") " pod="swift-kuttl-tests/barbican-db-sync-xtjgw" Mar 07 14:56:25 crc kubenswrapper[4943]: I0307 14:56:25.867143 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/76563709-9e41-463e-a16f-d30e9cd4abf8-db-sync-config-data\") pod \"barbican-db-sync-xtjgw\" (UID: \"76563709-9e41-463e-a16f-d30e9cd4abf8\") " pod="swift-kuttl-tests/barbican-db-sync-xtjgw" Mar 07 14:56:25 crc kubenswrapper[4943]: I0307 14:56:25.876019 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/76563709-9e41-463e-a16f-d30e9cd4abf8-db-sync-config-data\") pod \"barbican-db-sync-xtjgw\" (UID: \"76563709-9e41-463e-a16f-d30e9cd4abf8\") " pod="swift-kuttl-tests/barbican-db-sync-xtjgw" Mar 07 14:56:25 crc kubenswrapper[4943]: I0307 14:56:25.884892 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtbjm\" (UniqueName: \"kubernetes.io/projected/76563709-9e41-463e-a16f-d30e9cd4abf8-kube-api-access-gtbjm\") pod \"barbican-db-sync-xtjgw\" (UID: \"76563709-9e41-463e-a16f-d30e9cd4abf8\") " pod="swift-kuttl-tests/barbican-db-sync-xtjgw" Mar 07 14:56:26 crc kubenswrapper[4943]: I0307 14:56:26.000757 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/barbican-db-sync-xtjgw" Mar 07 14:56:26 crc kubenswrapper[4943]: I0307 14:56:26.868735 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-whqgv"] Mar 07 14:56:26 crc kubenswrapper[4943]: I0307 14:56:26.869712 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-whqgv" podUID="8ea7d2c3-7a0f-48dd-826f-5b4806612e66" containerName="registry-server" containerID="cri-o://e14311cea59eaafea745941fcaf7dbbaa13565700d7ecc2d05890c621bdb566f" gracePeriod=2 Mar 07 14:56:26 crc kubenswrapper[4943]: I0307 14:56:26.992104 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-zltk7" event={"ID":"89b40fec-692e-44d3-9290-66c7a2931e40","Type":"ContainerStarted","Data":"f047d36b71318ab6cc14057cdf2f66aeec4f4e3577be2fe456cd5024d3fe3964"} Mar 07 14:56:27 crc kubenswrapper[4943]: I0307 14:56:27.030063 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/barbican-db-sync-xtjgw"] Mar 07 14:56:27 crc kubenswrapper[4943]: I0307 14:56:27.035596 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-index-zltk7" podStartSLOduration=1.312310312 podStartE2EDuration="4.035582159s" podCreationTimestamp="2026-03-07 14:56:23 +0000 UTC" firstStartedPulling="2026-03-07 14:56:23.83157506 +0000 UTC m=+1025.783711558" lastFinishedPulling="2026-03-07 14:56:26.554846897 +0000 UTC m=+1028.506983405" observedRunningTime="2026-03-07 14:56:27.030598836 +0000 UTC m=+1028.982735334" watchObservedRunningTime="2026-03-07 14:56:27.035582159 +0000 UTC m=+1028.987718667" Mar 07 14:56:27 crc kubenswrapper[4943]: I0307 14:56:27.404530 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-whqgv" Mar 07 14:56:27 crc kubenswrapper[4943]: I0307 14:56:27.499212 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wbv6t\" (UniqueName: \"kubernetes.io/projected/8ea7d2c3-7a0f-48dd-826f-5b4806612e66-kube-api-access-wbv6t\") pod \"8ea7d2c3-7a0f-48dd-826f-5b4806612e66\" (UID: \"8ea7d2c3-7a0f-48dd-826f-5b4806612e66\") " Mar 07 14:56:27 crc kubenswrapper[4943]: I0307 14:56:27.499294 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ea7d2c3-7a0f-48dd-826f-5b4806612e66-utilities\") pod \"8ea7d2c3-7a0f-48dd-826f-5b4806612e66\" (UID: \"8ea7d2c3-7a0f-48dd-826f-5b4806612e66\") " Mar 07 14:56:27 crc kubenswrapper[4943]: I0307 14:56:27.499376 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ea7d2c3-7a0f-48dd-826f-5b4806612e66-catalog-content\") pod \"8ea7d2c3-7a0f-48dd-826f-5b4806612e66\" (UID: \"8ea7d2c3-7a0f-48dd-826f-5b4806612e66\") " Mar 07 14:56:27 crc kubenswrapper[4943]: I0307 14:56:27.500973 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ea7d2c3-7a0f-48dd-826f-5b4806612e66-utilities" (OuterVolumeSpecName: "utilities") pod "8ea7d2c3-7a0f-48dd-826f-5b4806612e66" (UID: "8ea7d2c3-7a0f-48dd-826f-5b4806612e66"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:56:27 crc kubenswrapper[4943]: I0307 14:56:27.507382 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ea7d2c3-7a0f-48dd-826f-5b4806612e66-kube-api-access-wbv6t" (OuterVolumeSpecName: "kube-api-access-wbv6t") pod "8ea7d2c3-7a0f-48dd-826f-5b4806612e66" (UID: "8ea7d2c3-7a0f-48dd-826f-5b4806612e66"). InnerVolumeSpecName "kube-api-access-wbv6t". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:56:27 crc kubenswrapper[4943]: I0307 14:56:27.600727 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wbv6t\" (UniqueName: \"kubernetes.io/projected/8ea7d2c3-7a0f-48dd-826f-5b4806612e66-kube-api-access-wbv6t\") on node \"crc\" DevicePath \"\"" Mar 07 14:56:27 crc kubenswrapper[4943]: I0307 14:56:27.600782 4943 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ea7d2c3-7a0f-48dd-826f-5b4806612e66-utilities\") on node \"crc\" DevicePath \"\"" Mar 07 14:56:27 crc kubenswrapper[4943]: I0307 14:56:27.664655 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ea7d2c3-7a0f-48dd-826f-5b4806612e66-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8ea7d2c3-7a0f-48dd-826f-5b4806612e66" (UID: "8ea7d2c3-7a0f-48dd-826f-5b4806612e66"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:56:27 crc kubenswrapper[4943]: I0307 14:56:27.705664 4943 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ea7d2c3-7a0f-48dd-826f-5b4806612e66-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 07 14:56:28 crc kubenswrapper[4943]: I0307 14:56:28.006022 4943 generic.go:334] "Generic (PLEG): container finished" podID="8ea7d2c3-7a0f-48dd-826f-5b4806612e66" containerID="e14311cea59eaafea745941fcaf7dbbaa13565700d7ecc2d05890c621bdb566f" exitCode=0 Mar 07 14:56:28 crc kubenswrapper[4943]: I0307 14:56:28.006138 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-whqgv" Mar 07 14:56:28 crc kubenswrapper[4943]: I0307 14:56:28.006309 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-whqgv" event={"ID":"8ea7d2c3-7a0f-48dd-826f-5b4806612e66","Type":"ContainerDied","Data":"e14311cea59eaafea745941fcaf7dbbaa13565700d7ecc2d05890c621bdb566f"} Mar 07 14:56:28 crc kubenswrapper[4943]: I0307 14:56:28.006402 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-whqgv" event={"ID":"8ea7d2c3-7a0f-48dd-826f-5b4806612e66","Type":"ContainerDied","Data":"cbd345bd6691301795a78042ca1943a3745dd7faef6fcb6dd1b14d1d37552603"} Mar 07 14:56:28 crc kubenswrapper[4943]: I0307 14:56:28.006435 4943 scope.go:117] "RemoveContainer" containerID="e14311cea59eaafea745941fcaf7dbbaa13565700d7ecc2d05890c621bdb566f" Mar 07 14:56:28 crc kubenswrapper[4943]: I0307 14:56:28.011113 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/barbican-db-sync-xtjgw" event={"ID":"76563709-9e41-463e-a16f-d30e9cd4abf8","Type":"ContainerStarted","Data":"950f68545b9cc8cb3b604fdd991dc827a68db3b0253238c63c85e98fc3531a2f"} Mar 07 14:56:28 crc kubenswrapper[4943]: I0307 14:56:28.026320 4943 scope.go:117] "RemoveContainer" containerID="457d722faf5a6dba954ea95543a456183feba10955952d148d3b54a614fcd140" Mar 07 14:56:28 crc kubenswrapper[4943]: I0307 14:56:28.045746 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-whqgv"] Mar 07 14:56:28 crc kubenswrapper[4943]: I0307 14:56:28.050443 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-whqgv"] Mar 07 14:56:28 crc kubenswrapper[4943]: I0307 14:56:28.065435 4943 scope.go:117] "RemoveContainer" containerID="3b8dd0f0cf7529b2c288ce9dd6453698eefb85758f94635f6e781656f18bee4a" Mar 07 14:56:28 crc kubenswrapper[4943]: I0307 14:56:28.092356 4943 scope.go:117] "RemoveContainer" containerID="e14311cea59eaafea745941fcaf7dbbaa13565700d7ecc2d05890c621bdb566f" Mar 07 14:56:28 crc kubenswrapper[4943]: E0307 14:56:28.092896 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e14311cea59eaafea745941fcaf7dbbaa13565700d7ecc2d05890c621bdb566f\": container with ID starting with e14311cea59eaafea745941fcaf7dbbaa13565700d7ecc2d05890c621bdb566f not found: ID does not exist" containerID="e14311cea59eaafea745941fcaf7dbbaa13565700d7ecc2d05890c621bdb566f" Mar 07 14:56:28 crc kubenswrapper[4943]: I0307 14:56:28.092956 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e14311cea59eaafea745941fcaf7dbbaa13565700d7ecc2d05890c621bdb566f"} err="failed to get container status \"e14311cea59eaafea745941fcaf7dbbaa13565700d7ecc2d05890c621bdb566f\": rpc error: code = NotFound desc = could not find container \"e14311cea59eaafea745941fcaf7dbbaa13565700d7ecc2d05890c621bdb566f\": container with ID starting with e14311cea59eaafea745941fcaf7dbbaa13565700d7ecc2d05890c621bdb566f not found: ID does not exist" Mar 07 14:56:28 crc kubenswrapper[4943]: I0307 14:56:28.093024 4943 scope.go:117] "RemoveContainer" containerID="457d722faf5a6dba954ea95543a456183feba10955952d148d3b54a614fcd140" Mar 07 14:56:28 crc kubenswrapper[4943]: E0307 14:56:28.093386 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"457d722faf5a6dba954ea95543a456183feba10955952d148d3b54a614fcd140\": container with ID starting with 457d722faf5a6dba954ea95543a456183feba10955952d148d3b54a614fcd140 not found: ID does not exist" containerID="457d722faf5a6dba954ea95543a456183feba10955952d148d3b54a614fcd140" Mar 07 14:56:28 crc kubenswrapper[4943]: I0307 14:56:28.093423 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"457d722faf5a6dba954ea95543a456183feba10955952d148d3b54a614fcd140"} err="failed to get container status \"457d722faf5a6dba954ea95543a456183feba10955952d148d3b54a614fcd140\": rpc error: code = NotFound desc = could not find container \"457d722faf5a6dba954ea95543a456183feba10955952d148d3b54a614fcd140\": container with ID starting with 457d722faf5a6dba954ea95543a456183feba10955952d148d3b54a614fcd140 not found: ID does not exist" Mar 07 14:56:28 crc kubenswrapper[4943]: I0307 14:56:28.093449 4943 scope.go:117] "RemoveContainer" containerID="3b8dd0f0cf7529b2c288ce9dd6453698eefb85758f94635f6e781656f18bee4a" Mar 07 14:56:28 crc kubenswrapper[4943]: E0307 14:56:28.093888 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b8dd0f0cf7529b2c288ce9dd6453698eefb85758f94635f6e781656f18bee4a\": container with ID starting with 3b8dd0f0cf7529b2c288ce9dd6453698eefb85758f94635f6e781656f18bee4a not found: ID does not exist" containerID="3b8dd0f0cf7529b2c288ce9dd6453698eefb85758f94635f6e781656f18bee4a" Mar 07 14:56:28 crc kubenswrapper[4943]: I0307 14:56:28.093911 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b8dd0f0cf7529b2c288ce9dd6453698eefb85758f94635f6e781656f18bee4a"} err="failed to get container status \"3b8dd0f0cf7529b2c288ce9dd6453698eefb85758f94635f6e781656f18bee4a\": rpc error: code = NotFound desc = could not find container \"3b8dd0f0cf7529b2c288ce9dd6453698eefb85758f94635f6e781656f18bee4a\": container with ID starting with 3b8dd0f0cf7529b2c288ce9dd6453698eefb85758f94635f6e781656f18bee4a not found: ID does not exist" Mar 07 14:56:28 crc kubenswrapper[4943]: I0307 14:56:28.767590 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ea7d2c3-7a0f-48dd-826f-5b4806612e66" path="/var/lib/kubelet/pods/8ea7d2c3-7a0f-48dd-826f-5b4806612e66/volumes" Mar 07 14:56:32 crc kubenswrapper[4943]: I0307 14:56:32.042012 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/barbican-db-sync-xtjgw" event={"ID":"76563709-9e41-463e-a16f-d30e9cd4abf8","Type":"ContainerStarted","Data":"17995ddb533005d0fdd1e4adeb1555b7567492101161dfe9e0fd06b9549e6619"} Mar 07 14:56:33 crc kubenswrapper[4943]: I0307 14:56:33.594402 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-index-zltk7" Mar 07 14:56:33 crc kubenswrapper[4943]: I0307 14:56:33.594880 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/swift-operator-index-zltk7" Mar 07 14:56:33 crc kubenswrapper[4943]: I0307 14:56:33.640056 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/swift-operator-index-zltk7" Mar 07 14:56:33 crc kubenswrapper[4943]: I0307 14:56:33.661791 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/barbican-db-sync-xtjgw" podStartSLOduration=4.222900146 podStartE2EDuration="8.66176847s" podCreationTimestamp="2026-03-07 14:56:25 +0000 UTC" firstStartedPulling="2026-03-07 14:56:27.048141438 +0000 UTC m=+1029.000277946" lastFinishedPulling="2026-03-07 14:56:31.487009762 +0000 UTC m=+1033.439146270" observedRunningTime="2026-03-07 14:56:32.070413548 +0000 UTC m=+1034.022550086" watchObservedRunningTime="2026-03-07 14:56:33.66176847 +0000 UTC m=+1035.613904968" Mar 07 14:56:34 crc kubenswrapper[4943]: I0307 14:56:34.124698 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-index-zltk7" Mar 07 14:56:35 crc kubenswrapper[4943]: I0307 14:56:35.096003 4943 generic.go:334] "Generic (PLEG): container finished" podID="76563709-9e41-463e-a16f-d30e9cd4abf8" containerID="17995ddb533005d0fdd1e4adeb1555b7567492101161dfe9e0fd06b9549e6619" exitCode=0 Mar 07 14:56:35 crc kubenswrapper[4943]: I0307 14:56:35.096088 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/barbican-db-sync-xtjgw" event={"ID":"76563709-9e41-463e-a16f-d30e9cd4abf8","Type":"ContainerDied","Data":"17995ddb533005d0fdd1e4adeb1555b7567492101161dfe9e0fd06b9549e6619"} Mar 07 14:56:36 crc kubenswrapper[4943]: I0307 14:56:36.407281 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/barbican-db-sync-xtjgw" Mar 07 14:56:36 crc kubenswrapper[4943]: I0307 14:56:36.455459 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gtbjm\" (UniqueName: \"kubernetes.io/projected/76563709-9e41-463e-a16f-d30e9cd4abf8-kube-api-access-gtbjm\") pod \"76563709-9e41-463e-a16f-d30e9cd4abf8\" (UID: \"76563709-9e41-463e-a16f-d30e9cd4abf8\") " Mar 07 14:56:36 crc kubenswrapper[4943]: I0307 14:56:36.455908 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/76563709-9e41-463e-a16f-d30e9cd4abf8-db-sync-config-data\") pod \"76563709-9e41-463e-a16f-d30e9cd4abf8\" (UID: \"76563709-9e41-463e-a16f-d30e9cd4abf8\") " Mar 07 14:56:36 crc kubenswrapper[4943]: I0307 14:56:36.466184 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76563709-9e41-463e-a16f-d30e9cd4abf8-kube-api-access-gtbjm" (OuterVolumeSpecName: "kube-api-access-gtbjm") pod "76563709-9e41-463e-a16f-d30e9cd4abf8" (UID: "76563709-9e41-463e-a16f-d30e9cd4abf8"). InnerVolumeSpecName "kube-api-access-gtbjm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:56:36 crc kubenswrapper[4943]: I0307 14:56:36.476109 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76563709-9e41-463e-a16f-d30e9cd4abf8-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "76563709-9e41-463e-a16f-d30e9cd4abf8" (UID: "76563709-9e41-463e-a16f-d30e9cd4abf8"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:56:36 crc kubenswrapper[4943]: I0307 14:56:36.557521 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gtbjm\" (UniqueName: \"kubernetes.io/projected/76563709-9e41-463e-a16f-d30e9cd4abf8-kube-api-access-gtbjm\") on node \"crc\" DevicePath \"\"" Mar 07 14:56:36 crc kubenswrapper[4943]: I0307 14:56:36.557564 4943 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/76563709-9e41-463e-a16f-d30e9cd4abf8-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.112878 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/barbican-db-sync-xtjgw" event={"ID":"76563709-9e41-463e-a16f-d30e9cd4abf8","Type":"ContainerDied","Data":"950f68545b9cc8cb3b604fdd991dc827a68db3b0253238c63c85e98fc3531a2f"} Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.112942 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="950f68545b9cc8cb3b604fdd991dc827a68db3b0253238c63c85e98fc3531a2f" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.113000 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/barbican-db-sync-xtjgw" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.266324 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/barbican-worker-8bf566dc7-r7nnf"] Mar 07 14:56:37 crc kubenswrapper[4943]: E0307 14:56:37.266649 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ea7d2c3-7a0f-48dd-826f-5b4806612e66" containerName="extract-utilities" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.266663 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ea7d2c3-7a0f-48dd-826f-5b4806612e66" containerName="extract-utilities" Mar 07 14:56:37 crc kubenswrapper[4943]: E0307 14:56:37.266676 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ea7d2c3-7a0f-48dd-826f-5b4806612e66" containerName="registry-server" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.266684 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ea7d2c3-7a0f-48dd-826f-5b4806612e66" containerName="registry-server" Mar 07 14:56:37 crc kubenswrapper[4943]: E0307 14:56:37.266700 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ea7d2c3-7a0f-48dd-826f-5b4806612e66" containerName="extract-content" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.266708 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ea7d2c3-7a0f-48dd-826f-5b4806612e66" containerName="extract-content" Mar 07 14:56:37 crc kubenswrapper[4943]: E0307 14:56:37.266719 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76563709-9e41-463e-a16f-d30e9cd4abf8" containerName="barbican-db-sync" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.266725 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="76563709-9e41-463e-a16f-d30e9cd4abf8" containerName="barbican-db-sync" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.266863 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="76563709-9e41-463e-a16f-d30e9cd4abf8" containerName="barbican-db-sync" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.266873 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ea7d2c3-7a0f-48dd-826f-5b4806612e66" containerName="registry-server" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.267620 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/barbican-worker-8bf566dc7-r7nnf" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.278019 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/barbican-keystone-listener-54998b9498-bh574"] Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.279255 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/barbican-keystone-listener-54998b9498-bh574" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.280545 4943 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"barbican-barbican-dockercfg-7cj95" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.282297 4943 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"barbican-keystone-listener-config-data" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.282562 4943 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"barbican-config-data" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.282990 4943 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"barbican-worker-config-data" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.300757 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/barbican-worker-8bf566dc7-r7nnf"] Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.326061 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/barbican-keystone-listener-54998b9498-bh574"] Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.369611 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2159c059-075c-4986-8244-bf6d57dc3cd9-config-data-custom\") pod \"barbican-worker-8bf566dc7-r7nnf\" (UID: \"2159c059-075c-4986-8244-bf6d57dc3cd9\") " pod="swift-kuttl-tests/barbican-worker-8bf566dc7-r7nnf" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.369657 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2159c059-075c-4986-8244-bf6d57dc3cd9-logs\") pod \"barbican-worker-8bf566dc7-r7nnf\" (UID: \"2159c059-075c-4986-8244-bf6d57dc3cd9\") " pod="swift-kuttl-tests/barbican-worker-8bf566dc7-r7nnf" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.369689 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d74297b6-b4e3-4dd8-8ee0-fd403bf79254-config-data-custom\") pod \"barbican-keystone-listener-54998b9498-bh574\" (UID: \"d74297b6-b4e3-4dd8-8ee0-fd403bf79254\") " pod="swift-kuttl-tests/barbican-keystone-listener-54998b9498-bh574" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.369713 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d74297b6-b4e3-4dd8-8ee0-fd403bf79254-logs\") pod \"barbican-keystone-listener-54998b9498-bh574\" (UID: \"d74297b6-b4e3-4dd8-8ee0-fd403bf79254\") " pod="swift-kuttl-tests/barbican-keystone-listener-54998b9498-bh574" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.369755 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2159c059-075c-4986-8244-bf6d57dc3cd9-config-data\") pod \"barbican-worker-8bf566dc7-r7nnf\" (UID: \"2159c059-075c-4986-8244-bf6d57dc3cd9\") " pod="swift-kuttl-tests/barbican-worker-8bf566dc7-r7nnf" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.369777 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgtcm\" (UniqueName: \"kubernetes.io/projected/d74297b6-b4e3-4dd8-8ee0-fd403bf79254-kube-api-access-cgtcm\") pod \"barbican-keystone-listener-54998b9498-bh574\" (UID: \"d74297b6-b4e3-4dd8-8ee0-fd403bf79254\") " pod="swift-kuttl-tests/barbican-keystone-listener-54998b9498-bh574" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.369801 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d74297b6-b4e3-4dd8-8ee0-fd403bf79254-config-data\") pod \"barbican-keystone-listener-54998b9498-bh574\" (UID: \"d74297b6-b4e3-4dd8-8ee0-fd403bf79254\") " pod="swift-kuttl-tests/barbican-keystone-listener-54998b9498-bh574" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.369829 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hx2qr\" (UniqueName: \"kubernetes.io/projected/2159c059-075c-4986-8244-bf6d57dc3cd9-kube-api-access-hx2qr\") pod \"barbican-worker-8bf566dc7-r7nnf\" (UID: \"2159c059-075c-4986-8244-bf6d57dc3cd9\") " pod="swift-kuttl-tests/barbican-worker-8bf566dc7-r7nnf" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.411586 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/barbican-api-57985b48cb-dqxs9"] Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.412806 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/barbican-api-57985b48cb-dqxs9" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.415642 4943 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"barbican-api-config-data" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.426801 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/barbican-api-57985b48cb-dqxs9"] Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.470777 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2159c059-075c-4986-8244-bf6d57dc3cd9-config-data-custom\") pod \"barbican-worker-8bf566dc7-r7nnf\" (UID: \"2159c059-075c-4986-8244-bf6d57dc3cd9\") " pod="swift-kuttl-tests/barbican-worker-8bf566dc7-r7nnf" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.470818 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2159c059-075c-4986-8244-bf6d57dc3cd9-logs\") pod \"barbican-worker-8bf566dc7-r7nnf\" (UID: \"2159c059-075c-4986-8244-bf6d57dc3cd9\") " pod="swift-kuttl-tests/barbican-worker-8bf566dc7-r7nnf" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.470856 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d74297b6-b4e3-4dd8-8ee0-fd403bf79254-config-data-custom\") pod \"barbican-keystone-listener-54998b9498-bh574\" (UID: \"d74297b6-b4e3-4dd8-8ee0-fd403bf79254\") " pod="swift-kuttl-tests/barbican-keystone-listener-54998b9498-bh574" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.470880 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d74297b6-b4e3-4dd8-8ee0-fd403bf79254-logs\") pod \"barbican-keystone-listener-54998b9498-bh574\" (UID: \"d74297b6-b4e3-4dd8-8ee0-fd403bf79254\") " pod="swift-kuttl-tests/barbican-keystone-listener-54998b9498-bh574" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.470906 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2r6k\" (UniqueName: \"kubernetes.io/projected/6f106573-1c59-4abf-aeb2-30bacbd53b04-kube-api-access-h2r6k\") pod \"barbican-api-57985b48cb-dqxs9\" (UID: \"6f106573-1c59-4abf-aeb2-30bacbd53b04\") " pod="swift-kuttl-tests/barbican-api-57985b48cb-dqxs9" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.470983 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f106573-1c59-4abf-aeb2-30bacbd53b04-config-data\") pod \"barbican-api-57985b48cb-dqxs9\" (UID: \"6f106573-1c59-4abf-aeb2-30bacbd53b04\") " pod="swift-kuttl-tests/barbican-api-57985b48cb-dqxs9" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.470999 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f106573-1c59-4abf-aeb2-30bacbd53b04-logs\") pod \"barbican-api-57985b48cb-dqxs9\" (UID: \"6f106573-1c59-4abf-aeb2-30bacbd53b04\") " pod="swift-kuttl-tests/barbican-api-57985b48cb-dqxs9" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.471021 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2159c059-075c-4986-8244-bf6d57dc3cd9-config-data\") pod \"barbican-worker-8bf566dc7-r7nnf\" (UID: \"2159c059-075c-4986-8244-bf6d57dc3cd9\") " pod="swift-kuttl-tests/barbican-worker-8bf566dc7-r7nnf" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.471036 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgtcm\" (UniqueName: \"kubernetes.io/projected/d74297b6-b4e3-4dd8-8ee0-fd403bf79254-kube-api-access-cgtcm\") pod \"barbican-keystone-listener-54998b9498-bh574\" (UID: \"d74297b6-b4e3-4dd8-8ee0-fd403bf79254\") " pod="swift-kuttl-tests/barbican-keystone-listener-54998b9498-bh574" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.471058 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d74297b6-b4e3-4dd8-8ee0-fd403bf79254-config-data\") pod \"barbican-keystone-listener-54998b9498-bh574\" (UID: \"d74297b6-b4e3-4dd8-8ee0-fd403bf79254\") " pod="swift-kuttl-tests/barbican-keystone-listener-54998b9498-bh574" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.471083 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hx2qr\" (UniqueName: \"kubernetes.io/projected/2159c059-075c-4986-8244-bf6d57dc3cd9-kube-api-access-hx2qr\") pod \"barbican-worker-8bf566dc7-r7nnf\" (UID: \"2159c059-075c-4986-8244-bf6d57dc3cd9\") " pod="swift-kuttl-tests/barbican-worker-8bf566dc7-r7nnf" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.471099 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6f106573-1c59-4abf-aeb2-30bacbd53b04-config-data-custom\") pod \"barbican-api-57985b48cb-dqxs9\" (UID: \"6f106573-1c59-4abf-aeb2-30bacbd53b04\") " pod="swift-kuttl-tests/barbican-api-57985b48cb-dqxs9" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.472137 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2159c059-075c-4986-8244-bf6d57dc3cd9-logs\") pod \"barbican-worker-8bf566dc7-r7nnf\" (UID: \"2159c059-075c-4986-8244-bf6d57dc3cd9\") " pod="swift-kuttl-tests/barbican-worker-8bf566dc7-r7nnf" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.473373 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d74297b6-b4e3-4dd8-8ee0-fd403bf79254-logs\") pod \"barbican-keystone-listener-54998b9498-bh574\" (UID: \"d74297b6-b4e3-4dd8-8ee0-fd403bf79254\") " pod="swift-kuttl-tests/barbican-keystone-listener-54998b9498-bh574" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.476940 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d74297b6-b4e3-4dd8-8ee0-fd403bf79254-config-data-custom\") pod \"barbican-keystone-listener-54998b9498-bh574\" (UID: \"d74297b6-b4e3-4dd8-8ee0-fd403bf79254\") " pod="swift-kuttl-tests/barbican-keystone-listener-54998b9498-bh574" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.477463 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2159c059-075c-4986-8244-bf6d57dc3cd9-config-data-custom\") pod \"barbican-worker-8bf566dc7-r7nnf\" (UID: \"2159c059-075c-4986-8244-bf6d57dc3cd9\") " pod="swift-kuttl-tests/barbican-worker-8bf566dc7-r7nnf" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.481629 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="swift-kuttl-tests/keystone-5c4bccddcc-5vjzw" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.489090 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgtcm\" (UniqueName: \"kubernetes.io/projected/d74297b6-b4e3-4dd8-8ee0-fd403bf79254-kube-api-access-cgtcm\") pod \"barbican-keystone-listener-54998b9498-bh574\" (UID: \"d74297b6-b4e3-4dd8-8ee0-fd403bf79254\") " pod="swift-kuttl-tests/barbican-keystone-listener-54998b9498-bh574" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.491082 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2159c059-075c-4986-8244-bf6d57dc3cd9-config-data\") pod \"barbican-worker-8bf566dc7-r7nnf\" (UID: \"2159c059-075c-4986-8244-bf6d57dc3cd9\") " pod="swift-kuttl-tests/barbican-worker-8bf566dc7-r7nnf" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.493367 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d74297b6-b4e3-4dd8-8ee0-fd403bf79254-config-data\") pod \"barbican-keystone-listener-54998b9498-bh574\" (UID: \"d74297b6-b4e3-4dd8-8ee0-fd403bf79254\") " pod="swift-kuttl-tests/barbican-keystone-listener-54998b9498-bh574" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.502687 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hx2qr\" (UniqueName: \"kubernetes.io/projected/2159c059-075c-4986-8244-bf6d57dc3cd9-kube-api-access-hx2qr\") pod \"barbican-worker-8bf566dc7-r7nnf\" (UID: \"2159c059-075c-4986-8244-bf6d57dc3cd9\") " pod="swift-kuttl-tests/barbican-worker-8bf566dc7-r7nnf" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.572160 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2r6k\" (UniqueName: \"kubernetes.io/projected/6f106573-1c59-4abf-aeb2-30bacbd53b04-kube-api-access-h2r6k\") pod \"barbican-api-57985b48cb-dqxs9\" (UID: \"6f106573-1c59-4abf-aeb2-30bacbd53b04\") " pod="swift-kuttl-tests/barbican-api-57985b48cb-dqxs9" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.572253 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f106573-1c59-4abf-aeb2-30bacbd53b04-config-data\") pod \"barbican-api-57985b48cb-dqxs9\" (UID: \"6f106573-1c59-4abf-aeb2-30bacbd53b04\") " pod="swift-kuttl-tests/barbican-api-57985b48cb-dqxs9" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.572276 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f106573-1c59-4abf-aeb2-30bacbd53b04-logs\") pod \"barbican-api-57985b48cb-dqxs9\" (UID: \"6f106573-1c59-4abf-aeb2-30bacbd53b04\") " pod="swift-kuttl-tests/barbican-api-57985b48cb-dqxs9" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.572362 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6f106573-1c59-4abf-aeb2-30bacbd53b04-config-data-custom\") pod \"barbican-api-57985b48cb-dqxs9\" (UID: \"6f106573-1c59-4abf-aeb2-30bacbd53b04\") " pod="swift-kuttl-tests/barbican-api-57985b48cb-dqxs9" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.573146 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f106573-1c59-4abf-aeb2-30bacbd53b04-logs\") pod \"barbican-api-57985b48cb-dqxs9\" (UID: \"6f106573-1c59-4abf-aeb2-30bacbd53b04\") " pod="swift-kuttl-tests/barbican-api-57985b48cb-dqxs9" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.577165 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f106573-1c59-4abf-aeb2-30bacbd53b04-config-data\") pod \"barbican-api-57985b48cb-dqxs9\" (UID: \"6f106573-1c59-4abf-aeb2-30bacbd53b04\") " pod="swift-kuttl-tests/barbican-api-57985b48cb-dqxs9" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.584757 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6f106573-1c59-4abf-aeb2-30bacbd53b04-config-data-custom\") pod \"barbican-api-57985b48cb-dqxs9\" (UID: \"6f106573-1c59-4abf-aeb2-30bacbd53b04\") " pod="swift-kuttl-tests/barbican-api-57985b48cb-dqxs9" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.591334 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2r6k\" (UniqueName: \"kubernetes.io/projected/6f106573-1c59-4abf-aeb2-30bacbd53b04-kube-api-access-h2r6k\") pod \"barbican-api-57985b48cb-dqxs9\" (UID: \"6f106573-1c59-4abf-aeb2-30bacbd53b04\") " pod="swift-kuttl-tests/barbican-api-57985b48cb-dqxs9" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.608514 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/barbican-worker-8bf566dc7-r7nnf" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.620860 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/barbican-keystone-listener-54998b9498-bh574" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.732732 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/barbican-api-57985b48cb-dqxs9" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.927310 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/8e67c57d4916ae7e348662cb38dd5155bd37a096e90f723dff4c6ec243ptqwj"] Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.928730 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/8e67c57d4916ae7e348662cb38dd5155bd37a096e90f723dff4c6ec243ptqwj" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.931344 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-k6gp2" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.939918 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/8e67c57d4916ae7e348662cb38dd5155bd37a096e90f723dff4c6ec243ptqwj"] Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.968781 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/barbican-keystone-listener-54998b9498-bh574"] Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.978665 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7f65812b-f0ff-4e87-845d-eafb6cc93743-util\") pod \"8e67c57d4916ae7e348662cb38dd5155bd37a096e90f723dff4c6ec243ptqwj\" (UID: \"7f65812b-f0ff-4e87-845d-eafb6cc93743\") " pod="openstack-operators/8e67c57d4916ae7e348662cb38dd5155bd37a096e90f723dff4c6ec243ptqwj" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.978726 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7f65812b-f0ff-4e87-845d-eafb6cc93743-bundle\") pod \"8e67c57d4916ae7e348662cb38dd5155bd37a096e90f723dff4c6ec243ptqwj\" (UID: \"7f65812b-f0ff-4e87-845d-eafb6cc93743\") " pod="openstack-operators/8e67c57d4916ae7e348662cb38dd5155bd37a096e90f723dff4c6ec243ptqwj" Mar 07 14:56:37 crc kubenswrapper[4943]: I0307 14:56:37.978803 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vnrr\" (UniqueName: \"kubernetes.io/projected/7f65812b-f0ff-4e87-845d-eafb6cc93743-kube-api-access-9vnrr\") pod \"8e67c57d4916ae7e348662cb38dd5155bd37a096e90f723dff4c6ec243ptqwj\" (UID: \"7f65812b-f0ff-4e87-845d-eafb6cc93743\") " pod="openstack-operators/8e67c57d4916ae7e348662cb38dd5155bd37a096e90f723dff4c6ec243ptqwj" Mar 07 14:56:38 crc kubenswrapper[4943]: I0307 14:56:38.045170 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/barbican-worker-8bf566dc7-r7nnf"] Mar 07 14:56:38 crc kubenswrapper[4943]: W0307 14:56:38.047324 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2159c059_075c_4986_8244_bf6d57dc3cd9.slice/crio-6eaf9093f6ca24b1c22edc141ee449e55049ca716cef30acee5515e246e43b10 WatchSource:0}: Error finding container 6eaf9093f6ca24b1c22edc141ee449e55049ca716cef30acee5515e246e43b10: Status 404 returned error can't find the container with id 6eaf9093f6ca24b1c22edc141ee449e55049ca716cef30acee5515e246e43b10 Mar 07 14:56:38 crc kubenswrapper[4943]: I0307 14:56:38.080207 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7f65812b-f0ff-4e87-845d-eafb6cc93743-util\") pod \"8e67c57d4916ae7e348662cb38dd5155bd37a096e90f723dff4c6ec243ptqwj\" (UID: \"7f65812b-f0ff-4e87-845d-eafb6cc93743\") " pod="openstack-operators/8e67c57d4916ae7e348662cb38dd5155bd37a096e90f723dff4c6ec243ptqwj" Mar 07 14:56:38 crc kubenswrapper[4943]: I0307 14:56:38.080263 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7f65812b-f0ff-4e87-845d-eafb6cc93743-bundle\") pod \"8e67c57d4916ae7e348662cb38dd5155bd37a096e90f723dff4c6ec243ptqwj\" (UID: \"7f65812b-f0ff-4e87-845d-eafb6cc93743\") " pod="openstack-operators/8e67c57d4916ae7e348662cb38dd5155bd37a096e90f723dff4c6ec243ptqwj" Mar 07 14:56:38 crc kubenswrapper[4943]: I0307 14:56:38.080350 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vnrr\" (UniqueName: \"kubernetes.io/projected/7f65812b-f0ff-4e87-845d-eafb6cc93743-kube-api-access-9vnrr\") pod \"8e67c57d4916ae7e348662cb38dd5155bd37a096e90f723dff4c6ec243ptqwj\" (UID: \"7f65812b-f0ff-4e87-845d-eafb6cc93743\") " pod="openstack-operators/8e67c57d4916ae7e348662cb38dd5155bd37a096e90f723dff4c6ec243ptqwj" Mar 07 14:56:38 crc kubenswrapper[4943]: I0307 14:56:38.081309 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7f65812b-f0ff-4e87-845d-eafb6cc93743-bundle\") pod \"8e67c57d4916ae7e348662cb38dd5155bd37a096e90f723dff4c6ec243ptqwj\" (UID: \"7f65812b-f0ff-4e87-845d-eafb6cc93743\") " pod="openstack-operators/8e67c57d4916ae7e348662cb38dd5155bd37a096e90f723dff4c6ec243ptqwj" Mar 07 14:56:38 crc kubenswrapper[4943]: I0307 14:56:38.081331 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7f65812b-f0ff-4e87-845d-eafb6cc93743-util\") pod \"8e67c57d4916ae7e348662cb38dd5155bd37a096e90f723dff4c6ec243ptqwj\" (UID: \"7f65812b-f0ff-4e87-845d-eafb6cc93743\") " pod="openstack-operators/8e67c57d4916ae7e348662cb38dd5155bd37a096e90f723dff4c6ec243ptqwj" Mar 07 14:56:38 crc kubenswrapper[4943]: I0307 14:56:38.097146 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vnrr\" (UniqueName: \"kubernetes.io/projected/7f65812b-f0ff-4e87-845d-eafb6cc93743-kube-api-access-9vnrr\") pod \"8e67c57d4916ae7e348662cb38dd5155bd37a096e90f723dff4c6ec243ptqwj\" (UID: \"7f65812b-f0ff-4e87-845d-eafb6cc93743\") " pod="openstack-operators/8e67c57d4916ae7e348662cb38dd5155bd37a096e90f723dff4c6ec243ptqwj" Mar 07 14:56:38 crc kubenswrapper[4943]: I0307 14:56:38.120052 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/barbican-keystone-listener-54998b9498-bh574" event={"ID":"d74297b6-b4e3-4dd8-8ee0-fd403bf79254","Type":"ContainerStarted","Data":"fa7243f7d739cb18ab80c487045224b6a6df7a5156dad5727839d1d564f19901"} Mar 07 14:56:38 crc kubenswrapper[4943]: I0307 14:56:38.121980 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/barbican-worker-8bf566dc7-r7nnf" event={"ID":"2159c059-075c-4986-8244-bf6d57dc3cd9","Type":"ContainerStarted","Data":"6eaf9093f6ca24b1c22edc141ee449e55049ca716cef30acee5515e246e43b10"} Mar 07 14:56:38 crc kubenswrapper[4943]: I0307 14:56:38.236292 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/barbican-api-57985b48cb-dqxs9"] Mar 07 14:56:38 crc kubenswrapper[4943]: W0307 14:56:38.247228 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6f106573_1c59_4abf_aeb2_30bacbd53b04.slice/crio-d1f5276dc600be40b8a67139768aafb6f0f2a78125e6fe1515f9387466b1a05c WatchSource:0}: Error finding container d1f5276dc600be40b8a67139768aafb6f0f2a78125e6fe1515f9387466b1a05c: Status 404 returned error can't find the container with id d1f5276dc600be40b8a67139768aafb6f0f2a78125e6fe1515f9387466b1a05c Mar 07 14:56:38 crc kubenswrapper[4943]: I0307 14:56:38.247287 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/8e67c57d4916ae7e348662cb38dd5155bd37a096e90f723dff4c6ec243ptqwj" Mar 07 14:56:38 crc kubenswrapper[4943]: I0307 14:56:38.561085 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/8e67c57d4916ae7e348662cb38dd5155bd37a096e90f723dff4c6ec243ptqwj"] Mar 07 14:56:39 crc kubenswrapper[4943]: I0307 14:56:39.128917 4943 generic.go:334] "Generic (PLEG): container finished" podID="7f65812b-f0ff-4e87-845d-eafb6cc93743" containerID="3f1cf3d4817f2c393551e21a38dd5208c14ab3b8d2eaa37b681029756956a9ea" exitCode=0 Mar 07 14:56:39 crc kubenswrapper[4943]: I0307 14:56:39.129015 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/8e67c57d4916ae7e348662cb38dd5155bd37a096e90f723dff4c6ec243ptqwj" event={"ID":"7f65812b-f0ff-4e87-845d-eafb6cc93743","Type":"ContainerDied","Data":"3f1cf3d4817f2c393551e21a38dd5208c14ab3b8d2eaa37b681029756956a9ea"} Mar 07 14:56:39 crc kubenswrapper[4943]: I0307 14:56:39.129379 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/8e67c57d4916ae7e348662cb38dd5155bd37a096e90f723dff4c6ec243ptqwj" event={"ID":"7f65812b-f0ff-4e87-845d-eafb6cc93743","Type":"ContainerStarted","Data":"59c6b2ddb14a619642ad8c45e0656b7f68f89e102c08a78dc299c4aa71e83976"} Mar 07 14:56:39 crc kubenswrapper[4943]: I0307 14:56:39.131602 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/barbican-api-57985b48cb-dqxs9" event={"ID":"6f106573-1c59-4abf-aeb2-30bacbd53b04","Type":"ContainerStarted","Data":"13d2ee653dc596f3a3d900ee562c4e13937053cce81f81bbd406efbe20f375b7"} Mar 07 14:56:39 crc kubenswrapper[4943]: I0307 14:56:39.131675 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/barbican-api-57985b48cb-dqxs9" event={"ID":"6f106573-1c59-4abf-aeb2-30bacbd53b04","Type":"ContainerStarted","Data":"fde5d3d3b001d952d781696bb7966112f3e2d9e03bb4dd355b072c50038a308e"} Mar 07 14:56:39 crc kubenswrapper[4943]: I0307 14:56:39.131700 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/barbican-api-57985b48cb-dqxs9" event={"ID":"6f106573-1c59-4abf-aeb2-30bacbd53b04","Type":"ContainerStarted","Data":"d1f5276dc600be40b8a67139768aafb6f0f2a78125e6fe1515f9387466b1a05c"} Mar 07 14:56:39 crc kubenswrapper[4943]: I0307 14:56:39.131749 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="swift-kuttl-tests/barbican-api-57985b48cb-dqxs9" Mar 07 14:56:39 crc kubenswrapper[4943]: I0307 14:56:39.167858 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/barbican-api-57985b48cb-dqxs9" podStartSLOduration=2.167840199 podStartE2EDuration="2.167840199s" podCreationTimestamp="2026-03-07 14:56:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:56:39.16382889 +0000 UTC m=+1041.115965388" watchObservedRunningTime="2026-03-07 14:56:39.167840199 +0000 UTC m=+1041.119976697" Mar 07 14:56:40 crc kubenswrapper[4943]: I0307 14:56:40.188592 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/barbican-keystone-listener-54998b9498-bh574" event={"ID":"d74297b6-b4e3-4dd8-8ee0-fd403bf79254","Type":"ContainerStarted","Data":"13c24a570b790ca41c69f239672b0ac58872ac085fd94221e80e5b8b5706ac3f"} Mar 07 14:56:40 crc kubenswrapper[4943]: I0307 14:56:40.241663 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/barbican-worker-8bf566dc7-r7nnf" event={"ID":"2159c059-075c-4986-8244-bf6d57dc3cd9","Type":"ContainerStarted","Data":"48509b6abc3593b122f60a7e0df6b62f712a73220b9c23d77e6cc5081b4ad6bc"} Mar 07 14:56:40 crc kubenswrapper[4943]: I0307 14:56:40.242316 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="swift-kuttl-tests/barbican-api-57985b48cb-dqxs9" Mar 07 14:56:41 crc kubenswrapper[4943]: I0307 14:56:41.258532 4943 generic.go:334] "Generic (PLEG): container finished" podID="7f65812b-f0ff-4e87-845d-eafb6cc93743" containerID="72692047f4c6a855bebbd29e5b1536491bc044cb640636a5c1dbe91847f68cc4" exitCode=0 Mar 07 14:56:41 crc kubenswrapper[4943]: I0307 14:56:41.258634 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/8e67c57d4916ae7e348662cb38dd5155bd37a096e90f723dff4c6ec243ptqwj" event={"ID":"7f65812b-f0ff-4e87-845d-eafb6cc93743","Type":"ContainerDied","Data":"72692047f4c6a855bebbd29e5b1536491bc044cb640636a5c1dbe91847f68cc4"} Mar 07 14:56:41 crc kubenswrapper[4943]: I0307 14:56:41.264147 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/barbican-worker-8bf566dc7-r7nnf" event={"ID":"2159c059-075c-4986-8244-bf6d57dc3cd9","Type":"ContainerStarted","Data":"8a067be7c8469e9d76ae7bcf609c62459fe88e10789660117cabb5dfe39052c3"} Mar 07 14:56:41 crc kubenswrapper[4943]: I0307 14:56:41.272251 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/barbican-keystone-listener-54998b9498-bh574" event={"ID":"d74297b6-b4e3-4dd8-8ee0-fd403bf79254","Type":"ContainerStarted","Data":"1b9d6f4196332fb4fa0f07a2c9e88444033b2abb9b0b42be40fab463b613fee0"} Mar 07 14:56:41 crc kubenswrapper[4943]: I0307 14:56:41.311116 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/barbican-worker-8bf566dc7-r7nnf" podStartSLOduration=2.808688237 podStartE2EDuration="4.311098112s" podCreationTimestamp="2026-03-07 14:56:37 +0000 UTC" firstStartedPulling="2026-03-07 14:56:38.049768034 +0000 UTC m=+1040.001904532" lastFinishedPulling="2026-03-07 14:56:39.552177909 +0000 UTC m=+1041.504314407" observedRunningTime="2026-03-07 14:56:41.309468152 +0000 UTC m=+1043.261604680" watchObservedRunningTime="2026-03-07 14:56:41.311098112 +0000 UTC m=+1043.263234610" Mar 07 14:56:41 crc kubenswrapper[4943]: I0307 14:56:41.336330 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/barbican-keystone-listener-54998b9498-bh574" podStartSLOduration=2.7495045019999997 podStartE2EDuration="4.336309422s" podCreationTimestamp="2026-03-07 14:56:37 +0000 UTC" firstStartedPulling="2026-03-07 14:56:37.9605494 +0000 UTC m=+1039.912685898" lastFinishedPulling="2026-03-07 14:56:39.54735432 +0000 UTC m=+1041.499490818" observedRunningTime="2026-03-07 14:56:41.329667179 +0000 UTC m=+1043.281803677" watchObservedRunningTime="2026-03-07 14:56:41.336309422 +0000 UTC m=+1043.288445930" Mar 07 14:56:42 crc kubenswrapper[4943]: I0307 14:56:42.282607 4943 generic.go:334] "Generic (PLEG): container finished" podID="7f65812b-f0ff-4e87-845d-eafb6cc93743" containerID="b9c7657c6fa4f87c53c62a2ece45cc762f9f75fc9e3b39240d05551d63a338b9" exitCode=0 Mar 07 14:56:42 crc kubenswrapper[4943]: I0307 14:56:42.282697 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/8e67c57d4916ae7e348662cb38dd5155bd37a096e90f723dff4c6ec243ptqwj" event={"ID":"7f65812b-f0ff-4e87-845d-eafb6cc93743","Type":"ContainerDied","Data":"b9c7657c6fa4f87c53c62a2ece45cc762f9f75fc9e3b39240d05551d63a338b9"} Mar 07 14:56:43 crc kubenswrapper[4943]: I0307 14:56:43.715633 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/8e67c57d4916ae7e348662cb38dd5155bd37a096e90f723dff4c6ec243ptqwj" Mar 07 14:56:43 crc kubenswrapper[4943]: I0307 14:56:43.888639 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7f65812b-f0ff-4e87-845d-eafb6cc93743-util\") pod \"7f65812b-f0ff-4e87-845d-eafb6cc93743\" (UID: \"7f65812b-f0ff-4e87-845d-eafb6cc93743\") " Mar 07 14:56:43 crc kubenswrapper[4943]: I0307 14:56:43.895514 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7f65812b-f0ff-4e87-845d-eafb6cc93743-bundle\") pod \"7f65812b-f0ff-4e87-845d-eafb6cc93743\" (UID: \"7f65812b-f0ff-4e87-845d-eafb6cc93743\") " Mar 07 14:56:43 crc kubenswrapper[4943]: I0307 14:56:43.895600 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9vnrr\" (UniqueName: \"kubernetes.io/projected/7f65812b-f0ff-4e87-845d-eafb6cc93743-kube-api-access-9vnrr\") pod \"7f65812b-f0ff-4e87-845d-eafb6cc93743\" (UID: \"7f65812b-f0ff-4e87-845d-eafb6cc93743\") " Mar 07 14:56:43 crc kubenswrapper[4943]: I0307 14:56:43.896531 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f65812b-f0ff-4e87-845d-eafb6cc93743-bundle" (OuterVolumeSpecName: "bundle") pod "7f65812b-f0ff-4e87-845d-eafb6cc93743" (UID: "7f65812b-f0ff-4e87-845d-eafb6cc93743"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:56:43 crc kubenswrapper[4943]: I0307 14:56:43.909571 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f65812b-f0ff-4e87-845d-eafb6cc93743-util" (OuterVolumeSpecName: "util") pod "7f65812b-f0ff-4e87-845d-eafb6cc93743" (UID: "7f65812b-f0ff-4e87-845d-eafb6cc93743"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:56:43 crc kubenswrapper[4943]: I0307 14:56:43.912436 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f65812b-f0ff-4e87-845d-eafb6cc93743-kube-api-access-9vnrr" (OuterVolumeSpecName: "kube-api-access-9vnrr") pod "7f65812b-f0ff-4e87-845d-eafb6cc93743" (UID: "7f65812b-f0ff-4e87-845d-eafb6cc93743"). InnerVolumeSpecName "kube-api-access-9vnrr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:56:43 crc kubenswrapper[4943]: I0307 14:56:43.999194 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9vnrr\" (UniqueName: \"kubernetes.io/projected/7f65812b-f0ff-4e87-845d-eafb6cc93743-kube-api-access-9vnrr\") on node \"crc\" DevicePath \"\"" Mar 07 14:56:43 crc kubenswrapper[4943]: I0307 14:56:43.999230 4943 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7f65812b-f0ff-4e87-845d-eafb6cc93743-util\") on node \"crc\" DevicePath \"\"" Mar 07 14:56:43 crc kubenswrapper[4943]: I0307 14:56:43.999243 4943 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7f65812b-f0ff-4e87-845d-eafb6cc93743-bundle\") on node \"crc\" DevicePath \"\"" Mar 07 14:56:44 crc kubenswrapper[4943]: I0307 14:56:44.301995 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/8e67c57d4916ae7e348662cb38dd5155bd37a096e90f723dff4c6ec243ptqwj" event={"ID":"7f65812b-f0ff-4e87-845d-eafb6cc93743","Type":"ContainerDied","Data":"59c6b2ddb14a619642ad8c45e0656b7f68f89e102c08a78dc299c4aa71e83976"} Mar 07 14:56:44 crc kubenswrapper[4943]: I0307 14:56:44.302376 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="59c6b2ddb14a619642ad8c45e0656b7f68f89e102c08a78dc299c4aa71e83976" Mar 07 14:56:44 crc kubenswrapper[4943]: I0307 14:56:44.302069 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/8e67c57d4916ae7e348662cb38dd5155bd37a096e90f723dff4c6ec243ptqwj" Mar 07 14:56:49 crc kubenswrapper[4943]: I0307 14:56:49.021961 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="swift-kuttl-tests/barbican-api-57985b48cb-dqxs9" Mar 07 14:56:49 crc kubenswrapper[4943]: I0307 14:56:49.299551 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="swift-kuttl-tests/barbican-api-57985b48cb-dqxs9" Mar 07 14:56:54 crc kubenswrapper[4943]: I0307 14:56:54.227014 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-9c8b7c9f6-wcwvx"] Mar 07 14:56:54 crc kubenswrapper[4943]: E0307 14:56:54.228043 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f65812b-f0ff-4e87-845d-eafb6cc93743" containerName="extract" Mar 07 14:56:54 crc kubenswrapper[4943]: I0307 14:56:54.228057 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f65812b-f0ff-4e87-845d-eafb6cc93743" containerName="extract" Mar 07 14:56:54 crc kubenswrapper[4943]: E0307 14:56:54.228070 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f65812b-f0ff-4e87-845d-eafb6cc93743" containerName="pull" Mar 07 14:56:54 crc kubenswrapper[4943]: I0307 14:56:54.228077 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f65812b-f0ff-4e87-845d-eafb6cc93743" containerName="pull" Mar 07 14:56:54 crc kubenswrapper[4943]: E0307 14:56:54.228096 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f65812b-f0ff-4e87-845d-eafb6cc93743" containerName="util" Mar 07 14:56:54 crc kubenswrapper[4943]: I0307 14:56:54.228102 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f65812b-f0ff-4e87-845d-eafb6cc93743" containerName="util" Mar 07 14:56:54 crc kubenswrapper[4943]: I0307 14:56:54.228216 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f65812b-f0ff-4e87-845d-eafb6cc93743" containerName="extract" Mar 07 14:56:54 crc kubenswrapper[4943]: I0307 14:56:54.228668 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-9c8b7c9f6-wcwvx" Mar 07 14:56:54 crc kubenswrapper[4943]: I0307 14:56:54.231504 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-2fpwb" Mar 07 14:56:54 crc kubenswrapper[4943]: I0307 14:56:54.232326 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-service-cert" Mar 07 14:56:54 crc kubenswrapper[4943]: I0307 14:56:54.292965 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-9c8b7c9f6-wcwvx"] Mar 07 14:56:54 crc kubenswrapper[4943]: I0307 14:56:54.364406 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5p5b\" (UniqueName: \"kubernetes.io/projected/125fcec4-ea6c-4199-b0cd-68bcd04c4580-kube-api-access-l5p5b\") pod \"swift-operator-controller-manager-9c8b7c9f6-wcwvx\" (UID: \"125fcec4-ea6c-4199-b0cd-68bcd04c4580\") " pod="openstack-operators/swift-operator-controller-manager-9c8b7c9f6-wcwvx" Mar 07 14:56:54 crc kubenswrapper[4943]: I0307 14:56:54.364477 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/125fcec4-ea6c-4199-b0cd-68bcd04c4580-webhook-cert\") pod \"swift-operator-controller-manager-9c8b7c9f6-wcwvx\" (UID: \"125fcec4-ea6c-4199-b0cd-68bcd04c4580\") " pod="openstack-operators/swift-operator-controller-manager-9c8b7c9f6-wcwvx" Mar 07 14:56:54 crc kubenswrapper[4943]: I0307 14:56:54.364528 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/125fcec4-ea6c-4199-b0cd-68bcd04c4580-apiservice-cert\") pod \"swift-operator-controller-manager-9c8b7c9f6-wcwvx\" (UID: \"125fcec4-ea6c-4199-b0cd-68bcd04c4580\") " pod="openstack-operators/swift-operator-controller-manager-9c8b7c9f6-wcwvx" Mar 07 14:56:54 crc kubenswrapper[4943]: I0307 14:56:54.465786 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5p5b\" (UniqueName: \"kubernetes.io/projected/125fcec4-ea6c-4199-b0cd-68bcd04c4580-kube-api-access-l5p5b\") pod \"swift-operator-controller-manager-9c8b7c9f6-wcwvx\" (UID: \"125fcec4-ea6c-4199-b0cd-68bcd04c4580\") " pod="openstack-operators/swift-operator-controller-manager-9c8b7c9f6-wcwvx" Mar 07 14:56:54 crc kubenswrapper[4943]: I0307 14:56:54.465887 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/125fcec4-ea6c-4199-b0cd-68bcd04c4580-webhook-cert\") pod \"swift-operator-controller-manager-9c8b7c9f6-wcwvx\" (UID: \"125fcec4-ea6c-4199-b0cd-68bcd04c4580\") " pod="openstack-operators/swift-operator-controller-manager-9c8b7c9f6-wcwvx" Mar 07 14:56:54 crc kubenswrapper[4943]: I0307 14:56:54.466054 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/125fcec4-ea6c-4199-b0cd-68bcd04c4580-apiservice-cert\") pod \"swift-operator-controller-manager-9c8b7c9f6-wcwvx\" (UID: \"125fcec4-ea6c-4199-b0cd-68bcd04c4580\") " pod="openstack-operators/swift-operator-controller-manager-9c8b7c9f6-wcwvx" Mar 07 14:56:54 crc kubenswrapper[4943]: I0307 14:56:54.474677 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/125fcec4-ea6c-4199-b0cd-68bcd04c4580-apiservice-cert\") pod \"swift-operator-controller-manager-9c8b7c9f6-wcwvx\" (UID: \"125fcec4-ea6c-4199-b0cd-68bcd04c4580\") " pod="openstack-operators/swift-operator-controller-manager-9c8b7c9f6-wcwvx" Mar 07 14:56:54 crc kubenswrapper[4943]: I0307 14:56:54.483200 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5p5b\" (UniqueName: \"kubernetes.io/projected/125fcec4-ea6c-4199-b0cd-68bcd04c4580-kube-api-access-l5p5b\") pod \"swift-operator-controller-manager-9c8b7c9f6-wcwvx\" (UID: \"125fcec4-ea6c-4199-b0cd-68bcd04c4580\") " pod="openstack-operators/swift-operator-controller-manager-9c8b7c9f6-wcwvx" Mar 07 14:56:54 crc kubenswrapper[4943]: I0307 14:56:54.490457 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/125fcec4-ea6c-4199-b0cd-68bcd04c4580-webhook-cert\") pod \"swift-operator-controller-manager-9c8b7c9f6-wcwvx\" (UID: \"125fcec4-ea6c-4199-b0cd-68bcd04c4580\") " pod="openstack-operators/swift-operator-controller-manager-9c8b7c9f6-wcwvx" Mar 07 14:56:54 crc kubenswrapper[4943]: I0307 14:56:54.547568 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-9c8b7c9f6-wcwvx" Mar 07 14:56:54 crc kubenswrapper[4943]: I0307 14:56:54.973263 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-9c8b7c9f6-wcwvx"] Mar 07 14:56:54 crc kubenswrapper[4943]: W0307 14:56:54.979965 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod125fcec4_ea6c_4199_b0cd_68bcd04c4580.slice/crio-0f1a3104543a2d300d8796be4f675ca4ee47a60a2ea8baa689bf865a237d428b WatchSource:0}: Error finding container 0f1a3104543a2d300d8796be4f675ca4ee47a60a2ea8baa689bf865a237d428b: Status 404 returned error can't find the container with id 0f1a3104543a2d300d8796be4f675ca4ee47a60a2ea8baa689bf865a237d428b Mar 07 14:56:55 crc kubenswrapper[4943]: I0307 14:56:55.406661 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9c8b7c9f6-wcwvx" event={"ID":"125fcec4-ea6c-4199-b0cd-68bcd04c4580","Type":"ContainerStarted","Data":"0f1a3104543a2d300d8796be4f675ca4ee47a60a2ea8baa689bf865a237d428b"} Mar 07 14:56:57 crc kubenswrapper[4943]: I0307 14:56:57.423967 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9c8b7c9f6-wcwvx" event={"ID":"125fcec4-ea6c-4199-b0cd-68bcd04c4580","Type":"ContainerStarted","Data":"3d5e8baebbeaac29ac758a34e9ac19481bcd427188e335f95acf742a1c933e06"} Mar 07 14:56:57 crc kubenswrapper[4943]: I0307 14:56:57.424295 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-9c8b7c9f6-wcwvx" Mar 07 14:56:57 crc kubenswrapper[4943]: I0307 14:56:57.465216 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-9c8b7c9f6-wcwvx" podStartSLOduration=1.959088445 podStartE2EDuration="3.465185331s" podCreationTimestamp="2026-03-07 14:56:54 +0000 UTC" firstStartedPulling="2026-03-07 14:56:54.984453739 +0000 UTC m=+1056.936590237" lastFinishedPulling="2026-03-07 14:56:56.490550585 +0000 UTC m=+1058.442687123" observedRunningTime="2026-03-07 14:56:57.455974105 +0000 UTC m=+1059.408110623" watchObservedRunningTime="2026-03-07 14:56:57.465185331 +0000 UTC m=+1059.417321869" Mar 07 14:57:04 crc kubenswrapper[4943]: I0307 14:57:04.557406 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-9c8b7c9f6-wcwvx" Mar 07 14:57:07 crc kubenswrapper[4943]: I0307 14:57:07.052013 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-storage-0"] Mar 07 14:57:07 crc kubenswrapper[4943]: I0307 14:57:07.057794 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-storage-0" Mar 07 14:57:07 crc kubenswrapper[4943]: I0307 14:57:07.061211 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-storage-config-data" Mar 07 14:57:07 crc kubenswrapper[4943]: I0307 14:57:07.061348 4943 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"swift-swift-dockercfg-87bfx" Mar 07 14:57:07 crc kubenswrapper[4943]: I0307 14:57:07.061435 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-files" Mar 07 14:57:07 crc kubenswrapper[4943]: I0307 14:57:07.066154 4943 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"swift-conf" Mar 07 14:57:07 crc kubenswrapper[4943]: I0307 14:57:07.071991 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-storage-0"] Mar 07 14:57:07 crc kubenswrapper[4943]: I0307 14:57:07.168793 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/61bfc108-bcc1-4712-b8c8-8bc58b22777a-lock\") pod \"swift-storage-0\" (UID: \"61bfc108-bcc1-4712-b8c8-8bc58b22777a\") " pod="swift-kuttl-tests/swift-storage-0" Mar 07 14:57:07 crc kubenswrapper[4943]: I0307 14:57:07.169293 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/61bfc108-bcc1-4712-b8c8-8bc58b22777a-cache\") pod \"swift-storage-0\" (UID: \"61bfc108-bcc1-4712-b8c8-8bc58b22777a\") " pod="swift-kuttl-tests/swift-storage-0" Mar 07 14:57:07 crc kubenswrapper[4943]: I0307 14:57:07.169485 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kksv6\" (UniqueName: \"kubernetes.io/projected/61bfc108-bcc1-4712-b8c8-8bc58b22777a-kube-api-access-kksv6\") pod \"swift-storage-0\" (UID: \"61bfc108-bcc1-4712-b8c8-8bc58b22777a\") " pod="swift-kuttl-tests/swift-storage-0" Mar 07 14:57:07 crc kubenswrapper[4943]: I0307 14:57:07.169604 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/61bfc108-bcc1-4712-b8c8-8bc58b22777a-etc-swift\") pod \"swift-storage-0\" (UID: \"61bfc108-bcc1-4712-b8c8-8bc58b22777a\") " pod="swift-kuttl-tests/swift-storage-0" Mar 07 14:57:07 crc kubenswrapper[4943]: I0307 14:57:07.169741 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"61bfc108-bcc1-4712-b8c8-8bc58b22777a\") " pod="swift-kuttl-tests/swift-storage-0" Mar 07 14:57:07 crc kubenswrapper[4943]: I0307 14:57:07.271512 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/61bfc108-bcc1-4712-b8c8-8bc58b22777a-lock\") pod \"swift-storage-0\" (UID: \"61bfc108-bcc1-4712-b8c8-8bc58b22777a\") " pod="swift-kuttl-tests/swift-storage-0" Mar 07 14:57:07 crc kubenswrapper[4943]: I0307 14:57:07.271608 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/61bfc108-bcc1-4712-b8c8-8bc58b22777a-cache\") pod \"swift-storage-0\" (UID: \"61bfc108-bcc1-4712-b8c8-8bc58b22777a\") " pod="swift-kuttl-tests/swift-storage-0" Mar 07 14:57:07 crc kubenswrapper[4943]: I0307 14:57:07.271655 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kksv6\" (UniqueName: \"kubernetes.io/projected/61bfc108-bcc1-4712-b8c8-8bc58b22777a-kube-api-access-kksv6\") pod \"swift-storage-0\" (UID: \"61bfc108-bcc1-4712-b8c8-8bc58b22777a\") " pod="swift-kuttl-tests/swift-storage-0" Mar 07 14:57:07 crc kubenswrapper[4943]: I0307 14:57:07.271672 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/61bfc108-bcc1-4712-b8c8-8bc58b22777a-etc-swift\") pod \"swift-storage-0\" (UID: \"61bfc108-bcc1-4712-b8c8-8bc58b22777a\") " pod="swift-kuttl-tests/swift-storage-0" Mar 07 14:57:07 crc kubenswrapper[4943]: I0307 14:57:07.271696 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"61bfc108-bcc1-4712-b8c8-8bc58b22777a\") " pod="swift-kuttl-tests/swift-storage-0" Mar 07 14:57:07 crc kubenswrapper[4943]: I0307 14:57:07.272145 4943 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"61bfc108-bcc1-4712-b8c8-8bc58b22777a\") device mount path \"/mnt/openstack/pv03\"" pod="swift-kuttl-tests/swift-storage-0" Mar 07 14:57:07 crc kubenswrapper[4943]: I0307 14:57:07.273251 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/61bfc108-bcc1-4712-b8c8-8bc58b22777a-lock\") pod \"swift-storage-0\" (UID: \"61bfc108-bcc1-4712-b8c8-8bc58b22777a\") " pod="swift-kuttl-tests/swift-storage-0" Mar 07 14:57:07 crc kubenswrapper[4943]: E0307 14:57:07.273442 4943 projected.go:288] Couldn't get configMap swift-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Mar 07 14:57:07 crc kubenswrapper[4943]: E0307 14:57:07.273463 4943 projected.go:194] Error preparing data for projected volume etc-swift for pod swift-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Mar 07 14:57:07 crc kubenswrapper[4943]: E0307 14:57:07.273517 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/61bfc108-bcc1-4712-b8c8-8bc58b22777a-etc-swift podName:61bfc108-bcc1-4712-b8c8-8bc58b22777a nodeName:}" failed. No retries permitted until 2026-03-07 14:57:07.773494314 +0000 UTC m=+1069.725630812 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/61bfc108-bcc1-4712-b8c8-8bc58b22777a-etc-swift") pod "swift-storage-0" (UID: "61bfc108-bcc1-4712-b8c8-8bc58b22777a") : configmap "swift-ring-files" not found Mar 07 14:57:07 crc kubenswrapper[4943]: I0307 14:57:07.273919 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/61bfc108-bcc1-4712-b8c8-8bc58b22777a-cache\") pod \"swift-storage-0\" (UID: \"61bfc108-bcc1-4712-b8c8-8bc58b22777a\") " pod="swift-kuttl-tests/swift-storage-0" Mar 07 14:57:07 crc kubenswrapper[4943]: I0307 14:57:07.301005 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"61bfc108-bcc1-4712-b8c8-8bc58b22777a\") " pod="swift-kuttl-tests/swift-storage-0" Mar 07 14:57:07 crc kubenswrapper[4943]: I0307 14:57:07.302030 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kksv6\" (UniqueName: \"kubernetes.io/projected/61bfc108-bcc1-4712-b8c8-8bc58b22777a-kube-api-access-kksv6\") pod \"swift-storage-0\" (UID: \"61bfc108-bcc1-4712-b8c8-8bc58b22777a\") " pod="swift-kuttl-tests/swift-storage-0" Mar 07 14:57:07 crc kubenswrapper[4943]: I0307 14:57:07.557314 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-proxy-76c998454c-9tgn8"] Mar 07 14:57:07 crc kubenswrapper[4943]: I0307 14:57:07.559011 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-proxy-76c998454c-9tgn8" Mar 07 14:57:07 crc kubenswrapper[4943]: I0307 14:57:07.564809 4943 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"swift-proxy-config-data" Mar 07 14:57:07 crc kubenswrapper[4943]: I0307 14:57:07.570798 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-proxy-76c998454c-9tgn8"] Mar 07 14:57:07 crc kubenswrapper[4943]: I0307 14:57:07.679329 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/81fd2fdf-28bf-40a5-92a9-8506339ed373-log-httpd\") pod \"swift-proxy-76c998454c-9tgn8\" (UID: \"81fd2fdf-28bf-40a5-92a9-8506339ed373\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-9tgn8" Mar 07 14:57:07 crc kubenswrapper[4943]: I0307 14:57:07.679387 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81fd2fdf-28bf-40a5-92a9-8506339ed373-config-data\") pod \"swift-proxy-76c998454c-9tgn8\" (UID: \"81fd2fdf-28bf-40a5-92a9-8506339ed373\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-9tgn8" Mar 07 14:57:07 crc kubenswrapper[4943]: I0307 14:57:07.679423 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/81fd2fdf-28bf-40a5-92a9-8506339ed373-etc-swift\") pod \"swift-proxy-76c998454c-9tgn8\" (UID: \"81fd2fdf-28bf-40a5-92a9-8506339ed373\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-9tgn8" Mar 07 14:57:07 crc kubenswrapper[4943]: I0307 14:57:07.679439 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/81fd2fdf-28bf-40a5-92a9-8506339ed373-run-httpd\") pod \"swift-proxy-76c998454c-9tgn8\" (UID: \"81fd2fdf-28bf-40a5-92a9-8506339ed373\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-9tgn8" Mar 07 14:57:07 crc kubenswrapper[4943]: I0307 14:57:07.679461 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ms495\" (UniqueName: \"kubernetes.io/projected/81fd2fdf-28bf-40a5-92a9-8506339ed373-kube-api-access-ms495\") pod \"swift-proxy-76c998454c-9tgn8\" (UID: \"81fd2fdf-28bf-40a5-92a9-8506339ed373\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-9tgn8" Mar 07 14:57:07 crc kubenswrapper[4943]: I0307 14:57:07.780434 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/81fd2fdf-28bf-40a5-92a9-8506339ed373-log-httpd\") pod \"swift-proxy-76c998454c-9tgn8\" (UID: \"81fd2fdf-28bf-40a5-92a9-8506339ed373\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-9tgn8" Mar 07 14:57:07 crc kubenswrapper[4943]: I0307 14:57:07.780527 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81fd2fdf-28bf-40a5-92a9-8506339ed373-config-data\") pod \"swift-proxy-76c998454c-9tgn8\" (UID: \"81fd2fdf-28bf-40a5-92a9-8506339ed373\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-9tgn8" Mar 07 14:57:07 crc kubenswrapper[4943]: I0307 14:57:07.780557 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/81fd2fdf-28bf-40a5-92a9-8506339ed373-etc-swift\") pod \"swift-proxy-76c998454c-9tgn8\" (UID: \"81fd2fdf-28bf-40a5-92a9-8506339ed373\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-9tgn8" Mar 07 14:57:07 crc kubenswrapper[4943]: I0307 14:57:07.780580 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/81fd2fdf-28bf-40a5-92a9-8506339ed373-run-httpd\") pod \"swift-proxy-76c998454c-9tgn8\" (UID: \"81fd2fdf-28bf-40a5-92a9-8506339ed373\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-9tgn8" Mar 07 14:57:07 crc kubenswrapper[4943]: I0307 14:57:07.780608 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ms495\" (UniqueName: \"kubernetes.io/projected/81fd2fdf-28bf-40a5-92a9-8506339ed373-kube-api-access-ms495\") pod \"swift-proxy-76c998454c-9tgn8\" (UID: \"81fd2fdf-28bf-40a5-92a9-8506339ed373\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-9tgn8" Mar 07 14:57:07 crc kubenswrapper[4943]: I0307 14:57:07.780657 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/61bfc108-bcc1-4712-b8c8-8bc58b22777a-etc-swift\") pod \"swift-storage-0\" (UID: \"61bfc108-bcc1-4712-b8c8-8bc58b22777a\") " pod="swift-kuttl-tests/swift-storage-0" Mar 07 14:57:07 crc kubenswrapper[4943]: E0307 14:57:07.780814 4943 projected.go:288] Couldn't get configMap swift-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Mar 07 14:57:07 crc kubenswrapper[4943]: E0307 14:57:07.780833 4943 projected.go:194] Error preparing data for projected volume etc-swift for pod swift-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Mar 07 14:57:07 crc kubenswrapper[4943]: E0307 14:57:07.780890 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/61bfc108-bcc1-4712-b8c8-8bc58b22777a-etc-swift podName:61bfc108-bcc1-4712-b8c8-8bc58b22777a nodeName:}" failed. No retries permitted until 2026-03-07 14:57:08.780870631 +0000 UTC m=+1070.733007149 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/61bfc108-bcc1-4712-b8c8-8bc58b22777a-etc-swift") pod "swift-storage-0" (UID: "61bfc108-bcc1-4712-b8c8-8bc58b22777a") : configmap "swift-ring-files" not found Mar 07 14:57:07 crc kubenswrapper[4943]: I0307 14:57:07.781755 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/81fd2fdf-28bf-40a5-92a9-8506339ed373-log-httpd\") pod \"swift-proxy-76c998454c-9tgn8\" (UID: \"81fd2fdf-28bf-40a5-92a9-8506339ed373\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-9tgn8" Mar 07 14:57:07 crc kubenswrapper[4943]: E0307 14:57:07.782397 4943 projected.go:288] Couldn't get configMap swift-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Mar 07 14:57:07 crc kubenswrapper[4943]: E0307 14:57:07.782418 4943 projected.go:194] Error preparing data for projected volume etc-swift for pod swift-kuttl-tests/swift-proxy-76c998454c-9tgn8: configmap "swift-ring-files" not found Mar 07 14:57:07 crc kubenswrapper[4943]: E0307 14:57:07.782460 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/81fd2fdf-28bf-40a5-92a9-8506339ed373-etc-swift podName:81fd2fdf-28bf-40a5-92a9-8506339ed373 nodeName:}" failed. No retries permitted until 2026-03-07 14:57:08.28244708 +0000 UTC m=+1070.234583598 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/81fd2fdf-28bf-40a5-92a9-8506339ed373-etc-swift") pod "swift-proxy-76c998454c-9tgn8" (UID: "81fd2fdf-28bf-40a5-92a9-8506339ed373") : configmap "swift-ring-files" not found Mar 07 14:57:07 crc kubenswrapper[4943]: I0307 14:57:07.782795 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/81fd2fdf-28bf-40a5-92a9-8506339ed373-run-httpd\") pod \"swift-proxy-76c998454c-9tgn8\" (UID: \"81fd2fdf-28bf-40a5-92a9-8506339ed373\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-9tgn8" Mar 07 14:57:07 crc kubenswrapper[4943]: I0307 14:57:07.785375 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81fd2fdf-28bf-40a5-92a9-8506339ed373-config-data\") pod \"swift-proxy-76c998454c-9tgn8\" (UID: \"81fd2fdf-28bf-40a5-92a9-8506339ed373\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-9tgn8" Mar 07 14:57:07 crc kubenswrapper[4943]: I0307 14:57:07.799539 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ms495\" (UniqueName: \"kubernetes.io/projected/81fd2fdf-28bf-40a5-92a9-8506339ed373-kube-api-access-ms495\") pod \"swift-proxy-76c998454c-9tgn8\" (UID: \"81fd2fdf-28bf-40a5-92a9-8506339ed373\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-9tgn8" Mar 07 14:57:08 crc kubenswrapper[4943]: I0307 14:57:08.288569 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/81fd2fdf-28bf-40a5-92a9-8506339ed373-etc-swift\") pod \"swift-proxy-76c998454c-9tgn8\" (UID: \"81fd2fdf-28bf-40a5-92a9-8506339ed373\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-9tgn8" Mar 07 14:57:08 crc kubenswrapper[4943]: E0307 14:57:08.288716 4943 projected.go:288] Couldn't get configMap swift-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Mar 07 14:57:08 crc kubenswrapper[4943]: E0307 14:57:08.289025 4943 projected.go:194] Error preparing data for projected volume etc-swift for pod swift-kuttl-tests/swift-proxy-76c998454c-9tgn8: configmap "swift-ring-files" not found Mar 07 14:57:08 crc kubenswrapper[4943]: E0307 14:57:08.289097 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/81fd2fdf-28bf-40a5-92a9-8506339ed373-etc-swift podName:81fd2fdf-28bf-40a5-92a9-8506339ed373 nodeName:}" failed. No retries permitted until 2026-03-07 14:57:09.289075397 +0000 UTC m=+1071.241211895 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/81fd2fdf-28bf-40a5-92a9-8506339ed373-etc-swift") pod "swift-proxy-76c998454c-9tgn8" (UID: "81fd2fdf-28bf-40a5-92a9-8506339ed373") : configmap "swift-ring-files" not found Mar 07 14:57:08 crc kubenswrapper[4943]: I0307 14:57:08.796123 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/61bfc108-bcc1-4712-b8c8-8bc58b22777a-etc-swift\") pod \"swift-storage-0\" (UID: \"61bfc108-bcc1-4712-b8c8-8bc58b22777a\") " pod="swift-kuttl-tests/swift-storage-0" Mar 07 14:57:08 crc kubenswrapper[4943]: E0307 14:57:08.796348 4943 projected.go:288] Couldn't get configMap swift-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Mar 07 14:57:08 crc kubenswrapper[4943]: E0307 14:57:08.796381 4943 projected.go:194] Error preparing data for projected volume etc-swift for pod swift-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Mar 07 14:57:08 crc kubenswrapper[4943]: E0307 14:57:08.796448 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/61bfc108-bcc1-4712-b8c8-8bc58b22777a-etc-swift podName:61bfc108-bcc1-4712-b8c8-8bc58b22777a nodeName:}" failed. No retries permitted until 2026-03-07 14:57:10.796430013 +0000 UTC m=+1072.748566511 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/61bfc108-bcc1-4712-b8c8-8bc58b22777a-etc-swift") pod "swift-storage-0" (UID: "61bfc108-bcc1-4712-b8c8-8bc58b22777a") : configmap "swift-ring-files" not found Mar 07 14:57:09 crc kubenswrapper[4943]: I0307 14:57:09.302560 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/81fd2fdf-28bf-40a5-92a9-8506339ed373-etc-swift\") pod \"swift-proxy-76c998454c-9tgn8\" (UID: \"81fd2fdf-28bf-40a5-92a9-8506339ed373\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-9tgn8" Mar 07 14:57:09 crc kubenswrapper[4943]: E0307 14:57:09.302841 4943 projected.go:288] Couldn't get configMap swift-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Mar 07 14:57:09 crc kubenswrapper[4943]: E0307 14:57:09.303400 4943 projected.go:194] Error preparing data for projected volume etc-swift for pod swift-kuttl-tests/swift-proxy-76c998454c-9tgn8: configmap "swift-ring-files" not found Mar 07 14:57:09 crc kubenswrapper[4943]: E0307 14:57:09.303519 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/81fd2fdf-28bf-40a5-92a9-8506339ed373-etc-swift podName:81fd2fdf-28bf-40a5-92a9-8506339ed373 nodeName:}" failed. No retries permitted until 2026-03-07 14:57:11.303499263 +0000 UTC m=+1073.255635761 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/81fd2fdf-28bf-40a5-92a9-8506339ed373-etc-swift") pod "swift-proxy-76c998454c-9tgn8" (UID: "81fd2fdf-28bf-40a5-92a9-8506339ed373") : configmap "swift-ring-files" not found Mar 07 14:57:10 crc kubenswrapper[4943]: I0307 14:57:10.827113 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/61bfc108-bcc1-4712-b8c8-8bc58b22777a-etc-swift\") pod \"swift-storage-0\" (UID: \"61bfc108-bcc1-4712-b8c8-8bc58b22777a\") " pod="swift-kuttl-tests/swift-storage-0" Mar 07 14:57:10 crc kubenswrapper[4943]: E0307 14:57:10.827522 4943 projected.go:288] Couldn't get configMap swift-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Mar 07 14:57:10 crc kubenswrapper[4943]: E0307 14:57:10.827599 4943 projected.go:194] Error preparing data for projected volume etc-swift for pod swift-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Mar 07 14:57:10 crc kubenswrapper[4943]: E0307 14:57:10.827794 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/61bfc108-bcc1-4712-b8c8-8bc58b22777a-etc-swift podName:61bfc108-bcc1-4712-b8c8-8bc58b22777a nodeName:}" failed. No retries permitted until 2026-03-07 14:57:14.827724965 +0000 UTC m=+1076.779861503 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/61bfc108-bcc1-4712-b8c8-8bc58b22777a-etc-swift") pod "swift-storage-0" (UID: "61bfc108-bcc1-4712-b8c8-8bc58b22777a") : configmap "swift-ring-files" not found Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.051101 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-ngvxn"] Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.052121 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-ngvxn" Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.065082 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-ngvxn"] Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.065316 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.066086 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 14:57:11 crc kubenswrapper[4943]: E0307 14:57:11.088471 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[dispersionconf etc-swift kube-api-access-xvs8m ring-data-devices scripts swiftconf], unattached volumes=[], failed to process volumes=[dispersionconf etc-swift kube-api-access-xvs8m ring-data-devices scripts swiftconf]: context canceled" pod="swift-kuttl-tests/swift-ring-rebalance-ngvxn" podUID="32846dc3-2efa-4301-9083-bb3292b8bec6" Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.107430 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-svwpq"] Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.108674 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-svwpq" Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.141814 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-svwpq"] Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.157621 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-ngvxn"] Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.234190 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvs8m\" (UniqueName: \"kubernetes.io/projected/32846dc3-2efa-4301-9083-bb3292b8bec6-kube-api-access-xvs8m\") pod \"swift-ring-rebalance-ngvxn\" (UID: \"32846dc3-2efa-4301-9083-bb3292b8bec6\") " pod="swift-kuttl-tests/swift-ring-rebalance-ngvxn" Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.234304 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d105da1c-5832-49c5-8c86-c12c20f3983f-dispersionconf\") pod \"swift-ring-rebalance-svwpq\" (UID: \"d105da1c-5832-49c5-8c86-c12c20f3983f\") " pod="swift-kuttl-tests/swift-ring-rebalance-svwpq" Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.234361 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/32846dc3-2efa-4301-9083-bb3292b8bec6-dispersionconf\") pod \"swift-ring-rebalance-ngvxn\" (UID: \"32846dc3-2efa-4301-9083-bb3292b8bec6\") " pod="swift-kuttl-tests/swift-ring-rebalance-ngvxn" Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.234479 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/32846dc3-2efa-4301-9083-bb3292b8bec6-swiftconf\") pod \"swift-ring-rebalance-ngvxn\" (UID: \"32846dc3-2efa-4301-9083-bb3292b8bec6\") " pod="swift-kuttl-tests/swift-ring-rebalance-ngvxn" Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.234524 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/32846dc3-2efa-4301-9083-bb3292b8bec6-ring-data-devices\") pod \"swift-ring-rebalance-ngvxn\" (UID: \"32846dc3-2efa-4301-9083-bb3292b8bec6\") " pod="swift-kuttl-tests/swift-ring-rebalance-ngvxn" Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.234558 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d105da1c-5832-49c5-8c86-c12c20f3983f-ring-data-devices\") pod \"swift-ring-rebalance-svwpq\" (UID: \"d105da1c-5832-49c5-8c86-c12c20f3983f\") " pod="swift-kuttl-tests/swift-ring-rebalance-svwpq" Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.234618 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d105da1c-5832-49c5-8c86-c12c20f3983f-swiftconf\") pod \"swift-ring-rebalance-svwpq\" (UID: \"d105da1c-5832-49c5-8c86-c12c20f3983f\") " pod="swift-kuttl-tests/swift-ring-rebalance-svwpq" Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.234673 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndp5w\" (UniqueName: \"kubernetes.io/projected/d105da1c-5832-49c5-8c86-c12c20f3983f-kube-api-access-ndp5w\") pod \"swift-ring-rebalance-svwpq\" (UID: \"d105da1c-5832-49c5-8c86-c12c20f3983f\") " pod="swift-kuttl-tests/swift-ring-rebalance-svwpq" Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.234724 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/32846dc3-2efa-4301-9083-bb3292b8bec6-scripts\") pod \"swift-ring-rebalance-ngvxn\" (UID: \"32846dc3-2efa-4301-9083-bb3292b8bec6\") " pod="swift-kuttl-tests/swift-ring-rebalance-ngvxn" Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.234758 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d105da1c-5832-49c5-8c86-c12c20f3983f-scripts\") pod \"swift-ring-rebalance-svwpq\" (UID: \"d105da1c-5832-49c5-8c86-c12c20f3983f\") " pod="swift-kuttl-tests/swift-ring-rebalance-svwpq" Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.234869 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d105da1c-5832-49c5-8c86-c12c20f3983f-etc-swift\") pod \"swift-ring-rebalance-svwpq\" (UID: \"d105da1c-5832-49c5-8c86-c12c20f3983f\") " pod="swift-kuttl-tests/swift-ring-rebalance-svwpq" Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.234947 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/32846dc3-2efa-4301-9083-bb3292b8bec6-etc-swift\") pod \"swift-ring-rebalance-ngvxn\" (UID: \"32846dc3-2efa-4301-9083-bb3292b8bec6\") " pod="swift-kuttl-tests/swift-ring-rebalance-ngvxn" Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.336582 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/32846dc3-2efa-4301-9083-bb3292b8bec6-ring-data-devices\") pod \"swift-ring-rebalance-ngvxn\" (UID: \"32846dc3-2efa-4301-9083-bb3292b8bec6\") " pod="swift-kuttl-tests/swift-ring-rebalance-ngvxn" Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.336636 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d105da1c-5832-49c5-8c86-c12c20f3983f-ring-data-devices\") pod \"swift-ring-rebalance-svwpq\" (UID: \"d105da1c-5832-49c5-8c86-c12c20f3983f\") " pod="swift-kuttl-tests/swift-ring-rebalance-svwpq" Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.336685 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d105da1c-5832-49c5-8c86-c12c20f3983f-swiftconf\") pod \"swift-ring-rebalance-svwpq\" (UID: \"d105da1c-5832-49c5-8c86-c12c20f3983f\") " pod="swift-kuttl-tests/swift-ring-rebalance-svwpq" Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.336713 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndp5w\" (UniqueName: \"kubernetes.io/projected/d105da1c-5832-49c5-8c86-c12c20f3983f-kube-api-access-ndp5w\") pod \"swift-ring-rebalance-svwpq\" (UID: \"d105da1c-5832-49c5-8c86-c12c20f3983f\") " pod="swift-kuttl-tests/swift-ring-rebalance-svwpq" Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.336735 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/32846dc3-2efa-4301-9083-bb3292b8bec6-scripts\") pod \"swift-ring-rebalance-ngvxn\" (UID: \"32846dc3-2efa-4301-9083-bb3292b8bec6\") " pod="swift-kuttl-tests/swift-ring-rebalance-ngvxn" Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.336760 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d105da1c-5832-49c5-8c86-c12c20f3983f-scripts\") pod \"swift-ring-rebalance-svwpq\" (UID: \"d105da1c-5832-49c5-8c86-c12c20f3983f\") " pod="swift-kuttl-tests/swift-ring-rebalance-svwpq" Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.336785 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d105da1c-5832-49c5-8c86-c12c20f3983f-etc-swift\") pod \"swift-ring-rebalance-svwpq\" (UID: \"d105da1c-5832-49c5-8c86-c12c20f3983f\") " pod="swift-kuttl-tests/swift-ring-rebalance-svwpq" Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.336806 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/32846dc3-2efa-4301-9083-bb3292b8bec6-etc-swift\") pod \"swift-ring-rebalance-ngvxn\" (UID: \"32846dc3-2efa-4301-9083-bb3292b8bec6\") " pod="swift-kuttl-tests/swift-ring-rebalance-ngvxn" Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.336848 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvs8m\" (UniqueName: \"kubernetes.io/projected/32846dc3-2efa-4301-9083-bb3292b8bec6-kube-api-access-xvs8m\") pod \"swift-ring-rebalance-ngvxn\" (UID: \"32846dc3-2efa-4301-9083-bb3292b8bec6\") " pod="swift-kuttl-tests/swift-ring-rebalance-ngvxn" Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.336894 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d105da1c-5832-49c5-8c86-c12c20f3983f-dispersionconf\") pod \"swift-ring-rebalance-svwpq\" (UID: \"d105da1c-5832-49c5-8c86-c12c20f3983f\") " pod="swift-kuttl-tests/swift-ring-rebalance-svwpq" Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.336920 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/32846dc3-2efa-4301-9083-bb3292b8bec6-dispersionconf\") pod \"swift-ring-rebalance-ngvxn\" (UID: \"32846dc3-2efa-4301-9083-bb3292b8bec6\") " pod="swift-kuttl-tests/swift-ring-rebalance-ngvxn" Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.336972 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/81fd2fdf-28bf-40a5-92a9-8506339ed373-etc-swift\") pod \"swift-proxy-76c998454c-9tgn8\" (UID: \"81fd2fdf-28bf-40a5-92a9-8506339ed373\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-9tgn8" Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.337003 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/32846dc3-2efa-4301-9083-bb3292b8bec6-swiftconf\") pod \"swift-ring-rebalance-ngvxn\" (UID: \"32846dc3-2efa-4301-9083-bb3292b8bec6\") " pod="swift-kuttl-tests/swift-ring-rebalance-ngvxn" Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.338320 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/32846dc3-2efa-4301-9083-bb3292b8bec6-scripts\") pod \"swift-ring-rebalance-ngvxn\" (UID: \"32846dc3-2efa-4301-9083-bb3292b8bec6\") " pod="swift-kuttl-tests/swift-ring-rebalance-ngvxn" Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.338675 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d105da1c-5832-49c5-8c86-c12c20f3983f-etc-swift\") pod \"swift-ring-rebalance-svwpq\" (UID: \"d105da1c-5832-49c5-8c86-c12c20f3983f\") " pod="swift-kuttl-tests/swift-ring-rebalance-svwpq" Mar 07 14:57:11 crc kubenswrapper[4943]: E0307 14:57:11.338971 4943 projected.go:288] Couldn't get configMap swift-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Mar 07 14:57:11 crc kubenswrapper[4943]: E0307 14:57:11.339116 4943 projected.go:194] Error preparing data for projected volume etc-swift for pod swift-kuttl-tests/swift-proxy-76c998454c-9tgn8: configmap "swift-ring-files" not found Mar 07 14:57:11 crc kubenswrapper[4943]: E0307 14:57:11.339296 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/81fd2fdf-28bf-40a5-92a9-8506339ed373-etc-swift podName:81fd2fdf-28bf-40a5-92a9-8506339ed373 nodeName:}" failed. No retries permitted until 2026-03-07 14:57:15.339271154 +0000 UTC m=+1077.291407692 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/81fd2fdf-28bf-40a5-92a9-8506339ed373-etc-swift") pod "swift-proxy-76c998454c-9tgn8" (UID: "81fd2fdf-28bf-40a5-92a9-8506339ed373") : configmap "swift-ring-files" not found Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.338990 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/32846dc3-2efa-4301-9083-bb3292b8bec6-etc-swift\") pod \"swift-ring-rebalance-ngvxn\" (UID: \"32846dc3-2efa-4301-9083-bb3292b8bec6\") " pod="swift-kuttl-tests/swift-ring-rebalance-ngvxn" Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.340023 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d105da1c-5832-49c5-8c86-c12c20f3983f-ring-data-devices\") pod \"swift-ring-rebalance-svwpq\" (UID: \"d105da1c-5832-49c5-8c86-c12c20f3983f\") " pod="swift-kuttl-tests/swift-ring-rebalance-svwpq" Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.340081 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d105da1c-5832-49c5-8c86-c12c20f3983f-scripts\") pod \"swift-ring-rebalance-svwpq\" (UID: \"d105da1c-5832-49c5-8c86-c12c20f3983f\") " pod="swift-kuttl-tests/swift-ring-rebalance-svwpq" Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.340380 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/32846dc3-2efa-4301-9083-bb3292b8bec6-ring-data-devices\") pod \"swift-ring-rebalance-ngvxn\" (UID: \"32846dc3-2efa-4301-9083-bb3292b8bec6\") " pod="swift-kuttl-tests/swift-ring-rebalance-ngvxn" Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.345735 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/32846dc3-2efa-4301-9083-bb3292b8bec6-swiftconf\") pod \"swift-ring-rebalance-ngvxn\" (UID: \"32846dc3-2efa-4301-9083-bb3292b8bec6\") " pod="swift-kuttl-tests/swift-ring-rebalance-ngvxn" Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.345776 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d105da1c-5832-49c5-8c86-c12c20f3983f-swiftconf\") pod \"swift-ring-rebalance-svwpq\" (UID: \"d105da1c-5832-49c5-8c86-c12c20f3983f\") " pod="swift-kuttl-tests/swift-ring-rebalance-svwpq" Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.346160 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/32846dc3-2efa-4301-9083-bb3292b8bec6-dispersionconf\") pod \"swift-ring-rebalance-ngvxn\" (UID: \"32846dc3-2efa-4301-9083-bb3292b8bec6\") " pod="swift-kuttl-tests/swift-ring-rebalance-ngvxn" Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.348438 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d105da1c-5832-49c5-8c86-c12c20f3983f-dispersionconf\") pod \"swift-ring-rebalance-svwpq\" (UID: \"d105da1c-5832-49c5-8c86-c12c20f3983f\") " pod="swift-kuttl-tests/swift-ring-rebalance-svwpq" Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.358800 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvs8m\" (UniqueName: \"kubernetes.io/projected/32846dc3-2efa-4301-9083-bb3292b8bec6-kube-api-access-xvs8m\") pod \"swift-ring-rebalance-ngvxn\" (UID: \"32846dc3-2efa-4301-9083-bb3292b8bec6\") " pod="swift-kuttl-tests/swift-ring-rebalance-ngvxn" Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.367029 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndp5w\" (UniqueName: \"kubernetes.io/projected/d105da1c-5832-49c5-8c86-c12c20f3983f-kube-api-access-ndp5w\") pod \"swift-ring-rebalance-svwpq\" (UID: \"d105da1c-5832-49c5-8c86-c12c20f3983f\") " pod="swift-kuttl-tests/swift-ring-rebalance-svwpq" Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.425826 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-svwpq" Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.534160 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-ngvxn" Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.548913 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-ngvxn" Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.640685 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/32846dc3-2efa-4301-9083-bb3292b8bec6-scripts\") pod \"32846dc3-2efa-4301-9083-bb3292b8bec6\" (UID: \"32846dc3-2efa-4301-9083-bb3292b8bec6\") " Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.640743 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/32846dc3-2efa-4301-9083-bb3292b8bec6-etc-swift\") pod \"32846dc3-2efa-4301-9083-bb3292b8bec6\" (UID: \"32846dc3-2efa-4301-9083-bb3292b8bec6\") " Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.640776 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/32846dc3-2efa-4301-9083-bb3292b8bec6-dispersionconf\") pod \"32846dc3-2efa-4301-9083-bb3292b8bec6\" (UID: \"32846dc3-2efa-4301-9083-bb3292b8bec6\") " Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.640839 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/32846dc3-2efa-4301-9083-bb3292b8bec6-ring-data-devices\") pod \"32846dc3-2efa-4301-9083-bb3292b8bec6\" (UID: \"32846dc3-2efa-4301-9083-bb3292b8bec6\") " Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.640985 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xvs8m\" (UniqueName: \"kubernetes.io/projected/32846dc3-2efa-4301-9083-bb3292b8bec6-kube-api-access-xvs8m\") pod \"32846dc3-2efa-4301-9083-bb3292b8bec6\" (UID: \"32846dc3-2efa-4301-9083-bb3292b8bec6\") " Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.641011 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/32846dc3-2efa-4301-9083-bb3292b8bec6-swiftconf\") pod \"32846dc3-2efa-4301-9083-bb3292b8bec6\" (UID: \"32846dc3-2efa-4301-9083-bb3292b8bec6\") " Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.641417 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/32846dc3-2efa-4301-9083-bb3292b8bec6-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "32846dc3-2efa-4301-9083-bb3292b8bec6" (UID: "32846dc3-2efa-4301-9083-bb3292b8bec6"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.641427 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32846dc3-2efa-4301-9083-bb3292b8bec6-scripts" (OuterVolumeSpecName: "scripts") pod "32846dc3-2efa-4301-9083-bb3292b8bec6" (UID: "32846dc3-2efa-4301-9083-bb3292b8bec6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.641898 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32846dc3-2efa-4301-9083-bb3292b8bec6-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "32846dc3-2efa-4301-9083-bb3292b8bec6" (UID: "32846dc3-2efa-4301-9083-bb3292b8bec6"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.648646 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32846dc3-2efa-4301-9083-bb3292b8bec6-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "32846dc3-2efa-4301-9083-bb3292b8bec6" (UID: "32846dc3-2efa-4301-9083-bb3292b8bec6"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.665661 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32846dc3-2efa-4301-9083-bb3292b8bec6-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "32846dc3-2efa-4301-9083-bb3292b8bec6" (UID: "32846dc3-2efa-4301-9083-bb3292b8bec6"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.667305 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32846dc3-2efa-4301-9083-bb3292b8bec6-kube-api-access-xvs8m" (OuterVolumeSpecName: "kube-api-access-xvs8m") pod "32846dc3-2efa-4301-9083-bb3292b8bec6" (UID: "32846dc3-2efa-4301-9083-bb3292b8bec6"). InnerVolumeSpecName "kube-api-access-xvs8m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.743256 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/32846dc3-2efa-4301-9083-bb3292b8bec6-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.743302 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/32846dc3-2efa-4301-9083-bb3292b8bec6-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.743322 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/32846dc3-2efa-4301-9083-bb3292b8bec6-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.743339 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/32846dc3-2efa-4301-9083-bb3292b8bec6-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.743357 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xvs8m\" (UniqueName: \"kubernetes.io/projected/32846dc3-2efa-4301-9083-bb3292b8bec6-kube-api-access-xvs8m\") on node \"crc\" DevicePath \"\"" Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.743369 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/32846dc3-2efa-4301-9083-bb3292b8bec6-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 14:57:11 crc kubenswrapper[4943]: I0307 14:57:11.997577 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-svwpq"] Mar 07 14:57:12 crc kubenswrapper[4943]: W0307 14:57:12.001767 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd105da1c_5832_49c5_8c86_c12c20f3983f.slice/crio-96aee1432befcae3dab12b241d063ba7a80552bde16dbe1c9ee9264b66188bb6 WatchSource:0}: Error finding container 96aee1432befcae3dab12b241d063ba7a80552bde16dbe1c9ee9264b66188bb6: Status 404 returned error can't find the container with id 96aee1432befcae3dab12b241d063ba7a80552bde16dbe1c9ee9264b66188bb6 Mar 07 14:57:12 crc kubenswrapper[4943]: I0307 14:57:12.548525 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-svwpq" event={"ID":"d105da1c-5832-49c5-8c86-c12c20f3983f","Type":"ContainerStarted","Data":"96aee1432befcae3dab12b241d063ba7a80552bde16dbe1c9ee9264b66188bb6"} Mar 07 14:57:12 crc kubenswrapper[4943]: I0307 14:57:12.548690 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-ngvxn" Mar 07 14:57:12 crc kubenswrapper[4943]: I0307 14:57:12.622912 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-ngvxn"] Mar 07 14:57:12 crc kubenswrapper[4943]: I0307 14:57:12.632441 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-ngvxn"] Mar 07 14:57:12 crc kubenswrapper[4943]: I0307 14:57:12.767605 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32846dc3-2efa-4301-9083-bb3292b8bec6" path="/var/lib/kubelet/pods/32846dc3-2efa-4301-9083-bb3292b8bec6/volumes" Mar 07 14:57:14 crc kubenswrapper[4943]: I0307 14:57:14.848130 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/61bfc108-bcc1-4712-b8c8-8bc58b22777a-etc-swift\") pod \"swift-storage-0\" (UID: \"61bfc108-bcc1-4712-b8c8-8bc58b22777a\") " pod="swift-kuttl-tests/swift-storage-0" Mar 07 14:57:14 crc kubenswrapper[4943]: E0307 14:57:14.850022 4943 projected.go:288] Couldn't get configMap swift-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Mar 07 14:57:14 crc kubenswrapper[4943]: E0307 14:57:14.850094 4943 projected.go:194] Error preparing data for projected volume etc-swift for pod swift-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Mar 07 14:57:14 crc kubenswrapper[4943]: E0307 14:57:14.850188 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/61bfc108-bcc1-4712-b8c8-8bc58b22777a-etc-swift podName:61bfc108-bcc1-4712-b8c8-8bc58b22777a nodeName:}" failed. No retries permitted until 2026-03-07 14:57:22.850172769 +0000 UTC m=+1084.802309267 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/61bfc108-bcc1-4712-b8c8-8bc58b22777a-etc-swift") pod "swift-storage-0" (UID: "61bfc108-bcc1-4712-b8c8-8bc58b22777a") : configmap "swift-ring-files" not found Mar 07 14:57:15 crc kubenswrapper[4943]: I0307 14:57:15.358211 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/81fd2fdf-28bf-40a5-92a9-8506339ed373-etc-swift\") pod \"swift-proxy-76c998454c-9tgn8\" (UID: \"81fd2fdf-28bf-40a5-92a9-8506339ed373\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-9tgn8" Mar 07 14:57:15 crc kubenswrapper[4943]: E0307 14:57:15.358478 4943 projected.go:288] Couldn't get configMap swift-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Mar 07 14:57:15 crc kubenswrapper[4943]: E0307 14:57:15.358535 4943 projected.go:194] Error preparing data for projected volume etc-swift for pod swift-kuttl-tests/swift-proxy-76c998454c-9tgn8: configmap "swift-ring-files" not found Mar 07 14:57:15 crc kubenswrapper[4943]: E0307 14:57:15.358645 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/81fd2fdf-28bf-40a5-92a9-8506339ed373-etc-swift podName:81fd2fdf-28bf-40a5-92a9-8506339ed373 nodeName:}" failed. No retries permitted until 2026-03-07 14:57:23.358612311 +0000 UTC m=+1085.310748849 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/81fd2fdf-28bf-40a5-92a9-8506339ed373-etc-swift") pod "swift-proxy-76c998454c-9tgn8" (UID: "81fd2fdf-28bf-40a5-92a9-8506339ed373") : configmap "swift-ring-files" not found Mar 07 14:57:16 crc kubenswrapper[4943]: I0307 14:57:16.598917 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-svwpq" event={"ID":"d105da1c-5832-49c5-8c86-c12c20f3983f","Type":"ContainerStarted","Data":"462f33b8284a4b8f9e8effbfe35a02663da8eb314ff89b451e7780bf15a3a839"} Mar 07 14:57:16 crc kubenswrapper[4943]: I0307 14:57:16.628410 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-svwpq" podStartSLOduration=2.197461137 podStartE2EDuration="5.628394016s" podCreationTimestamp="2026-03-07 14:57:11 +0000 UTC" firstStartedPulling="2026-03-07 14:57:12.005374783 +0000 UTC m=+1073.957511281" lastFinishedPulling="2026-03-07 14:57:15.436307632 +0000 UTC m=+1077.388444160" observedRunningTime="2026-03-07 14:57:16.626880819 +0000 UTC m=+1078.579017327" watchObservedRunningTime="2026-03-07 14:57:16.628394016 +0000 UTC m=+1078.580530524" Mar 07 14:57:22 crc kubenswrapper[4943]: I0307 14:57:22.663601 4943 generic.go:334] "Generic (PLEG): container finished" podID="d105da1c-5832-49c5-8c86-c12c20f3983f" containerID="462f33b8284a4b8f9e8effbfe35a02663da8eb314ff89b451e7780bf15a3a839" exitCode=0 Mar 07 14:57:22 crc kubenswrapper[4943]: I0307 14:57:22.664361 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-svwpq" event={"ID":"d105da1c-5832-49c5-8c86-c12c20f3983f","Type":"ContainerDied","Data":"462f33b8284a4b8f9e8effbfe35a02663da8eb314ff89b451e7780bf15a3a839"} Mar 07 14:57:22 crc kubenswrapper[4943]: I0307 14:57:22.901662 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/61bfc108-bcc1-4712-b8c8-8bc58b22777a-etc-swift\") pod \"swift-storage-0\" (UID: \"61bfc108-bcc1-4712-b8c8-8bc58b22777a\") " pod="swift-kuttl-tests/swift-storage-0" Mar 07 14:57:22 crc kubenswrapper[4943]: I0307 14:57:22.912967 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/61bfc108-bcc1-4712-b8c8-8bc58b22777a-etc-swift\") pod \"swift-storage-0\" (UID: \"61bfc108-bcc1-4712-b8c8-8bc58b22777a\") " pod="swift-kuttl-tests/swift-storage-0" Mar 07 14:57:22 crc kubenswrapper[4943]: I0307 14:57:22.972579 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-storage-0" Mar 07 14:57:23 crc kubenswrapper[4943]: I0307 14:57:23.410168 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/81fd2fdf-28bf-40a5-92a9-8506339ed373-etc-swift\") pod \"swift-proxy-76c998454c-9tgn8\" (UID: \"81fd2fdf-28bf-40a5-92a9-8506339ed373\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-9tgn8" Mar 07 14:57:23 crc kubenswrapper[4943]: I0307 14:57:23.417831 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/81fd2fdf-28bf-40a5-92a9-8506339ed373-etc-swift\") pod \"swift-proxy-76c998454c-9tgn8\" (UID: \"81fd2fdf-28bf-40a5-92a9-8506339ed373\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-9tgn8" Mar 07 14:57:23 crc kubenswrapper[4943]: I0307 14:57:23.477062 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-storage-0"] Mar 07 14:57:23 crc kubenswrapper[4943]: I0307 14:57:23.480575 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-proxy-76c998454c-9tgn8" Mar 07 14:57:23 crc kubenswrapper[4943]: I0307 14:57:23.697792 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"61bfc108-bcc1-4712-b8c8-8bc58b22777a","Type":"ContainerStarted","Data":"ebd4939e3816860bb570de03484c98b55a451688036c6832ff420c3f48a74250"} Mar 07 14:57:23 crc kubenswrapper[4943]: I0307 14:57:23.986142 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-proxy-76c998454c-9tgn8"] Mar 07 14:57:24 crc kubenswrapper[4943]: I0307 14:57:24.185435 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-svwpq" Mar 07 14:57:24 crc kubenswrapper[4943]: I0307 14:57:24.344351 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d105da1c-5832-49c5-8c86-c12c20f3983f-scripts\") pod \"d105da1c-5832-49c5-8c86-c12c20f3983f\" (UID: \"d105da1c-5832-49c5-8c86-c12c20f3983f\") " Mar 07 14:57:24 crc kubenswrapper[4943]: I0307 14:57:24.344721 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d105da1c-5832-49c5-8c86-c12c20f3983f-dispersionconf\") pod \"d105da1c-5832-49c5-8c86-c12c20f3983f\" (UID: \"d105da1c-5832-49c5-8c86-c12c20f3983f\") " Mar 07 14:57:24 crc kubenswrapper[4943]: I0307 14:57:24.344758 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d105da1c-5832-49c5-8c86-c12c20f3983f-ring-data-devices\") pod \"d105da1c-5832-49c5-8c86-c12c20f3983f\" (UID: \"d105da1c-5832-49c5-8c86-c12c20f3983f\") " Mar 07 14:57:24 crc kubenswrapper[4943]: I0307 14:57:24.344869 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d105da1c-5832-49c5-8c86-c12c20f3983f-etc-swift\") pod \"d105da1c-5832-49c5-8c86-c12c20f3983f\" (UID: \"d105da1c-5832-49c5-8c86-c12c20f3983f\") " Mar 07 14:57:24 crc kubenswrapper[4943]: I0307 14:57:24.344906 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d105da1c-5832-49c5-8c86-c12c20f3983f-swiftconf\") pod \"d105da1c-5832-49c5-8c86-c12c20f3983f\" (UID: \"d105da1c-5832-49c5-8c86-c12c20f3983f\") " Mar 07 14:57:24 crc kubenswrapper[4943]: I0307 14:57:24.345075 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ndp5w\" (UniqueName: \"kubernetes.io/projected/d105da1c-5832-49c5-8c86-c12c20f3983f-kube-api-access-ndp5w\") pod \"d105da1c-5832-49c5-8c86-c12c20f3983f\" (UID: \"d105da1c-5832-49c5-8c86-c12c20f3983f\") " Mar 07 14:57:24 crc kubenswrapper[4943]: I0307 14:57:24.347485 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d105da1c-5832-49c5-8c86-c12c20f3983f-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "d105da1c-5832-49c5-8c86-c12c20f3983f" (UID: "d105da1c-5832-49c5-8c86-c12c20f3983f"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:57:24 crc kubenswrapper[4943]: I0307 14:57:24.347860 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d105da1c-5832-49c5-8c86-c12c20f3983f-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "d105da1c-5832-49c5-8c86-c12c20f3983f" (UID: "d105da1c-5832-49c5-8c86-c12c20f3983f"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:57:24 crc kubenswrapper[4943]: I0307 14:57:24.349921 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d105da1c-5832-49c5-8c86-c12c20f3983f-kube-api-access-ndp5w" (OuterVolumeSpecName: "kube-api-access-ndp5w") pod "d105da1c-5832-49c5-8c86-c12c20f3983f" (UID: "d105da1c-5832-49c5-8c86-c12c20f3983f"). InnerVolumeSpecName "kube-api-access-ndp5w". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:57:24 crc kubenswrapper[4943]: I0307 14:57:24.368789 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d105da1c-5832-49c5-8c86-c12c20f3983f-scripts" (OuterVolumeSpecName: "scripts") pod "d105da1c-5832-49c5-8c86-c12c20f3983f" (UID: "d105da1c-5832-49c5-8c86-c12c20f3983f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:57:24 crc kubenswrapper[4943]: I0307 14:57:24.373301 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d105da1c-5832-49c5-8c86-c12c20f3983f-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "d105da1c-5832-49c5-8c86-c12c20f3983f" (UID: "d105da1c-5832-49c5-8c86-c12c20f3983f"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:57:24 crc kubenswrapper[4943]: I0307 14:57:24.391359 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d105da1c-5832-49c5-8c86-c12c20f3983f-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "d105da1c-5832-49c5-8c86-c12c20f3983f" (UID: "d105da1c-5832-49c5-8c86-c12c20f3983f"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:57:24 crc kubenswrapper[4943]: I0307 14:57:24.447219 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d105da1c-5832-49c5-8c86-c12c20f3983f-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 14:57:24 crc kubenswrapper[4943]: I0307 14:57:24.447278 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d105da1c-5832-49c5-8c86-c12c20f3983f-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 14:57:24 crc kubenswrapper[4943]: I0307 14:57:24.447291 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d105da1c-5832-49c5-8c86-c12c20f3983f-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 14:57:24 crc kubenswrapper[4943]: I0307 14:57:24.447302 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ndp5w\" (UniqueName: \"kubernetes.io/projected/d105da1c-5832-49c5-8c86-c12c20f3983f-kube-api-access-ndp5w\") on node \"crc\" DevicePath \"\"" Mar 07 14:57:24 crc kubenswrapper[4943]: I0307 14:57:24.447313 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d105da1c-5832-49c5-8c86-c12c20f3983f-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 14:57:24 crc kubenswrapper[4943]: I0307 14:57:24.447323 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d105da1c-5832-49c5-8c86-c12c20f3983f-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 14:57:24 crc kubenswrapper[4943]: I0307 14:57:24.715053 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-svwpq" event={"ID":"d105da1c-5832-49c5-8c86-c12c20f3983f","Type":"ContainerDied","Data":"96aee1432befcae3dab12b241d063ba7a80552bde16dbe1c9ee9264b66188bb6"} Mar 07 14:57:24 crc kubenswrapper[4943]: I0307 14:57:24.715133 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="96aee1432befcae3dab12b241d063ba7a80552bde16dbe1c9ee9264b66188bb6" Mar 07 14:57:24 crc kubenswrapper[4943]: I0307 14:57:24.715208 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-svwpq" Mar 07 14:57:24 crc kubenswrapper[4943]: I0307 14:57:24.732194 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"61bfc108-bcc1-4712-b8c8-8bc58b22777a","Type":"ContainerStarted","Data":"3e06c0aeda3fbc189f9882999ab5a9183aa33314c68104b0f48b5857654c2239"} Mar 07 14:57:24 crc kubenswrapper[4943]: I0307 14:57:24.742694 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-proxy-76c998454c-9tgn8" event={"ID":"81fd2fdf-28bf-40a5-92a9-8506339ed373","Type":"ContainerStarted","Data":"d9db52dc67a7e088584da372ecf3e67c4da6b04503b49db9620b4cff5327c002"} Mar 07 14:57:24 crc kubenswrapper[4943]: I0307 14:57:24.742756 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-proxy-76c998454c-9tgn8" event={"ID":"81fd2fdf-28bf-40a5-92a9-8506339ed373","Type":"ContainerStarted","Data":"97c1aa81636e58fd90feba198be51760eed9720497251b34d30f4c9141ea8409"} Mar 07 14:57:24 crc kubenswrapper[4943]: I0307 14:57:24.742769 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-proxy-76c998454c-9tgn8" event={"ID":"81fd2fdf-28bf-40a5-92a9-8506339ed373","Type":"ContainerStarted","Data":"a01bfe2f9f4f71fa07d531b560b7aa1faea31b5eca72a5c024fad6a7ea4f78e4"} Mar 07 14:57:24 crc kubenswrapper[4943]: I0307 14:57:24.743185 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="swift-kuttl-tests/swift-proxy-76c998454c-9tgn8" Mar 07 14:57:24 crc kubenswrapper[4943]: I0307 14:57:24.743242 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="swift-kuttl-tests/swift-proxy-76c998454c-9tgn8" Mar 07 14:57:24 crc kubenswrapper[4943]: I0307 14:57:24.772993 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-proxy-76c998454c-9tgn8" podStartSLOduration=17.772963838 podStartE2EDuration="17.772963838s" podCreationTimestamp="2026-03-07 14:57:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:57:24.76369236 +0000 UTC m=+1086.715828878" watchObservedRunningTime="2026-03-07 14:57:24.772963838 +0000 UTC m=+1086.725100366" Mar 07 14:57:24 crc kubenswrapper[4943]: I0307 14:57:24.931757 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-ring-rebalance-svwpq_d105da1c-5832-49c5-8c86-c12c20f3983f/swift-ring-rebalance/0.log" Mar 07 14:57:25 crc kubenswrapper[4943]: I0307 14:57:25.752650 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"61bfc108-bcc1-4712-b8c8-8bc58b22777a","Type":"ContainerStarted","Data":"68e7c594bb36cc709f03800c5db0a2d96729018353fac6d60e8cb7a4ba2fa238"} Mar 07 14:57:25 crc kubenswrapper[4943]: I0307 14:57:25.752970 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"61bfc108-bcc1-4712-b8c8-8bc58b22777a","Type":"ContainerStarted","Data":"6f41a0b66118e9f7bdbd7717f791fbb3f6762f06eeb7e0aff3405d58daa4958a"} Mar 07 14:57:25 crc kubenswrapper[4943]: I0307 14:57:25.752981 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"61bfc108-bcc1-4712-b8c8-8bc58b22777a","Type":"ContainerStarted","Data":"e6a97829dc442d8db3046d52420793bbe248c6e4dacb29a69942321938a60bfa"} Mar 07 14:57:26 crc kubenswrapper[4943]: I0307 14:57:26.596324 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-ring-rebalance-svwpq_d105da1c-5832-49c5-8c86-c12c20f3983f/swift-ring-rebalance/0.log" Mar 07 14:57:26 crc kubenswrapper[4943]: I0307 14:57:26.779989 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"61bfc108-bcc1-4712-b8c8-8bc58b22777a","Type":"ContainerStarted","Data":"aa111fa2f0651801d3a0703bbef1b6650d2d7b0c5e863ca4573074b7061bd9e1"} Mar 07 14:57:27 crc kubenswrapper[4943]: I0307 14:57:27.797298 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"61bfc108-bcc1-4712-b8c8-8bc58b22777a","Type":"ContainerStarted","Data":"1d86529c5cd723c960db81582958aec90ea60c5ae034f987b60f532622638e91"} Mar 07 14:57:27 crc kubenswrapper[4943]: I0307 14:57:27.797698 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"61bfc108-bcc1-4712-b8c8-8bc58b22777a","Type":"ContainerStarted","Data":"a8781652607147e67c9640dde21b8ab12919f2616b8841b8eb11c1a3990f61ff"} Mar 07 14:57:27 crc kubenswrapper[4943]: I0307 14:57:27.797720 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"61bfc108-bcc1-4712-b8c8-8bc58b22777a","Type":"ContainerStarted","Data":"6cd56e8740db694809a028cca63c9a4433743dfb0ebe9e04a1967f2b37e9d293"} Mar 07 14:57:28 crc kubenswrapper[4943]: I0307 14:57:28.254650 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-ring-rebalance-svwpq_d105da1c-5832-49c5-8c86-c12c20f3983f/swift-ring-rebalance/0.log" Mar 07 14:57:28 crc kubenswrapper[4943]: I0307 14:57:28.824054 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"61bfc108-bcc1-4712-b8c8-8bc58b22777a","Type":"ContainerStarted","Data":"8a21f246526c030a1c8871a1c9d61670fe1dc6f835254920aaa17f9ba7229d99"} Mar 07 14:57:28 crc kubenswrapper[4943]: I0307 14:57:28.824450 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"61bfc108-bcc1-4712-b8c8-8bc58b22777a","Type":"ContainerStarted","Data":"d93dddf647b1cf741991cbef8e4bb89720ea9c094dfde025aa45574b1b46d1b3"} Mar 07 14:57:29 crc kubenswrapper[4943]: I0307 14:57:29.844094 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"61bfc108-bcc1-4712-b8c8-8bc58b22777a","Type":"ContainerStarted","Data":"95bb5217f844e378e57512480c7f0c59e7b85d4265619fb46dcdc132dc57a739"} Mar 07 14:57:29 crc kubenswrapper[4943]: I0307 14:57:29.844151 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"61bfc108-bcc1-4712-b8c8-8bc58b22777a","Type":"ContainerStarted","Data":"6efa5c1c20f09b9e71b1f4e99886eb7db8e5d70a8d132abf7a9908a54e7dbbea"} Mar 07 14:57:29 crc kubenswrapper[4943]: I0307 14:57:29.844163 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"61bfc108-bcc1-4712-b8c8-8bc58b22777a","Type":"ContainerStarted","Data":"26f55f0f6682744f5ed31577037541f7b99894bdd3dd19a114e1590e2b8697ab"} Mar 07 14:57:29 crc kubenswrapper[4943]: I0307 14:57:29.844173 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"61bfc108-bcc1-4712-b8c8-8bc58b22777a","Type":"ContainerStarted","Data":"7b896439aae01ff52fa43a69d231483bddd5f7624f25904fc8ae68de8c2dc3cc"} Mar 07 14:57:29 crc kubenswrapper[4943]: I0307 14:57:29.844184 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"61bfc108-bcc1-4712-b8c8-8bc58b22777a","Type":"ContainerStarted","Data":"b5daa7f5307510498acd32d81ac9f7c0aab7c6308c53dc697686fd07d1776777"} Mar 07 14:57:29 crc kubenswrapper[4943]: I0307 14:57:29.853954 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-ring-rebalance-svwpq_d105da1c-5832-49c5-8c86-c12c20f3983f/swift-ring-rebalance/0.log" Mar 07 14:57:29 crc kubenswrapper[4943]: I0307 14:57:29.922367 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-storage-0" podStartSLOduration=19.169846465 podStartE2EDuration="23.922331353s" podCreationTimestamp="2026-03-07 14:57:06 +0000 UTC" firstStartedPulling="2026-03-07 14:57:23.504407852 +0000 UTC m=+1085.456544390" lastFinishedPulling="2026-03-07 14:57:28.25689277 +0000 UTC m=+1090.209029278" observedRunningTime="2026-03-07 14:57:29.915389223 +0000 UTC m=+1091.867525761" watchObservedRunningTime="2026-03-07 14:57:29.922331353 +0000 UTC m=+1091.874467881" Mar 07 14:57:31 crc kubenswrapper[4943]: I0307 14:57:31.420861 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-ring-rebalance-svwpq_d105da1c-5832-49c5-8c86-c12c20f3983f/swift-ring-rebalance/0.log" Mar 07 14:57:33 crc kubenswrapper[4943]: I0307 14:57:33.081875 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-ring-rebalance-svwpq_d105da1c-5832-49c5-8c86-c12c20f3983f/swift-ring-rebalance/0.log" Mar 07 14:57:33 crc kubenswrapper[4943]: I0307 14:57:33.486354 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="swift-kuttl-tests/swift-proxy-76c998454c-9tgn8" Mar 07 14:57:33 crc kubenswrapper[4943]: I0307 14:57:33.488598 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="swift-kuttl-tests/swift-proxy-76c998454c-9tgn8" Mar 07 14:57:34 crc kubenswrapper[4943]: I0307 14:57:34.673806 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-ring-rebalance-svwpq_d105da1c-5832-49c5-8c86-c12c20f3983f/swift-ring-rebalance/0.log" Mar 07 14:57:36 crc kubenswrapper[4943]: I0307 14:57:36.271228 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-ring-rebalance-svwpq_d105da1c-5832-49c5-8c86-c12c20f3983f/swift-ring-rebalance/0.log" Mar 07 14:57:37 crc kubenswrapper[4943]: I0307 14:57:37.688812 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-storage-1"] Mar 07 14:57:37 crc kubenswrapper[4943]: E0307 14:57:37.689342 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d105da1c-5832-49c5-8c86-c12c20f3983f" containerName="swift-ring-rebalance" Mar 07 14:57:37 crc kubenswrapper[4943]: I0307 14:57:37.689364 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="d105da1c-5832-49c5-8c86-c12c20f3983f" containerName="swift-ring-rebalance" Mar 07 14:57:37 crc kubenswrapper[4943]: I0307 14:57:37.689571 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="d105da1c-5832-49c5-8c86-c12c20f3983f" containerName="swift-ring-rebalance" Mar 07 14:57:37 crc kubenswrapper[4943]: I0307 14:57:37.703974 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-storage-1" Mar 07 14:57:37 crc kubenswrapper[4943]: I0307 14:57:37.750205 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-storage-2"] Mar 07 14:57:37 crc kubenswrapper[4943]: I0307 14:57:37.761957 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-storage-1"] Mar 07 14:57:37 crc kubenswrapper[4943]: I0307 14:57:37.762122 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-storage-2" Mar 07 14:57:37 crc kubenswrapper[4943]: I0307 14:57:37.803551 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-storage-2"] Mar 07 14:57:37 crc kubenswrapper[4943]: I0307 14:57:37.870628 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-1\" (UID: \"db16093b-eede-415e-8145-56e579a8782a\") " pod="swift-kuttl-tests/swift-storage-1" Mar 07 14:57:37 crc kubenswrapper[4943]: I0307 14:57:37.870741 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"swift-storage-2\" (UID: \"0e83c041-7178-43e4-b05d-e1bcc0ec36bd\") " pod="swift-kuttl-tests/swift-storage-2" Mar 07 14:57:37 crc kubenswrapper[4943]: I0307 14:57:37.870778 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/0e83c041-7178-43e4-b05d-e1bcc0ec36bd-lock\") pod \"swift-storage-2\" (UID: \"0e83c041-7178-43e4-b05d-e1bcc0ec36bd\") " pod="swift-kuttl-tests/swift-storage-2" Mar 07 14:57:37 crc kubenswrapper[4943]: I0307 14:57:37.870818 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/db16093b-eede-415e-8145-56e579a8782a-etc-swift\") pod \"swift-storage-1\" (UID: \"db16093b-eede-415e-8145-56e579a8782a\") " pod="swift-kuttl-tests/swift-storage-1" Mar 07 14:57:37 crc kubenswrapper[4943]: I0307 14:57:37.870842 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nknnt\" (UniqueName: \"kubernetes.io/projected/db16093b-eede-415e-8145-56e579a8782a-kube-api-access-nknnt\") pod \"swift-storage-1\" (UID: \"db16093b-eede-415e-8145-56e579a8782a\") " pod="swift-kuttl-tests/swift-storage-1" Mar 07 14:57:37 crc kubenswrapper[4943]: I0307 14:57:37.870860 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/0e83c041-7178-43e4-b05d-e1bcc0ec36bd-cache\") pod \"swift-storage-2\" (UID: \"0e83c041-7178-43e4-b05d-e1bcc0ec36bd\") " pod="swift-kuttl-tests/swift-storage-2" Mar 07 14:57:37 crc kubenswrapper[4943]: I0307 14:57:37.870896 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/db16093b-eede-415e-8145-56e579a8782a-cache\") pod \"swift-storage-1\" (UID: \"db16093b-eede-415e-8145-56e579a8782a\") " pod="swift-kuttl-tests/swift-storage-1" Mar 07 14:57:37 crc kubenswrapper[4943]: I0307 14:57:37.870909 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/db16093b-eede-415e-8145-56e579a8782a-lock\") pod \"swift-storage-1\" (UID: \"db16093b-eede-415e-8145-56e579a8782a\") " pod="swift-kuttl-tests/swift-storage-1" Mar 07 14:57:37 crc kubenswrapper[4943]: I0307 14:57:37.870949 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0e83c041-7178-43e4-b05d-e1bcc0ec36bd-etc-swift\") pod \"swift-storage-2\" (UID: \"0e83c041-7178-43e4-b05d-e1bcc0ec36bd\") " pod="swift-kuttl-tests/swift-storage-2" Mar 07 14:57:37 crc kubenswrapper[4943]: I0307 14:57:37.870972 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llm5v\" (UniqueName: \"kubernetes.io/projected/0e83c041-7178-43e4-b05d-e1bcc0ec36bd-kube-api-access-llm5v\") pod \"swift-storage-2\" (UID: \"0e83c041-7178-43e4-b05d-e1bcc0ec36bd\") " pod="swift-kuttl-tests/swift-storage-2" Mar 07 14:57:37 crc kubenswrapper[4943]: I0307 14:57:37.971945 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nknnt\" (UniqueName: \"kubernetes.io/projected/db16093b-eede-415e-8145-56e579a8782a-kube-api-access-nknnt\") pod \"swift-storage-1\" (UID: \"db16093b-eede-415e-8145-56e579a8782a\") " pod="swift-kuttl-tests/swift-storage-1" Mar 07 14:57:37 crc kubenswrapper[4943]: I0307 14:57:37.971985 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/0e83c041-7178-43e4-b05d-e1bcc0ec36bd-cache\") pod \"swift-storage-2\" (UID: \"0e83c041-7178-43e4-b05d-e1bcc0ec36bd\") " pod="swift-kuttl-tests/swift-storage-2" Mar 07 14:57:37 crc kubenswrapper[4943]: I0307 14:57:37.972011 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/db16093b-eede-415e-8145-56e579a8782a-cache\") pod \"swift-storage-1\" (UID: \"db16093b-eede-415e-8145-56e579a8782a\") " pod="swift-kuttl-tests/swift-storage-1" Mar 07 14:57:37 crc kubenswrapper[4943]: I0307 14:57:37.972024 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/db16093b-eede-415e-8145-56e579a8782a-lock\") pod \"swift-storage-1\" (UID: \"db16093b-eede-415e-8145-56e579a8782a\") " pod="swift-kuttl-tests/swift-storage-1" Mar 07 14:57:37 crc kubenswrapper[4943]: I0307 14:57:37.972042 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0e83c041-7178-43e4-b05d-e1bcc0ec36bd-etc-swift\") pod \"swift-storage-2\" (UID: \"0e83c041-7178-43e4-b05d-e1bcc0ec36bd\") " pod="swift-kuttl-tests/swift-storage-2" Mar 07 14:57:37 crc kubenswrapper[4943]: I0307 14:57:37.972065 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llm5v\" (UniqueName: \"kubernetes.io/projected/0e83c041-7178-43e4-b05d-e1bcc0ec36bd-kube-api-access-llm5v\") pod \"swift-storage-2\" (UID: \"0e83c041-7178-43e4-b05d-e1bcc0ec36bd\") " pod="swift-kuttl-tests/swift-storage-2" Mar 07 14:57:37 crc kubenswrapper[4943]: I0307 14:57:37.972107 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-1\" (UID: \"db16093b-eede-415e-8145-56e579a8782a\") " pod="swift-kuttl-tests/swift-storage-1" Mar 07 14:57:37 crc kubenswrapper[4943]: I0307 14:57:37.972143 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"swift-storage-2\" (UID: \"0e83c041-7178-43e4-b05d-e1bcc0ec36bd\") " pod="swift-kuttl-tests/swift-storage-2" Mar 07 14:57:37 crc kubenswrapper[4943]: I0307 14:57:37.972164 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/0e83c041-7178-43e4-b05d-e1bcc0ec36bd-lock\") pod \"swift-storage-2\" (UID: \"0e83c041-7178-43e4-b05d-e1bcc0ec36bd\") " pod="swift-kuttl-tests/swift-storage-2" Mar 07 14:57:37 crc kubenswrapper[4943]: I0307 14:57:37.972189 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/db16093b-eede-415e-8145-56e579a8782a-etc-swift\") pod \"swift-storage-1\" (UID: \"db16093b-eede-415e-8145-56e579a8782a\") " pod="swift-kuttl-tests/swift-storage-1" Mar 07 14:57:37 crc kubenswrapper[4943]: I0307 14:57:37.972521 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/db16093b-eede-415e-8145-56e579a8782a-cache\") pod \"swift-storage-1\" (UID: \"db16093b-eede-415e-8145-56e579a8782a\") " pod="swift-kuttl-tests/swift-storage-1" Mar 07 14:57:37 crc kubenswrapper[4943]: I0307 14:57:37.972552 4943 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-1\" (UID: \"db16093b-eede-415e-8145-56e579a8782a\") device mount path \"/mnt/openstack/pv04\"" pod="swift-kuttl-tests/swift-storage-1" Mar 07 14:57:37 crc kubenswrapper[4943]: I0307 14:57:37.972596 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/db16093b-eede-415e-8145-56e579a8782a-lock\") pod \"swift-storage-1\" (UID: \"db16093b-eede-415e-8145-56e579a8782a\") " pod="swift-kuttl-tests/swift-storage-1" Mar 07 14:57:37 crc kubenswrapper[4943]: I0307 14:57:37.972647 4943 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"swift-storage-2\" (UID: \"0e83c041-7178-43e4-b05d-e1bcc0ec36bd\") device mount path \"/mnt/openstack/pv05\"" pod="swift-kuttl-tests/swift-storage-2" Mar 07 14:57:37 crc kubenswrapper[4943]: I0307 14:57:37.973917 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/0e83c041-7178-43e4-b05d-e1bcc0ec36bd-lock\") pod \"swift-storage-2\" (UID: \"0e83c041-7178-43e4-b05d-e1bcc0ec36bd\") " pod="swift-kuttl-tests/swift-storage-2" Mar 07 14:57:37 crc kubenswrapper[4943]: I0307 14:57:37.974279 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/0e83c041-7178-43e4-b05d-e1bcc0ec36bd-cache\") pod \"swift-storage-2\" (UID: \"0e83c041-7178-43e4-b05d-e1bcc0ec36bd\") " pod="swift-kuttl-tests/swift-storage-2" Mar 07 14:57:37 crc kubenswrapper[4943]: I0307 14:57:37.980179 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0e83c041-7178-43e4-b05d-e1bcc0ec36bd-etc-swift\") pod \"swift-storage-2\" (UID: \"0e83c041-7178-43e4-b05d-e1bcc0ec36bd\") " pod="swift-kuttl-tests/swift-storage-2" Mar 07 14:57:37 crc kubenswrapper[4943]: I0307 14:57:37.980367 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/db16093b-eede-415e-8145-56e579a8782a-etc-swift\") pod \"swift-storage-1\" (UID: \"db16093b-eede-415e-8145-56e579a8782a\") " pod="swift-kuttl-tests/swift-storage-1" Mar 07 14:57:37 crc kubenswrapper[4943]: I0307 14:57:37.991339 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llm5v\" (UniqueName: \"kubernetes.io/projected/0e83c041-7178-43e4-b05d-e1bcc0ec36bd-kube-api-access-llm5v\") pod \"swift-storage-2\" (UID: \"0e83c041-7178-43e4-b05d-e1bcc0ec36bd\") " pod="swift-kuttl-tests/swift-storage-2" Mar 07 14:57:37 crc kubenswrapper[4943]: I0307 14:57:37.994977 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nknnt\" (UniqueName: \"kubernetes.io/projected/db16093b-eede-415e-8145-56e579a8782a-kube-api-access-nknnt\") pod \"swift-storage-1\" (UID: \"db16093b-eede-415e-8145-56e579a8782a\") " pod="swift-kuttl-tests/swift-storage-1" Mar 07 14:57:37 crc kubenswrapper[4943]: I0307 14:57:37.997396 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"swift-storage-2\" (UID: \"0e83c041-7178-43e4-b05d-e1bcc0ec36bd\") " pod="swift-kuttl-tests/swift-storage-2" Mar 07 14:57:37 crc kubenswrapper[4943]: I0307 14:57:37.997452 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-1\" (UID: \"db16093b-eede-415e-8145-56e579a8782a\") " pod="swift-kuttl-tests/swift-storage-1" Mar 07 14:57:38 crc kubenswrapper[4943]: I0307 14:57:38.062003 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-storage-1" Mar 07 14:57:38 crc kubenswrapper[4943]: I0307 14:57:38.086766 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-storage-2" Mar 07 14:57:38 crc kubenswrapper[4943]: I0307 14:57:38.581703 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-storage-1"] Mar 07 14:57:38 crc kubenswrapper[4943]: W0307 14:57:38.589969 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddb16093b_eede_415e_8145_56e579a8782a.slice/crio-f00d6313816b2a63ab9071197558bf9bdce2ddd970da2a5f725ea0264364e1b6 WatchSource:0}: Error finding container f00d6313816b2a63ab9071197558bf9bdce2ddd970da2a5f725ea0264364e1b6: Status 404 returned error can't find the container with id f00d6313816b2a63ab9071197558bf9bdce2ddd970da2a5f725ea0264364e1b6 Mar 07 14:57:38 crc kubenswrapper[4943]: I0307 14:57:38.659045 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-storage-2"] Mar 07 14:57:38 crc kubenswrapper[4943]: W0307 14:57:38.692159 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0e83c041_7178_43e4_b05d_e1bcc0ec36bd.slice/crio-9bd03386ba29d23157b54d68484fe1fc385e534892b2e7d7d73856be11bbea58 WatchSource:0}: Error finding container 9bd03386ba29d23157b54d68484fe1fc385e534892b2e7d7d73856be11bbea58: Status 404 returned error can't find the container with id 9bd03386ba29d23157b54d68484fe1fc385e534892b2e7d7d73856be11bbea58 Mar 07 14:57:38 crc kubenswrapper[4943]: I0307 14:57:38.936814 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"db16093b-eede-415e-8145-56e579a8782a","Type":"ContainerStarted","Data":"e346ec17aa2d2cdfc931a3fd9158238065a5520b87ac1b27e152b7596b21a812"} Mar 07 14:57:38 crc kubenswrapper[4943]: I0307 14:57:38.936859 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"db16093b-eede-415e-8145-56e579a8782a","Type":"ContainerStarted","Data":"f00d6313816b2a63ab9071197558bf9bdce2ddd970da2a5f725ea0264364e1b6"} Mar 07 14:57:38 crc kubenswrapper[4943]: I0307 14:57:38.938287 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"0e83c041-7178-43e4-b05d-e1bcc0ec36bd","Type":"ContainerStarted","Data":"378e9e0fc8363d06ebee218138d567cd679e69eab4d30a8c35d8d6ba0c009165"} Mar 07 14:57:38 crc kubenswrapper[4943]: I0307 14:57:38.938308 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"0e83c041-7178-43e4-b05d-e1bcc0ec36bd","Type":"ContainerStarted","Data":"9bd03386ba29d23157b54d68484fe1fc385e534892b2e7d7d73856be11bbea58"} Mar 07 14:57:39 crc kubenswrapper[4943]: I0307 14:57:39.959103 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"db16093b-eede-415e-8145-56e579a8782a","Type":"ContainerStarted","Data":"4115dfe7381bf3fa3917d022ff124a7aa33326a18d5838b3caf284409aed99a2"} Mar 07 14:57:39 crc kubenswrapper[4943]: I0307 14:57:39.959692 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"db16093b-eede-415e-8145-56e579a8782a","Type":"ContainerStarted","Data":"e651e02c4c021413625c666e0f2aae32c58a1f777868324172ec3a75e9fbdd1d"} Mar 07 14:57:39 crc kubenswrapper[4943]: I0307 14:57:39.959709 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"db16093b-eede-415e-8145-56e579a8782a","Type":"ContainerStarted","Data":"f325a942655495f7cfc8dda622634e063c67e9ae302d7bc629c67eafa8aa4902"} Mar 07 14:57:39 crc kubenswrapper[4943]: I0307 14:57:39.959722 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"db16093b-eede-415e-8145-56e579a8782a","Type":"ContainerStarted","Data":"c89ce52dcf78c4288953c7557fe606114f20cd3a3351e3c89c2537818ce3be10"} Mar 07 14:57:39 crc kubenswrapper[4943]: I0307 14:57:39.959735 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"db16093b-eede-415e-8145-56e579a8782a","Type":"ContainerStarted","Data":"fde39a16da2d051fe287b573e0e4cdba6b70fea2e8f7445ef402a9a97f173ca7"} Mar 07 14:57:39 crc kubenswrapper[4943]: I0307 14:57:39.959747 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"db16093b-eede-415e-8145-56e579a8782a","Type":"ContainerStarted","Data":"1e42ab3da5309b6ad0251fd6b7b50713d2bf82db8ebe73f72da280111db9a63a"} Mar 07 14:57:39 crc kubenswrapper[4943]: I0307 14:57:39.959759 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"db16093b-eede-415e-8145-56e579a8782a","Type":"ContainerStarted","Data":"c7d049e1848ff61a6aa9a2cd8b145da1ddf8e0d5675e4ddaf1b547116fa455b2"} Mar 07 14:57:39 crc kubenswrapper[4943]: I0307 14:57:39.967102 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"0e83c041-7178-43e4-b05d-e1bcc0ec36bd","Type":"ContainerStarted","Data":"8a933cae33841023d45b5c9e60e92f252afc264f4c9d875e212fd937dfe8a709"} Mar 07 14:57:39 crc kubenswrapper[4943]: I0307 14:57:39.967180 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"0e83c041-7178-43e4-b05d-e1bcc0ec36bd","Type":"ContainerStarted","Data":"b1ab2126394629d1a8dbbaa0f7f865780384c107928dc42b81568566f59b1cdd"} Mar 07 14:57:39 crc kubenswrapper[4943]: I0307 14:57:39.967195 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"0e83c041-7178-43e4-b05d-e1bcc0ec36bd","Type":"ContainerStarted","Data":"466fe0dc2e8ddc693dc18033ff08abb79536de9d9c21b6015040b06e7454b177"} Mar 07 14:57:39 crc kubenswrapper[4943]: I0307 14:57:39.967208 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"0e83c041-7178-43e4-b05d-e1bcc0ec36bd","Type":"ContainerStarted","Data":"386892a8a101df1e91f2ce3827d6706952226a956fdbb507d91575d4831f42b3"} Mar 07 14:57:39 crc kubenswrapper[4943]: I0307 14:57:39.967219 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"0e83c041-7178-43e4-b05d-e1bcc0ec36bd","Type":"ContainerStarted","Data":"a36e51a00bdb5a305ca9e42e0cf8e5db2c0802a50081b0774e6d53712e2848da"} Mar 07 14:57:40 crc kubenswrapper[4943]: I0307 14:57:40.992083 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"0e83c041-7178-43e4-b05d-e1bcc0ec36bd","Type":"ContainerStarted","Data":"185a3b1b195164b0391a65fb58c65f24f848fd66987f773e5d4f2031d582ad76"} Mar 07 14:57:40 crc kubenswrapper[4943]: I0307 14:57:40.992127 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"0e83c041-7178-43e4-b05d-e1bcc0ec36bd","Type":"ContainerStarted","Data":"ba8857ea0c02e4a8248cffb1f9d8e7c2b66181c038e948a8f69e9f91e89979a7"} Mar 07 14:57:40 crc kubenswrapper[4943]: I0307 14:57:40.992136 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"0e83c041-7178-43e4-b05d-e1bcc0ec36bd","Type":"ContainerStarted","Data":"7b3105463e57fbdbaa2e126058c6aaa054286f515ac013e5b967046d33b29603"} Mar 07 14:57:40 crc kubenswrapper[4943]: I0307 14:57:40.992442 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"0e83c041-7178-43e4-b05d-e1bcc0ec36bd","Type":"ContainerStarted","Data":"e571f26eb8b825f06454b7946c558c7a1eb1f69258f0d8ada3b654c53c97b55b"} Mar 07 14:57:40 crc kubenswrapper[4943]: I0307 14:57:40.992450 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"0e83c041-7178-43e4-b05d-e1bcc0ec36bd","Type":"ContainerStarted","Data":"257741f8218f29b0b1e5e8d5ef95ecf3f28bee16cdc292d77de8f3326bbe123d"} Mar 07 14:57:40 crc kubenswrapper[4943]: I0307 14:57:40.992458 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"0e83c041-7178-43e4-b05d-e1bcc0ec36bd","Type":"ContainerStarted","Data":"0f1f14f4fa5711a1d3b92eef66fde7b260fbee7a7bbeed91f897f960ef2e7c3e"} Mar 07 14:57:41 crc kubenswrapper[4943]: I0307 14:57:41.000447 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"db16093b-eede-415e-8145-56e579a8782a","Type":"ContainerStarted","Data":"1258864ced31fd4d6f607ce7df12ec446e74a8d5bc0a46cf2122b7b2d5b1972d"} Mar 07 14:57:41 crc kubenswrapper[4943]: I0307 14:57:41.000498 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"db16093b-eede-415e-8145-56e579a8782a","Type":"ContainerStarted","Data":"1676e9d2423174fdbd7bb0a1ca19bd1ce772561c44dd6229f48cf4244b017fd8"} Mar 07 14:57:41 crc kubenswrapper[4943]: I0307 14:57:41.000507 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"db16093b-eede-415e-8145-56e579a8782a","Type":"ContainerStarted","Data":"8d4253973cf92b8f134d333111c459dd45a28b58115e11f17fba582a8fce9cfd"} Mar 07 14:57:41 crc kubenswrapper[4943]: I0307 14:57:41.000515 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"db16093b-eede-415e-8145-56e579a8782a","Type":"ContainerStarted","Data":"635ececc7dce091575d667fc55ff70dae5854a71fa610c71180581ff05c44e85"} Mar 07 14:57:41 crc kubenswrapper[4943]: I0307 14:57:41.000524 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"db16093b-eede-415e-8145-56e579a8782a","Type":"ContainerStarted","Data":"6f3ce1ef2dba1d61c13484a1020aa353732fd7e4b8aca4d62f081d429dd9cf7f"} Mar 07 14:57:41 crc kubenswrapper[4943]: I0307 14:57:41.610856 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-svwpq"] Mar 07 14:57:41 crc kubenswrapper[4943]: I0307 14:57:41.619672 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-svwpq"] Mar 07 14:57:41 crc kubenswrapper[4943]: I0307 14:57:41.651995 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-89s8j"] Mar 07 14:57:41 crc kubenswrapper[4943]: I0307 14:57:41.653312 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-89s8j" Mar 07 14:57:41 crc kubenswrapper[4943]: I0307 14:57:41.656144 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 14:57:41 crc kubenswrapper[4943]: I0307 14:57:41.656234 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 14:57:41 crc kubenswrapper[4943]: I0307 14:57:41.659662 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-89s8j"] Mar 07 14:57:41 crc kubenswrapper[4943]: E0307 14:57:41.675969 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[dispersionconf etc-swift kube-api-access-k5hgr ring-data-devices scripts swiftconf], unattached volumes=[], failed to process volumes=[dispersionconf etc-swift kube-api-access-k5hgr ring-data-devices scripts swiftconf]: context canceled" pod="swift-kuttl-tests/swift-ring-rebalance-89s8j" podUID="b43e5152-86c3-4e1b-8fe1-d440d2e37b82" Mar 07 14:57:41 crc kubenswrapper[4943]: I0307 14:57:41.687208 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-xds8n"] Mar 07 14:57:41 crc kubenswrapper[4943]: I0307 14:57:41.688638 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-xds8n" Mar 07 14:57:41 crc kubenswrapper[4943]: I0307 14:57:41.695676 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-89s8j"] Mar 07 14:57:41 crc kubenswrapper[4943]: I0307 14:57:41.731701 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b43e5152-86c3-4e1b-8fe1-d440d2e37b82-ring-data-devices\") pod \"swift-ring-rebalance-89s8j\" (UID: \"b43e5152-86c3-4e1b-8fe1-d440d2e37b82\") " pod="swift-kuttl-tests/swift-ring-rebalance-89s8j" Mar 07 14:57:41 crc kubenswrapper[4943]: I0307 14:57:41.731769 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b43e5152-86c3-4e1b-8fe1-d440d2e37b82-swiftconf\") pod \"swift-ring-rebalance-89s8j\" (UID: \"b43e5152-86c3-4e1b-8fe1-d440d2e37b82\") " pod="swift-kuttl-tests/swift-ring-rebalance-89s8j" Mar 07 14:57:41 crc kubenswrapper[4943]: I0307 14:57:41.731795 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b43e5152-86c3-4e1b-8fe1-d440d2e37b82-dispersionconf\") pod \"swift-ring-rebalance-89s8j\" (UID: \"b43e5152-86c3-4e1b-8fe1-d440d2e37b82\") " pod="swift-kuttl-tests/swift-ring-rebalance-89s8j" Mar 07 14:57:41 crc kubenswrapper[4943]: I0307 14:57:41.731819 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b43e5152-86c3-4e1b-8fe1-d440d2e37b82-scripts\") pod \"swift-ring-rebalance-89s8j\" (UID: \"b43e5152-86c3-4e1b-8fe1-d440d2e37b82\") " pod="swift-kuttl-tests/swift-ring-rebalance-89s8j" Mar 07 14:57:41 crc kubenswrapper[4943]: I0307 14:57:41.732059 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b43e5152-86c3-4e1b-8fe1-d440d2e37b82-etc-swift\") pod \"swift-ring-rebalance-89s8j\" (UID: \"b43e5152-86c3-4e1b-8fe1-d440d2e37b82\") " pod="swift-kuttl-tests/swift-ring-rebalance-89s8j" Mar 07 14:57:41 crc kubenswrapper[4943]: I0307 14:57:41.732345 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5hgr\" (UniqueName: \"kubernetes.io/projected/b43e5152-86c3-4e1b-8fe1-d440d2e37b82-kube-api-access-k5hgr\") pod \"swift-ring-rebalance-89s8j\" (UID: \"b43e5152-86c3-4e1b-8fe1-d440d2e37b82\") " pod="swift-kuttl-tests/swift-ring-rebalance-89s8j" Mar 07 14:57:41 crc kubenswrapper[4943]: I0307 14:57:41.742268 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-xds8n"] Mar 07 14:57:41 crc kubenswrapper[4943]: I0307 14:57:41.833861 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b43e5152-86c3-4e1b-8fe1-d440d2e37b82-ring-data-devices\") pod \"swift-ring-rebalance-89s8j\" (UID: \"b43e5152-86c3-4e1b-8fe1-d440d2e37b82\") " pod="swift-kuttl-tests/swift-ring-rebalance-89s8j" Mar 07 14:57:41 crc kubenswrapper[4943]: I0307 14:57:41.833996 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b43e5152-86c3-4e1b-8fe1-d440d2e37b82-swiftconf\") pod \"swift-ring-rebalance-89s8j\" (UID: \"b43e5152-86c3-4e1b-8fe1-d440d2e37b82\") " pod="swift-kuttl-tests/swift-ring-rebalance-89s8j" Mar 07 14:57:41 crc kubenswrapper[4943]: I0307 14:57:41.834034 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6eaa7147-1483-4399-ae55-b0f27c44a0e6-ring-data-devices\") pod \"swift-ring-rebalance-xds8n\" (UID: \"6eaa7147-1483-4399-ae55-b0f27c44a0e6\") " pod="swift-kuttl-tests/swift-ring-rebalance-xds8n" Mar 07 14:57:41 crc kubenswrapper[4943]: I0307 14:57:41.834060 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b43e5152-86c3-4e1b-8fe1-d440d2e37b82-dispersionconf\") pod \"swift-ring-rebalance-89s8j\" (UID: \"b43e5152-86c3-4e1b-8fe1-d440d2e37b82\") " pod="swift-kuttl-tests/swift-ring-rebalance-89s8j" Mar 07 14:57:41 crc kubenswrapper[4943]: I0307 14:57:41.834091 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6eaa7147-1483-4399-ae55-b0f27c44a0e6-swiftconf\") pod \"swift-ring-rebalance-xds8n\" (UID: \"6eaa7147-1483-4399-ae55-b0f27c44a0e6\") " pod="swift-kuttl-tests/swift-ring-rebalance-xds8n" Mar 07 14:57:41 crc kubenswrapper[4943]: I0307 14:57:41.834118 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b43e5152-86c3-4e1b-8fe1-d440d2e37b82-scripts\") pod \"swift-ring-rebalance-89s8j\" (UID: \"b43e5152-86c3-4e1b-8fe1-d440d2e37b82\") " pod="swift-kuttl-tests/swift-ring-rebalance-89s8j" Mar 07 14:57:41 crc kubenswrapper[4943]: I0307 14:57:41.834170 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b43e5152-86c3-4e1b-8fe1-d440d2e37b82-etc-swift\") pod \"swift-ring-rebalance-89s8j\" (UID: \"b43e5152-86c3-4e1b-8fe1-d440d2e37b82\") " pod="swift-kuttl-tests/swift-ring-rebalance-89s8j" Mar 07 14:57:41 crc kubenswrapper[4943]: I0307 14:57:41.834202 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86v2q\" (UniqueName: \"kubernetes.io/projected/6eaa7147-1483-4399-ae55-b0f27c44a0e6-kube-api-access-86v2q\") pod \"swift-ring-rebalance-xds8n\" (UID: \"6eaa7147-1483-4399-ae55-b0f27c44a0e6\") " pod="swift-kuttl-tests/swift-ring-rebalance-xds8n" Mar 07 14:57:41 crc kubenswrapper[4943]: I0307 14:57:41.834455 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6eaa7147-1483-4399-ae55-b0f27c44a0e6-scripts\") pod \"swift-ring-rebalance-xds8n\" (UID: \"6eaa7147-1483-4399-ae55-b0f27c44a0e6\") " pod="swift-kuttl-tests/swift-ring-rebalance-xds8n" Mar 07 14:57:41 crc kubenswrapper[4943]: I0307 14:57:41.834547 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5hgr\" (UniqueName: \"kubernetes.io/projected/b43e5152-86c3-4e1b-8fe1-d440d2e37b82-kube-api-access-k5hgr\") pod \"swift-ring-rebalance-89s8j\" (UID: \"b43e5152-86c3-4e1b-8fe1-d440d2e37b82\") " pod="swift-kuttl-tests/swift-ring-rebalance-89s8j" Mar 07 14:57:41 crc kubenswrapper[4943]: I0307 14:57:41.834586 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6eaa7147-1483-4399-ae55-b0f27c44a0e6-dispersionconf\") pod \"swift-ring-rebalance-xds8n\" (UID: \"6eaa7147-1483-4399-ae55-b0f27c44a0e6\") " pod="swift-kuttl-tests/swift-ring-rebalance-xds8n" Mar 07 14:57:41 crc kubenswrapper[4943]: I0307 14:57:41.834648 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6eaa7147-1483-4399-ae55-b0f27c44a0e6-etc-swift\") pod \"swift-ring-rebalance-xds8n\" (UID: \"6eaa7147-1483-4399-ae55-b0f27c44a0e6\") " pod="swift-kuttl-tests/swift-ring-rebalance-xds8n" Mar 07 14:57:41 crc kubenswrapper[4943]: I0307 14:57:41.834991 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b43e5152-86c3-4e1b-8fe1-d440d2e37b82-ring-data-devices\") pod \"swift-ring-rebalance-89s8j\" (UID: \"b43e5152-86c3-4e1b-8fe1-d440d2e37b82\") " pod="swift-kuttl-tests/swift-ring-rebalance-89s8j" Mar 07 14:57:41 crc kubenswrapper[4943]: I0307 14:57:41.835228 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b43e5152-86c3-4e1b-8fe1-d440d2e37b82-scripts\") pod \"swift-ring-rebalance-89s8j\" (UID: \"b43e5152-86c3-4e1b-8fe1-d440d2e37b82\") " pod="swift-kuttl-tests/swift-ring-rebalance-89s8j" Mar 07 14:57:41 crc kubenswrapper[4943]: I0307 14:57:41.835679 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b43e5152-86c3-4e1b-8fe1-d440d2e37b82-etc-swift\") pod \"swift-ring-rebalance-89s8j\" (UID: \"b43e5152-86c3-4e1b-8fe1-d440d2e37b82\") " pod="swift-kuttl-tests/swift-ring-rebalance-89s8j" Mar 07 14:57:41 crc kubenswrapper[4943]: I0307 14:57:41.841371 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b43e5152-86c3-4e1b-8fe1-d440d2e37b82-swiftconf\") pod \"swift-ring-rebalance-89s8j\" (UID: \"b43e5152-86c3-4e1b-8fe1-d440d2e37b82\") " pod="swift-kuttl-tests/swift-ring-rebalance-89s8j" Mar 07 14:57:41 crc kubenswrapper[4943]: I0307 14:57:41.842265 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b43e5152-86c3-4e1b-8fe1-d440d2e37b82-dispersionconf\") pod \"swift-ring-rebalance-89s8j\" (UID: \"b43e5152-86c3-4e1b-8fe1-d440d2e37b82\") " pod="swift-kuttl-tests/swift-ring-rebalance-89s8j" Mar 07 14:57:41 crc kubenswrapper[4943]: I0307 14:57:41.850072 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5hgr\" (UniqueName: \"kubernetes.io/projected/b43e5152-86c3-4e1b-8fe1-d440d2e37b82-kube-api-access-k5hgr\") pod \"swift-ring-rebalance-89s8j\" (UID: \"b43e5152-86c3-4e1b-8fe1-d440d2e37b82\") " pod="swift-kuttl-tests/swift-ring-rebalance-89s8j" Mar 07 14:57:41 crc kubenswrapper[4943]: I0307 14:57:41.936545 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6eaa7147-1483-4399-ae55-b0f27c44a0e6-dispersionconf\") pod \"swift-ring-rebalance-xds8n\" (UID: \"6eaa7147-1483-4399-ae55-b0f27c44a0e6\") " pod="swift-kuttl-tests/swift-ring-rebalance-xds8n" Mar 07 14:57:41 crc kubenswrapper[4943]: I0307 14:57:41.936651 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6eaa7147-1483-4399-ae55-b0f27c44a0e6-etc-swift\") pod \"swift-ring-rebalance-xds8n\" (UID: \"6eaa7147-1483-4399-ae55-b0f27c44a0e6\") " pod="swift-kuttl-tests/swift-ring-rebalance-xds8n" Mar 07 14:57:41 crc kubenswrapper[4943]: I0307 14:57:41.936745 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6eaa7147-1483-4399-ae55-b0f27c44a0e6-ring-data-devices\") pod \"swift-ring-rebalance-xds8n\" (UID: \"6eaa7147-1483-4399-ae55-b0f27c44a0e6\") " pod="swift-kuttl-tests/swift-ring-rebalance-xds8n" Mar 07 14:57:41 crc kubenswrapper[4943]: I0307 14:57:41.936790 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6eaa7147-1483-4399-ae55-b0f27c44a0e6-swiftconf\") pod \"swift-ring-rebalance-xds8n\" (UID: \"6eaa7147-1483-4399-ae55-b0f27c44a0e6\") " pod="swift-kuttl-tests/swift-ring-rebalance-xds8n" Mar 07 14:57:41 crc kubenswrapper[4943]: I0307 14:57:41.936891 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86v2q\" (UniqueName: \"kubernetes.io/projected/6eaa7147-1483-4399-ae55-b0f27c44a0e6-kube-api-access-86v2q\") pod \"swift-ring-rebalance-xds8n\" (UID: \"6eaa7147-1483-4399-ae55-b0f27c44a0e6\") " pod="swift-kuttl-tests/swift-ring-rebalance-xds8n" Mar 07 14:57:41 crc kubenswrapper[4943]: I0307 14:57:41.936998 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6eaa7147-1483-4399-ae55-b0f27c44a0e6-scripts\") pod \"swift-ring-rebalance-xds8n\" (UID: \"6eaa7147-1483-4399-ae55-b0f27c44a0e6\") " pod="swift-kuttl-tests/swift-ring-rebalance-xds8n" Mar 07 14:57:41 crc kubenswrapper[4943]: I0307 14:57:41.938835 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6eaa7147-1483-4399-ae55-b0f27c44a0e6-etc-swift\") pod \"swift-ring-rebalance-xds8n\" (UID: \"6eaa7147-1483-4399-ae55-b0f27c44a0e6\") " pod="swift-kuttl-tests/swift-ring-rebalance-xds8n" Mar 07 14:57:41 crc kubenswrapper[4943]: I0307 14:57:41.938973 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6eaa7147-1483-4399-ae55-b0f27c44a0e6-scripts\") pod \"swift-ring-rebalance-xds8n\" (UID: \"6eaa7147-1483-4399-ae55-b0f27c44a0e6\") " pod="swift-kuttl-tests/swift-ring-rebalance-xds8n" Mar 07 14:57:41 crc kubenswrapper[4943]: I0307 14:57:41.939046 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6eaa7147-1483-4399-ae55-b0f27c44a0e6-ring-data-devices\") pod \"swift-ring-rebalance-xds8n\" (UID: \"6eaa7147-1483-4399-ae55-b0f27c44a0e6\") " pod="swift-kuttl-tests/swift-ring-rebalance-xds8n" Mar 07 14:57:41 crc kubenswrapper[4943]: I0307 14:57:41.942389 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6eaa7147-1483-4399-ae55-b0f27c44a0e6-dispersionconf\") pod \"swift-ring-rebalance-xds8n\" (UID: \"6eaa7147-1483-4399-ae55-b0f27c44a0e6\") " pod="swift-kuttl-tests/swift-ring-rebalance-xds8n" Mar 07 14:57:41 crc kubenswrapper[4943]: I0307 14:57:41.957147 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6eaa7147-1483-4399-ae55-b0f27c44a0e6-swiftconf\") pod \"swift-ring-rebalance-xds8n\" (UID: \"6eaa7147-1483-4399-ae55-b0f27c44a0e6\") " pod="swift-kuttl-tests/swift-ring-rebalance-xds8n" Mar 07 14:57:41 crc kubenswrapper[4943]: I0307 14:57:41.971994 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86v2q\" (UniqueName: \"kubernetes.io/projected/6eaa7147-1483-4399-ae55-b0f27c44a0e6-kube-api-access-86v2q\") pod \"swift-ring-rebalance-xds8n\" (UID: \"6eaa7147-1483-4399-ae55-b0f27c44a0e6\") " pod="swift-kuttl-tests/swift-ring-rebalance-xds8n" Mar 07 14:57:42 crc kubenswrapper[4943]: I0307 14:57:42.023975 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"db16093b-eede-415e-8145-56e579a8782a","Type":"ContainerStarted","Data":"a89ba23a18bbf856e369b9997db5b15c2e073e42db3f087a7b8b5524d27f75c6"} Mar 07 14:57:42 crc kubenswrapper[4943]: I0307 14:57:42.024050 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"db16093b-eede-415e-8145-56e579a8782a","Type":"ContainerStarted","Data":"968bf9296a3299fbf29b116fdf1d701bb11bdad0ce42eb97dd829de219f7c9d3"} Mar 07 14:57:42 crc kubenswrapper[4943]: I0307 14:57:42.036089 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"0e83c041-7178-43e4-b05d-e1bcc0ec36bd","Type":"ContainerStarted","Data":"7c67a0fd742d59f47a0d1f6a924f182e1e84266f97330d060208d8f777bad883"} Mar 07 14:57:42 crc kubenswrapper[4943]: I0307 14:57:42.036150 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-89s8j" Mar 07 14:57:42 crc kubenswrapper[4943]: I0307 14:57:42.036159 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"0e83c041-7178-43e4-b05d-e1bcc0ec36bd","Type":"ContainerStarted","Data":"737a49457d09cd8479c49fa391100fe4f3b79bf41392edda0bf94b505382ab04"} Mar 07 14:57:42 crc kubenswrapper[4943]: I0307 14:57:42.036293 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"0e83c041-7178-43e4-b05d-e1bcc0ec36bd","Type":"ContainerStarted","Data":"2ee77dbb77dbb65d1523838e6afd9d5a7b72613d080864ff79215b4a508ae442"} Mar 07 14:57:42 crc kubenswrapper[4943]: I0307 14:57:42.042715 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-xds8n" Mar 07 14:57:42 crc kubenswrapper[4943]: I0307 14:57:42.068213 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-89s8j" Mar 07 14:57:42 crc kubenswrapper[4943]: I0307 14:57:42.076390 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-storage-1" podStartSLOduration=6.07636143 podStartE2EDuration="6.07636143s" podCreationTimestamp="2026-03-07 14:57:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:57:42.06943352 +0000 UTC m=+1104.021570048" watchObservedRunningTime="2026-03-07 14:57:42.07636143 +0000 UTC m=+1104.028497968" Mar 07 14:57:42 crc kubenswrapper[4943]: I0307 14:57:42.137470 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-storage-2" podStartSLOduration=6.137448512 podStartE2EDuration="6.137448512s" podCreationTimestamp="2026-03-07 14:57:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:57:42.122424213 +0000 UTC m=+1104.074560731" watchObservedRunningTime="2026-03-07 14:57:42.137448512 +0000 UTC m=+1104.089585020" Mar 07 14:57:42 crc kubenswrapper[4943]: I0307 14:57:42.139447 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b43e5152-86c3-4e1b-8fe1-d440d2e37b82-ring-data-devices\") pod \"b43e5152-86c3-4e1b-8fe1-d440d2e37b82\" (UID: \"b43e5152-86c3-4e1b-8fe1-d440d2e37b82\") " Mar 07 14:57:42 crc kubenswrapper[4943]: I0307 14:57:42.139547 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k5hgr\" (UniqueName: \"kubernetes.io/projected/b43e5152-86c3-4e1b-8fe1-d440d2e37b82-kube-api-access-k5hgr\") pod \"b43e5152-86c3-4e1b-8fe1-d440d2e37b82\" (UID: \"b43e5152-86c3-4e1b-8fe1-d440d2e37b82\") " Mar 07 14:57:42 crc kubenswrapper[4943]: I0307 14:57:42.139636 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b43e5152-86c3-4e1b-8fe1-d440d2e37b82-dispersionconf\") pod \"b43e5152-86c3-4e1b-8fe1-d440d2e37b82\" (UID: \"b43e5152-86c3-4e1b-8fe1-d440d2e37b82\") " Mar 07 14:57:42 crc kubenswrapper[4943]: I0307 14:57:42.139706 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b43e5152-86c3-4e1b-8fe1-d440d2e37b82-etc-swift\") pod \"b43e5152-86c3-4e1b-8fe1-d440d2e37b82\" (UID: \"b43e5152-86c3-4e1b-8fe1-d440d2e37b82\") " Mar 07 14:57:42 crc kubenswrapper[4943]: I0307 14:57:42.139732 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b43e5152-86c3-4e1b-8fe1-d440d2e37b82-swiftconf\") pod \"b43e5152-86c3-4e1b-8fe1-d440d2e37b82\" (UID: \"b43e5152-86c3-4e1b-8fe1-d440d2e37b82\") " Mar 07 14:57:42 crc kubenswrapper[4943]: I0307 14:57:42.139767 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b43e5152-86c3-4e1b-8fe1-d440d2e37b82-scripts\") pod \"b43e5152-86c3-4e1b-8fe1-d440d2e37b82\" (UID: \"b43e5152-86c3-4e1b-8fe1-d440d2e37b82\") " Mar 07 14:57:42 crc kubenswrapper[4943]: I0307 14:57:42.141891 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b43e5152-86c3-4e1b-8fe1-d440d2e37b82-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "b43e5152-86c3-4e1b-8fe1-d440d2e37b82" (UID: "b43e5152-86c3-4e1b-8fe1-d440d2e37b82"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:57:42 crc kubenswrapper[4943]: I0307 14:57:42.144426 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b43e5152-86c3-4e1b-8fe1-d440d2e37b82-scripts" (OuterVolumeSpecName: "scripts") pod "b43e5152-86c3-4e1b-8fe1-d440d2e37b82" (UID: "b43e5152-86c3-4e1b-8fe1-d440d2e37b82"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:57:42 crc kubenswrapper[4943]: I0307 14:57:42.151546 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b43e5152-86c3-4e1b-8fe1-d440d2e37b82-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "b43e5152-86c3-4e1b-8fe1-d440d2e37b82" (UID: "b43e5152-86c3-4e1b-8fe1-d440d2e37b82"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:57:42 crc kubenswrapper[4943]: I0307 14:57:42.152777 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b43e5152-86c3-4e1b-8fe1-d440d2e37b82-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "b43e5152-86c3-4e1b-8fe1-d440d2e37b82" (UID: "b43e5152-86c3-4e1b-8fe1-d440d2e37b82"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:57:42 crc kubenswrapper[4943]: I0307 14:57:42.154734 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b43e5152-86c3-4e1b-8fe1-d440d2e37b82-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "b43e5152-86c3-4e1b-8fe1-d440d2e37b82" (UID: "b43e5152-86c3-4e1b-8fe1-d440d2e37b82"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:57:42 crc kubenswrapper[4943]: I0307 14:57:42.155134 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b43e5152-86c3-4e1b-8fe1-d440d2e37b82-kube-api-access-k5hgr" (OuterVolumeSpecName: "kube-api-access-k5hgr") pod "b43e5152-86c3-4e1b-8fe1-d440d2e37b82" (UID: "b43e5152-86c3-4e1b-8fe1-d440d2e37b82"). InnerVolumeSpecName "kube-api-access-k5hgr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:57:42 crc kubenswrapper[4943]: I0307 14:57:42.241361 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b43e5152-86c3-4e1b-8fe1-d440d2e37b82-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 14:57:42 crc kubenswrapper[4943]: I0307 14:57:42.241385 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b43e5152-86c3-4e1b-8fe1-d440d2e37b82-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 14:57:42 crc kubenswrapper[4943]: I0307 14:57:42.241394 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b43e5152-86c3-4e1b-8fe1-d440d2e37b82-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 14:57:42 crc kubenswrapper[4943]: I0307 14:57:42.241404 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b43e5152-86c3-4e1b-8fe1-d440d2e37b82-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 14:57:42 crc kubenswrapper[4943]: I0307 14:57:42.241412 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b43e5152-86c3-4e1b-8fe1-d440d2e37b82-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 14:57:42 crc kubenswrapper[4943]: I0307 14:57:42.241421 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k5hgr\" (UniqueName: \"kubernetes.io/projected/b43e5152-86c3-4e1b-8fe1-d440d2e37b82-kube-api-access-k5hgr\") on node \"crc\" DevicePath \"\"" Mar 07 14:57:42 crc kubenswrapper[4943]: I0307 14:57:42.582999 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-xds8n"] Mar 07 14:57:42 crc kubenswrapper[4943]: W0307 14:57:42.587398 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6eaa7147_1483_4399_ae55_b0f27c44a0e6.slice/crio-3d8b140ebdab2a0fd5604fff15e35d3556845e247fc488e4735417a767f1802a WatchSource:0}: Error finding container 3d8b140ebdab2a0fd5604fff15e35d3556845e247fc488e4735417a767f1802a: Status 404 returned error can't find the container with id 3d8b140ebdab2a0fd5604fff15e35d3556845e247fc488e4735417a767f1802a Mar 07 14:57:42 crc kubenswrapper[4943]: I0307 14:57:42.773691 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d105da1c-5832-49c5-8c86-c12c20f3983f" path="/var/lib/kubelet/pods/d105da1c-5832-49c5-8c86-c12c20f3983f/volumes" Mar 07 14:57:43 crc kubenswrapper[4943]: I0307 14:57:43.049816 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-xds8n" event={"ID":"6eaa7147-1483-4399-ae55-b0f27c44a0e6","Type":"ContainerStarted","Data":"7122613220e2db80a53f9f6dce66813c8891cd55870903238eb8dae123d96919"} Mar 07 14:57:43 crc kubenswrapper[4943]: I0307 14:57:43.049892 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-xds8n" event={"ID":"6eaa7147-1483-4399-ae55-b0f27c44a0e6","Type":"ContainerStarted","Data":"3d8b140ebdab2a0fd5604fff15e35d3556845e247fc488e4735417a767f1802a"} Mar 07 14:57:43 crc kubenswrapper[4943]: I0307 14:57:43.052547 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-89s8j" Mar 07 14:57:43 crc kubenswrapper[4943]: I0307 14:57:43.092002 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-xds8n" podStartSLOduration=2.091966962 podStartE2EDuration="2.091966962s" podCreationTimestamp="2026-03-07 14:57:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:57:43.079567819 +0000 UTC m=+1105.031704357" watchObservedRunningTime="2026-03-07 14:57:43.091966962 +0000 UTC m=+1105.044103490" Mar 07 14:57:43 crc kubenswrapper[4943]: I0307 14:57:43.136062 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-89s8j"] Mar 07 14:57:43 crc kubenswrapper[4943]: I0307 14:57:43.144165 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-89s8j"] Mar 07 14:57:44 crc kubenswrapper[4943]: I0307 14:57:44.772236 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b43e5152-86c3-4e1b-8fe1-d440d2e37b82" path="/var/lib/kubelet/pods/b43e5152-86c3-4e1b-8fe1-d440d2e37b82/volumes" Mar 07 14:57:52 crc kubenswrapper[4943]: I0307 14:57:52.147459 4943 generic.go:334] "Generic (PLEG): container finished" podID="6eaa7147-1483-4399-ae55-b0f27c44a0e6" containerID="7122613220e2db80a53f9f6dce66813c8891cd55870903238eb8dae123d96919" exitCode=0 Mar 07 14:57:52 crc kubenswrapper[4943]: I0307 14:57:52.147528 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-xds8n" event={"ID":"6eaa7147-1483-4399-ae55-b0f27c44a0e6","Type":"ContainerDied","Data":"7122613220e2db80a53f9f6dce66813c8891cd55870903238eb8dae123d96919"} Mar 07 14:57:53 crc kubenswrapper[4943]: I0307 14:57:53.514785 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-xds8n" Mar 07 14:57:53 crc kubenswrapper[4943]: I0307 14:57:53.714774 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6eaa7147-1483-4399-ae55-b0f27c44a0e6-swiftconf\") pod \"6eaa7147-1483-4399-ae55-b0f27c44a0e6\" (UID: \"6eaa7147-1483-4399-ae55-b0f27c44a0e6\") " Mar 07 14:57:53 crc kubenswrapper[4943]: I0307 14:57:53.715696 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6eaa7147-1483-4399-ae55-b0f27c44a0e6-dispersionconf\") pod \"6eaa7147-1483-4399-ae55-b0f27c44a0e6\" (UID: \"6eaa7147-1483-4399-ae55-b0f27c44a0e6\") " Mar 07 14:57:53 crc kubenswrapper[4943]: I0307 14:57:53.715788 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-86v2q\" (UniqueName: \"kubernetes.io/projected/6eaa7147-1483-4399-ae55-b0f27c44a0e6-kube-api-access-86v2q\") pod \"6eaa7147-1483-4399-ae55-b0f27c44a0e6\" (UID: \"6eaa7147-1483-4399-ae55-b0f27c44a0e6\") " Mar 07 14:57:53 crc kubenswrapper[4943]: I0307 14:57:53.715857 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6eaa7147-1483-4399-ae55-b0f27c44a0e6-scripts\") pod \"6eaa7147-1483-4399-ae55-b0f27c44a0e6\" (UID: \"6eaa7147-1483-4399-ae55-b0f27c44a0e6\") " Mar 07 14:57:53 crc kubenswrapper[4943]: I0307 14:57:53.716321 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6eaa7147-1483-4399-ae55-b0f27c44a0e6-ring-data-devices\") pod \"6eaa7147-1483-4399-ae55-b0f27c44a0e6\" (UID: \"6eaa7147-1483-4399-ae55-b0f27c44a0e6\") " Mar 07 14:57:53 crc kubenswrapper[4943]: I0307 14:57:53.716415 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6eaa7147-1483-4399-ae55-b0f27c44a0e6-etc-swift\") pod \"6eaa7147-1483-4399-ae55-b0f27c44a0e6\" (UID: \"6eaa7147-1483-4399-ae55-b0f27c44a0e6\") " Mar 07 14:57:53 crc kubenswrapper[4943]: I0307 14:57:53.716898 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6eaa7147-1483-4399-ae55-b0f27c44a0e6-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "6eaa7147-1483-4399-ae55-b0f27c44a0e6" (UID: "6eaa7147-1483-4399-ae55-b0f27c44a0e6"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:57:53 crc kubenswrapper[4943]: I0307 14:57:53.717458 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6eaa7147-1483-4399-ae55-b0f27c44a0e6-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 14:57:53 crc kubenswrapper[4943]: I0307 14:57:53.718036 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6eaa7147-1483-4399-ae55-b0f27c44a0e6-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "6eaa7147-1483-4399-ae55-b0f27c44a0e6" (UID: "6eaa7147-1483-4399-ae55-b0f27c44a0e6"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:57:53 crc kubenswrapper[4943]: I0307 14:57:53.723213 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6eaa7147-1483-4399-ae55-b0f27c44a0e6-kube-api-access-86v2q" (OuterVolumeSpecName: "kube-api-access-86v2q") pod "6eaa7147-1483-4399-ae55-b0f27c44a0e6" (UID: "6eaa7147-1483-4399-ae55-b0f27c44a0e6"). InnerVolumeSpecName "kube-api-access-86v2q". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:57:53 crc kubenswrapper[4943]: I0307 14:57:53.746522 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6eaa7147-1483-4399-ae55-b0f27c44a0e6-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "6eaa7147-1483-4399-ae55-b0f27c44a0e6" (UID: "6eaa7147-1483-4399-ae55-b0f27c44a0e6"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:57:53 crc kubenswrapper[4943]: I0307 14:57:53.750345 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6eaa7147-1483-4399-ae55-b0f27c44a0e6-scripts" (OuterVolumeSpecName: "scripts") pod "6eaa7147-1483-4399-ae55-b0f27c44a0e6" (UID: "6eaa7147-1483-4399-ae55-b0f27c44a0e6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:57:53 crc kubenswrapper[4943]: I0307 14:57:53.765552 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6eaa7147-1483-4399-ae55-b0f27c44a0e6-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "6eaa7147-1483-4399-ae55-b0f27c44a0e6" (UID: "6eaa7147-1483-4399-ae55-b0f27c44a0e6"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:57:53 crc kubenswrapper[4943]: I0307 14:57:53.819033 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-86v2q\" (UniqueName: \"kubernetes.io/projected/6eaa7147-1483-4399-ae55-b0f27c44a0e6-kube-api-access-86v2q\") on node \"crc\" DevicePath \"\"" Mar 07 14:57:53 crc kubenswrapper[4943]: I0307 14:57:53.819085 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6eaa7147-1483-4399-ae55-b0f27c44a0e6-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 14:57:53 crc kubenswrapper[4943]: I0307 14:57:53.819104 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6eaa7147-1483-4399-ae55-b0f27c44a0e6-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 14:57:53 crc kubenswrapper[4943]: I0307 14:57:53.819121 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6eaa7147-1483-4399-ae55-b0f27c44a0e6-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 14:57:53 crc kubenswrapper[4943]: I0307 14:57:53.819137 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6eaa7147-1483-4399-ae55-b0f27c44a0e6-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 14:57:54 crc kubenswrapper[4943]: I0307 14:57:54.180074 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-xds8n" event={"ID":"6eaa7147-1483-4399-ae55-b0f27c44a0e6","Type":"ContainerDied","Data":"3d8b140ebdab2a0fd5604fff15e35d3556845e247fc488e4735417a767f1802a"} Mar 07 14:57:54 crc kubenswrapper[4943]: I0307 14:57:54.180125 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3d8b140ebdab2a0fd5604fff15e35d3556845e247fc488e4735417a767f1802a" Mar 07 14:57:54 crc kubenswrapper[4943]: I0307 14:57:54.180280 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-xds8n" Mar 07 14:57:54 crc kubenswrapper[4943]: I0307 14:57:54.421090 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-m5fhr"] Mar 07 14:57:54 crc kubenswrapper[4943]: E0307 14:57:54.421487 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6eaa7147-1483-4399-ae55-b0f27c44a0e6" containerName="swift-ring-rebalance" Mar 07 14:57:54 crc kubenswrapper[4943]: I0307 14:57:54.421506 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="6eaa7147-1483-4399-ae55-b0f27c44a0e6" containerName="swift-ring-rebalance" Mar 07 14:57:54 crc kubenswrapper[4943]: I0307 14:57:54.421684 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="6eaa7147-1483-4399-ae55-b0f27c44a0e6" containerName="swift-ring-rebalance" Mar 07 14:57:54 crc kubenswrapper[4943]: I0307 14:57:54.422311 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-m5fhr" Mar 07 14:57:54 crc kubenswrapper[4943]: I0307 14:57:54.425097 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 14:57:54 crc kubenswrapper[4943]: I0307 14:57:54.425200 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 14:57:54 crc kubenswrapper[4943]: I0307 14:57:54.443996 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-m5fhr"] Mar 07 14:57:54 crc kubenswrapper[4943]: I0307 14:57:54.528765 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/aabde81f-aac9-492f-b8f1-8193cac34b4f-etc-swift\") pod \"swift-ring-rebalance-debug-m5fhr\" (UID: \"aabde81f-aac9-492f-b8f1-8193cac34b4f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m5fhr" Mar 07 14:57:54 crc kubenswrapper[4943]: I0307 14:57:54.528828 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/aabde81f-aac9-492f-b8f1-8193cac34b4f-ring-data-devices\") pod \"swift-ring-rebalance-debug-m5fhr\" (UID: \"aabde81f-aac9-492f-b8f1-8193cac34b4f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m5fhr" Mar 07 14:57:54 crc kubenswrapper[4943]: I0307 14:57:54.529032 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cnxpc\" (UniqueName: \"kubernetes.io/projected/aabde81f-aac9-492f-b8f1-8193cac34b4f-kube-api-access-cnxpc\") pod \"swift-ring-rebalance-debug-m5fhr\" (UID: \"aabde81f-aac9-492f-b8f1-8193cac34b4f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m5fhr" Mar 07 14:57:54 crc kubenswrapper[4943]: I0307 14:57:54.529078 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/aabde81f-aac9-492f-b8f1-8193cac34b4f-swiftconf\") pod \"swift-ring-rebalance-debug-m5fhr\" (UID: \"aabde81f-aac9-492f-b8f1-8193cac34b4f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m5fhr" Mar 07 14:57:54 crc kubenswrapper[4943]: I0307 14:57:54.529102 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/aabde81f-aac9-492f-b8f1-8193cac34b4f-dispersionconf\") pod \"swift-ring-rebalance-debug-m5fhr\" (UID: \"aabde81f-aac9-492f-b8f1-8193cac34b4f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m5fhr" Mar 07 14:57:54 crc kubenswrapper[4943]: I0307 14:57:54.529124 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aabde81f-aac9-492f-b8f1-8193cac34b4f-scripts\") pod \"swift-ring-rebalance-debug-m5fhr\" (UID: \"aabde81f-aac9-492f-b8f1-8193cac34b4f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m5fhr" Mar 07 14:57:54 crc kubenswrapper[4943]: I0307 14:57:54.630860 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/aabde81f-aac9-492f-b8f1-8193cac34b4f-swiftconf\") pod \"swift-ring-rebalance-debug-m5fhr\" (UID: \"aabde81f-aac9-492f-b8f1-8193cac34b4f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m5fhr" Mar 07 14:57:54 crc kubenswrapper[4943]: I0307 14:57:54.630968 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/aabde81f-aac9-492f-b8f1-8193cac34b4f-dispersionconf\") pod \"swift-ring-rebalance-debug-m5fhr\" (UID: \"aabde81f-aac9-492f-b8f1-8193cac34b4f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m5fhr" Mar 07 14:57:54 crc kubenswrapper[4943]: I0307 14:57:54.631015 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aabde81f-aac9-492f-b8f1-8193cac34b4f-scripts\") pod \"swift-ring-rebalance-debug-m5fhr\" (UID: \"aabde81f-aac9-492f-b8f1-8193cac34b4f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m5fhr" Mar 07 14:57:54 crc kubenswrapper[4943]: I0307 14:57:54.631133 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/aabde81f-aac9-492f-b8f1-8193cac34b4f-etc-swift\") pod \"swift-ring-rebalance-debug-m5fhr\" (UID: \"aabde81f-aac9-492f-b8f1-8193cac34b4f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m5fhr" Mar 07 14:57:54 crc kubenswrapper[4943]: I0307 14:57:54.631177 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/aabde81f-aac9-492f-b8f1-8193cac34b4f-ring-data-devices\") pod \"swift-ring-rebalance-debug-m5fhr\" (UID: \"aabde81f-aac9-492f-b8f1-8193cac34b4f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m5fhr" Mar 07 14:57:54 crc kubenswrapper[4943]: I0307 14:57:54.631334 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cnxpc\" (UniqueName: \"kubernetes.io/projected/aabde81f-aac9-492f-b8f1-8193cac34b4f-kube-api-access-cnxpc\") pod \"swift-ring-rebalance-debug-m5fhr\" (UID: \"aabde81f-aac9-492f-b8f1-8193cac34b4f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m5fhr" Mar 07 14:57:54 crc kubenswrapper[4943]: I0307 14:57:54.632280 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/aabde81f-aac9-492f-b8f1-8193cac34b4f-etc-swift\") pod \"swift-ring-rebalance-debug-m5fhr\" (UID: \"aabde81f-aac9-492f-b8f1-8193cac34b4f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m5fhr" Mar 07 14:57:54 crc kubenswrapper[4943]: I0307 14:57:54.632642 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/aabde81f-aac9-492f-b8f1-8193cac34b4f-ring-data-devices\") pod \"swift-ring-rebalance-debug-m5fhr\" (UID: \"aabde81f-aac9-492f-b8f1-8193cac34b4f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m5fhr" Mar 07 14:57:54 crc kubenswrapper[4943]: I0307 14:57:54.633179 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aabde81f-aac9-492f-b8f1-8193cac34b4f-scripts\") pod \"swift-ring-rebalance-debug-m5fhr\" (UID: \"aabde81f-aac9-492f-b8f1-8193cac34b4f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m5fhr" Mar 07 14:57:54 crc kubenswrapper[4943]: I0307 14:57:54.640294 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/aabde81f-aac9-492f-b8f1-8193cac34b4f-swiftconf\") pod \"swift-ring-rebalance-debug-m5fhr\" (UID: \"aabde81f-aac9-492f-b8f1-8193cac34b4f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m5fhr" Mar 07 14:57:54 crc kubenswrapper[4943]: I0307 14:57:54.640883 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/aabde81f-aac9-492f-b8f1-8193cac34b4f-dispersionconf\") pod \"swift-ring-rebalance-debug-m5fhr\" (UID: \"aabde81f-aac9-492f-b8f1-8193cac34b4f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m5fhr" Mar 07 14:57:54 crc kubenswrapper[4943]: I0307 14:57:54.650686 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cnxpc\" (UniqueName: \"kubernetes.io/projected/aabde81f-aac9-492f-b8f1-8193cac34b4f-kube-api-access-cnxpc\") pod \"swift-ring-rebalance-debug-m5fhr\" (UID: \"aabde81f-aac9-492f-b8f1-8193cac34b4f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m5fhr" Mar 07 14:57:54 crc kubenswrapper[4943]: I0307 14:57:54.736611 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-m5fhr" Mar 07 14:57:55 crc kubenswrapper[4943]: I0307 14:57:55.240265 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-m5fhr"] Mar 07 14:57:55 crc kubenswrapper[4943]: W0307 14:57:55.261542 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaabde81f_aac9_492f_b8f1_8193cac34b4f.slice/crio-e0fb5583204a65b6d0fba46e6edb4cc3b1494de6bc9ae491a532513ddbf8a306 WatchSource:0}: Error finding container e0fb5583204a65b6d0fba46e6edb4cc3b1494de6bc9ae491a532513ddbf8a306: Status 404 returned error can't find the container with id e0fb5583204a65b6d0fba46e6edb4cc3b1494de6bc9ae491a532513ddbf8a306 Mar 07 14:57:56 crc kubenswrapper[4943]: I0307 14:57:56.201891 4943 generic.go:334] "Generic (PLEG): container finished" podID="aabde81f-aac9-492f-b8f1-8193cac34b4f" containerID="39de95413cade901e51f50b2d8f9abb6980ab8da312acae8fa47accbc002bb30" exitCode=0 Mar 07 14:57:56 crc kubenswrapper[4943]: I0307 14:57:56.202079 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-m5fhr" event={"ID":"aabde81f-aac9-492f-b8f1-8193cac34b4f","Type":"ContainerDied","Data":"39de95413cade901e51f50b2d8f9abb6980ab8da312acae8fa47accbc002bb30"} Mar 07 14:57:56 crc kubenswrapper[4943]: I0307 14:57:56.202349 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-m5fhr" event={"ID":"aabde81f-aac9-492f-b8f1-8193cac34b4f","Type":"ContainerStarted","Data":"e0fb5583204a65b6d0fba46e6edb4cc3b1494de6bc9ae491a532513ddbf8a306"} Mar 07 14:57:57 crc kubenswrapper[4943]: I0307 14:57:57.163330 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-m5fhr"] Mar 07 14:57:57 crc kubenswrapper[4943]: I0307 14:57:57.193637 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-m5fhr"] Mar 07 14:57:57 crc kubenswrapper[4943]: I0307 14:57:57.484477 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-m5fhr" Mar 07 14:57:57 crc kubenswrapper[4943]: I0307 14:57:57.496673 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/aabde81f-aac9-492f-b8f1-8193cac34b4f-ring-data-devices\") pod \"aabde81f-aac9-492f-b8f1-8193cac34b4f\" (UID: \"aabde81f-aac9-492f-b8f1-8193cac34b4f\") " Mar 07 14:57:57 crc kubenswrapper[4943]: I0307 14:57:57.496741 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aabde81f-aac9-492f-b8f1-8193cac34b4f-scripts\") pod \"aabde81f-aac9-492f-b8f1-8193cac34b4f\" (UID: \"aabde81f-aac9-492f-b8f1-8193cac34b4f\") " Mar 07 14:57:57 crc kubenswrapper[4943]: I0307 14:57:57.496820 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/aabde81f-aac9-492f-b8f1-8193cac34b4f-swiftconf\") pod \"aabde81f-aac9-492f-b8f1-8193cac34b4f\" (UID: \"aabde81f-aac9-492f-b8f1-8193cac34b4f\") " Mar 07 14:57:57 crc kubenswrapper[4943]: I0307 14:57:57.496851 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/aabde81f-aac9-492f-b8f1-8193cac34b4f-etc-swift\") pod \"aabde81f-aac9-492f-b8f1-8193cac34b4f\" (UID: \"aabde81f-aac9-492f-b8f1-8193cac34b4f\") " Mar 07 14:57:57 crc kubenswrapper[4943]: I0307 14:57:57.496918 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/aabde81f-aac9-492f-b8f1-8193cac34b4f-dispersionconf\") pod \"aabde81f-aac9-492f-b8f1-8193cac34b4f\" (UID: \"aabde81f-aac9-492f-b8f1-8193cac34b4f\") " Mar 07 14:57:57 crc kubenswrapper[4943]: I0307 14:57:57.496983 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cnxpc\" (UniqueName: \"kubernetes.io/projected/aabde81f-aac9-492f-b8f1-8193cac34b4f-kube-api-access-cnxpc\") pod \"aabde81f-aac9-492f-b8f1-8193cac34b4f\" (UID: \"aabde81f-aac9-492f-b8f1-8193cac34b4f\") " Mar 07 14:57:57 crc kubenswrapper[4943]: I0307 14:57:57.497307 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aabde81f-aac9-492f-b8f1-8193cac34b4f-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "aabde81f-aac9-492f-b8f1-8193cac34b4f" (UID: "aabde81f-aac9-492f-b8f1-8193cac34b4f"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:57:57 crc kubenswrapper[4943]: I0307 14:57:57.497667 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aabde81f-aac9-492f-b8f1-8193cac34b4f-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "aabde81f-aac9-492f-b8f1-8193cac34b4f" (UID: "aabde81f-aac9-492f-b8f1-8193cac34b4f"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:57:57 crc kubenswrapper[4943]: I0307 14:57:57.502462 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aabde81f-aac9-492f-b8f1-8193cac34b4f-kube-api-access-cnxpc" (OuterVolumeSpecName: "kube-api-access-cnxpc") pod "aabde81f-aac9-492f-b8f1-8193cac34b4f" (UID: "aabde81f-aac9-492f-b8f1-8193cac34b4f"). InnerVolumeSpecName "kube-api-access-cnxpc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:57:57 crc kubenswrapper[4943]: I0307 14:57:57.518057 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aabde81f-aac9-492f-b8f1-8193cac34b4f-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "aabde81f-aac9-492f-b8f1-8193cac34b4f" (UID: "aabde81f-aac9-492f-b8f1-8193cac34b4f"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:57:57 crc kubenswrapper[4943]: I0307 14:57:57.520495 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aabde81f-aac9-492f-b8f1-8193cac34b4f-scripts" (OuterVolumeSpecName: "scripts") pod "aabde81f-aac9-492f-b8f1-8193cac34b4f" (UID: "aabde81f-aac9-492f-b8f1-8193cac34b4f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:57:57 crc kubenswrapper[4943]: I0307 14:57:57.525749 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aabde81f-aac9-492f-b8f1-8193cac34b4f-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "aabde81f-aac9-492f-b8f1-8193cac34b4f" (UID: "aabde81f-aac9-492f-b8f1-8193cac34b4f"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:57:57 crc kubenswrapper[4943]: I0307 14:57:57.598214 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/aabde81f-aac9-492f-b8f1-8193cac34b4f-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 14:57:57 crc kubenswrapper[4943]: I0307 14:57:57.598286 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cnxpc\" (UniqueName: \"kubernetes.io/projected/aabde81f-aac9-492f-b8f1-8193cac34b4f-kube-api-access-cnxpc\") on node \"crc\" DevicePath \"\"" Mar 07 14:57:57 crc kubenswrapper[4943]: I0307 14:57:57.598302 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/aabde81f-aac9-492f-b8f1-8193cac34b4f-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 14:57:57 crc kubenswrapper[4943]: I0307 14:57:57.598317 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aabde81f-aac9-492f-b8f1-8193cac34b4f-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 14:57:57 crc kubenswrapper[4943]: I0307 14:57:57.598328 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/aabde81f-aac9-492f-b8f1-8193cac34b4f-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 14:57:57 crc kubenswrapper[4943]: I0307 14:57:57.598340 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/aabde81f-aac9-492f-b8f1-8193cac34b4f-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 14:57:58 crc kubenswrapper[4943]: I0307 14:57:58.223179 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e0fb5583204a65b6d0fba46e6edb4cc3b1494de6bc9ae491a532513ddbf8a306" Mar 07 14:57:58 crc kubenswrapper[4943]: I0307 14:57:58.223310 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-m5fhr" Mar 07 14:57:58 crc kubenswrapper[4943]: I0307 14:57:58.590375 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-dxspq"] Mar 07 14:57:58 crc kubenswrapper[4943]: E0307 14:57:58.590994 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aabde81f-aac9-492f-b8f1-8193cac34b4f" containerName="swift-ring-rebalance" Mar 07 14:57:58 crc kubenswrapper[4943]: I0307 14:57:58.591008 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="aabde81f-aac9-492f-b8f1-8193cac34b4f" containerName="swift-ring-rebalance" Mar 07 14:57:58 crc kubenswrapper[4943]: I0307 14:57:58.591171 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="aabde81f-aac9-492f-b8f1-8193cac34b4f" containerName="swift-ring-rebalance" Mar 07 14:57:58 crc kubenswrapper[4943]: I0307 14:57:58.591671 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-dxspq" Mar 07 14:57:58 crc kubenswrapper[4943]: I0307 14:57:58.594200 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 14:57:58 crc kubenswrapper[4943]: I0307 14:57:58.595337 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 14:57:58 crc kubenswrapper[4943]: I0307 14:57:58.607010 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-dxspq"] Mar 07 14:57:58 crc kubenswrapper[4943]: I0307 14:57:58.616322 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2907f386-4042-4031-bbb1-327ac38b4173-dispersionconf\") pod \"swift-ring-rebalance-debug-dxspq\" (UID: \"2907f386-4042-4031-bbb1-327ac38b4173\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dxspq" Mar 07 14:57:58 crc kubenswrapper[4943]: I0307 14:57:58.616407 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2907f386-4042-4031-bbb1-327ac38b4173-etc-swift\") pod \"swift-ring-rebalance-debug-dxspq\" (UID: \"2907f386-4042-4031-bbb1-327ac38b4173\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dxspq" Mar 07 14:57:58 crc kubenswrapper[4943]: I0307 14:57:58.616506 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2907f386-4042-4031-bbb1-327ac38b4173-swiftconf\") pod \"swift-ring-rebalance-debug-dxspq\" (UID: \"2907f386-4042-4031-bbb1-327ac38b4173\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dxspq" Mar 07 14:57:58 crc kubenswrapper[4943]: I0307 14:57:58.616552 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2907f386-4042-4031-bbb1-327ac38b4173-scripts\") pod \"swift-ring-rebalance-debug-dxspq\" (UID: \"2907f386-4042-4031-bbb1-327ac38b4173\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dxspq" Mar 07 14:57:58 crc kubenswrapper[4943]: I0307 14:57:58.616604 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2907f386-4042-4031-bbb1-327ac38b4173-ring-data-devices\") pod \"swift-ring-rebalance-debug-dxspq\" (UID: \"2907f386-4042-4031-bbb1-327ac38b4173\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dxspq" Mar 07 14:57:58 crc kubenswrapper[4943]: I0307 14:57:58.616643 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfnhz\" (UniqueName: \"kubernetes.io/projected/2907f386-4042-4031-bbb1-327ac38b4173-kube-api-access-vfnhz\") pod \"swift-ring-rebalance-debug-dxspq\" (UID: \"2907f386-4042-4031-bbb1-327ac38b4173\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dxspq" Mar 07 14:57:58 crc kubenswrapper[4943]: I0307 14:57:58.718312 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2907f386-4042-4031-bbb1-327ac38b4173-dispersionconf\") pod \"swift-ring-rebalance-debug-dxspq\" (UID: \"2907f386-4042-4031-bbb1-327ac38b4173\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dxspq" Mar 07 14:57:58 crc kubenswrapper[4943]: I0307 14:57:58.718362 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2907f386-4042-4031-bbb1-327ac38b4173-etc-swift\") pod \"swift-ring-rebalance-debug-dxspq\" (UID: \"2907f386-4042-4031-bbb1-327ac38b4173\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dxspq" Mar 07 14:57:58 crc kubenswrapper[4943]: I0307 14:57:58.718392 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2907f386-4042-4031-bbb1-327ac38b4173-swiftconf\") pod \"swift-ring-rebalance-debug-dxspq\" (UID: \"2907f386-4042-4031-bbb1-327ac38b4173\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dxspq" Mar 07 14:57:58 crc kubenswrapper[4943]: I0307 14:57:58.718412 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2907f386-4042-4031-bbb1-327ac38b4173-scripts\") pod \"swift-ring-rebalance-debug-dxspq\" (UID: \"2907f386-4042-4031-bbb1-327ac38b4173\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dxspq" Mar 07 14:57:58 crc kubenswrapper[4943]: I0307 14:57:58.718438 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2907f386-4042-4031-bbb1-327ac38b4173-ring-data-devices\") pod \"swift-ring-rebalance-debug-dxspq\" (UID: \"2907f386-4042-4031-bbb1-327ac38b4173\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dxspq" Mar 07 14:57:58 crc kubenswrapper[4943]: I0307 14:57:58.718456 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfnhz\" (UniqueName: \"kubernetes.io/projected/2907f386-4042-4031-bbb1-327ac38b4173-kube-api-access-vfnhz\") pod \"swift-ring-rebalance-debug-dxspq\" (UID: \"2907f386-4042-4031-bbb1-327ac38b4173\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dxspq" Mar 07 14:57:58 crc kubenswrapper[4943]: I0307 14:57:58.719839 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2907f386-4042-4031-bbb1-327ac38b4173-etc-swift\") pod \"swift-ring-rebalance-debug-dxspq\" (UID: \"2907f386-4042-4031-bbb1-327ac38b4173\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dxspq" Mar 07 14:57:58 crc kubenswrapper[4943]: I0307 14:57:58.720204 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2907f386-4042-4031-bbb1-327ac38b4173-scripts\") pod \"swift-ring-rebalance-debug-dxspq\" (UID: \"2907f386-4042-4031-bbb1-327ac38b4173\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dxspq" Mar 07 14:57:58 crc kubenswrapper[4943]: I0307 14:57:58.720203 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2907f386-4042-4031-bbb1-327ac38b4173-ring-data-devices\") pod \"swift-ring-rebalance-debug-dxspq\" (UID: \"2907f386-4042-4031-bbb1-327ac38b4173\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dxspq" Mar 07 14:57:58 crc kubenswrapper[4943]: I0307 14:57:58.728668 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2907f386-4042-4031-bbb1-327ac38b4173-dispersionconf\") pod \"swift-ring-rebalance-debug-dxspq\" (UID: \"2907f386-4042-4031-bbb1-327ac38b4173\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dxspq" Mar 07 14:57:58 crc kubenswrapper[4943]: I0307 14:57:58.731179 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2907f386-4042-4031-bbb1-327ac38b4173-swiftconf\") pod \"swift-ring-rebalance-debug-dxspq\" (UID: \"2907f386-4042-4031-bbb1-327ac38b4173\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dxspq" Mar 07 14:57:58 crc kubenswrapper[4943]: I0307 14:57:58.745986 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfnhz\" (UniqueName: \"kubernetes.io/projected/2907f386-4042-4031-bbb1-327ac38b4173-kube-api-access-vfnhz\") pod \"swift-ring-rebalance-debug-dxspq\" (UID: \"2907f386-4042-4031-bbb1-327ac38b4173\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dxspq" Mar 07 14:57:58 crc kubenswrapper[4943]: I0307 14:57:58.771768 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aabde81f-aac9-492f-b8f1-8193cac34b4f" path="/var/lib/kubelet/pods/aabde81f-aac9-492f-b8f1-8193cac34b4f/volumes" Mar 07 14:57:58 crc kubenswrapper[4943]: I0307 14:57:58.921065 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-dxspq" Mar 07 14:57:59 crc kubenswrapper[4943]: I0307 14:57:59.237284 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-dxspq"] Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.125678 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29548258-spxr6"] Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.127199 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548258-spxr6" Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.134130 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29548258-spxr6"] Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.163141 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-qqlk5" Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.163330 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.163426 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.246024 4943 generic.go:334] "Generic (PLEG): container finished" podID="2907f386-4042-4031-bbb1-327ac38b4173" containerID="326cea0169c4db33a7d753742d878523652301a137ebd599243c6d0c184e2a3e" exitCode=0 Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.246068 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-dxspq" event={"ID":"2907f386-4042-4031-bbb1-327ac38b4173","Type":"ContainerDied","Data":"326cea0169c4db33a7d753742d878523652301a137ebd599243c6d0c184e2a3e"} Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.246095 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-dxspq" event={"ID":"2907f386-4042-4031-bbb1-327ac38b4173","Type":"ContainerStarted","Data":"b3e66036d103ed0df83c5d3e260a6f125fcdae26acb083b66c6f46c8ce781da6"} Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.263785 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hk686\" (UniqueName: \"kubernetes.io/projected/6621055b-9adc-4057-bc14-63b4848edbca-kube-api-access-hk686\") pod \"auto-csr-approver-29548258-spxr6\" (UID: \"6621055b-9adc-4057-bc14-63b4848edbca\") " pod="openshift-infra/auto-csr-approver-29548258-spxr6" Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.305280 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-dxspq"] Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.315396 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-dxspq"] Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.366600 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hk686\" (UniqueName: \"kubernetes.io/projected/6621055b-9adc-4057-bc14-63b4848edbca-kube-api-access-hk686\") pod \"auto-csr-approver-29548258-spxr6\" (UID: \"6621055b-9adc-4057-bc14-63b4848edbca\") " pod="openshift-infra/auto-csr-approver-29548258-spxr6" Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.398002 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-storage-2"] Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.398486 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-2" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="account-server" containerID="cri-o://378e9e0fc8363d06ebee218138d567cd679e69eab4d30a8c35d8d6ba0c009165" gracePeriod=30 Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.398860 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-2" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="swift-recon-cron" containerID="cri-o://7c67a0fd742d59f47a0d1f6a924f182e1e84266f97330d060208d8f777bad883" gracePeriod=30 Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.398916 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-2" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="rsync" containerID="cri-o://737a49457d09cd8479c49fa391100fe4f3b79bf41392edda0bf94b505382ab04" gracePeriod=30 Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.398968 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-2" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="object-expirer" containerID="cri-o://2ee77dbb77dbb65d1523838e6afd9d5a7b72613d080864ff79215b4a508ae442" gracePeriod=30 Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.399003 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-2" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="object-updater" containerID="cri-o://ba8857ea0c02e4a8248cffb1f9d8e7c2b66181c038e948a8f69e9f91e89979a7" gracePeriod=30 Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.399037 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-2" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="object-auditor" containerID="cri-o://7b3105463e57fbdbaa2e126058c6aaa054286f515ac013e5b967046d33b29603" gracePeriod=30 Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.399069 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-2" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="object-replicator" containerID="cri-o://e571f26eb8b825f06454b7946c558c7a1eb1f69258f0d8ada3b654c53c97b55b" gracePeriod=30 Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.399101 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-2" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="object-server" containerID="cri-o://257741f8218f29b0b1e5e8d5ef95ecf3f28bee16cdc292d77de8f3326bbe123d" gracePeriod=30 Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.399130 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-2" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="container-updater" containerID="cri-o://0f1f14f4fa5711a1d3b92eef66fde7b260fbee7a7bbeed91f897f960ef2e7c3e" gracePeriod=30 Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.399158 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-2" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="container-auditor" containerID="cri-o://185a3b1b195164b0391a65fb58c65f24f848fd66987f773e5d4f2031d582ad76" gracePeriod=30 Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.399187 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-2" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="container-replicator" containerID="cri-o://8a933cae33841023d45b5c9e60e92f252afc264f4c9d875e212fd937dfe8a709" gracePeriod=30 Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.399217 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-2" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="container-server" containerID="cri-o://b1ab2126394629d1a8dbbaa0f7f865780384c107928dc42b81568566f59b1cdd" gracePeriod=30 Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.399282 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-2" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="account-reaper" containerID="cri-o://466fe0dc2e8ddc693dc18033ff08abb79536de9d9c21b6015040b06e7454b177" gracePeriod=30 Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.399316 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-2" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="account-auditor" containerID="cri-o://386892a8a101df1e91f2ce3827d6706952226a956fdbb507d91575d4831f42b3" gracePeriod=30 Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.399344 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-2" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="account-replicator" containerID="cri-o://a36e51a00bdb5a305ca9e42e0cf8e5db2c0802a50081b0774e6d53712e2848da" gracePeriod=30 Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.416286 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-storage-0"] Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.416807 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="account-server" containerID="cri-o://3e06c0aeda3fbc189f9882999ab5a9183aa33314c68104b0f48b5857654c2239" gracePeriod=30 Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.419130 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="container-updater" containerID="cri-o://1d86529c5cd723c960db81582958aec90ea60c5ae034f987b60f532622638e91" gracePeriod=30 Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.419306 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="swift-recon-cron" containerID="cri-o://95bb5217f844e378e57512480c7f0c59e7b85d4265619fb46dcdc132dc57a739" gracePeriod=30 Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.419367 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="rsync" containerID="cri-o://6efa5c1c20f09b9e71b1f4e99886eb7db8e5d70a8d132abf7a9908a54e7dbbea" gracePeriod=30 Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.419429 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="object-expirer" containerID="cri-o://26f55f0f6682744f5ed31577037541f7b99894bdd3dd19a114e1590e2b8697ab" gracePeriod=30 Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.419474 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="object-updater" containerID="cri-o://7b896439aae01ff52fa43a69d231483bddd5f7624f25904fc8ae68de8c2dc3cc" gracePeriod=30 Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.419517 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="object-auditor" containerID="cri-o://b5daa7f5307510498acd32d81ac9f7c0aab7c6308c53dc697686fd07d1776777" gracePeriod=30 Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.419559 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="object-replicator" containerID="cri-o://8a21f246526c030a1c8871a1c9d61670fe1dc6f835254920aaa17f9ba7229d99" gracePeriod=30 Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.419621 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="object-server" containerID="cri-o://d93dddf647b1cf741991cbef8e4bb89720ea9c094dfde025aa45574b1b46d1b3" gracePeriod=30 Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.419693 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="account-reaper" containerID="cri-o://68e7c594bb36cc709f03800c5db0a2d96729018353fac6d60e8cb7a4ba2fa238" gracePeriod=30 Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.419746 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="container-auditor" containerID="cri-o://a8781652607147e67c9640dde21b8ab12919f2616b8841b8eb11c1a3990f61ff" gracePeriod=30 Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.419794 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="container-replicator" containerID="cri-o://6cd56e8740db694809a028cca63c9a4433743dfb0ebe9e04a1967f2b37e9d293" gracePeriod=30 Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.419836 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="container-server" containerID="cri-o://aa111fa2f0651801d3a0703bbef1b6650d2d7b0c5e863ca4573074b7061bd9e1" gracePeriod=30 Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.419904 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="account-replicator" containerID="cri-o://e6a97829dc442d8db3046d52420793bbe248c6e4dacb29a69942321938a60bfa" gracePeriod=30 Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.419967 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="account-auditor" containerID="cri-o://6f41a0b66118e9f7bdbd7717f791fbb3f6762f06eeb7e0aff3405d58daa4958a" gracePeriod=30 Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.426883 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-storage-1"] Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.427323 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-1" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="account-server" containerID="cri-o://e346ec17aa2d2cdfc931a3fd9158238065a5520b87ac1b27e152b7596b21a812" gracePeriod=30 Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.427343 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-1" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="object-updater" containerID="cri-o://1676e9d2423174fdbd7bb0a1ca19bd1ce772561c44dd6229f48cf4244b017fd8" gracePeriod=30 Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.427427 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-1" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="object-auditor" containerID="cri-o://8d4253973cf92b8f134d333111c459dd45a28b58115e11f17fba582a8fce9cfd" gracePeriod=30 Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.427459 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-1" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="swift-recon-cron" containerID="cri-o://a89ba23a18bbf856e369b9997db5b15c2e073e42db3f087a7b8b5524d27f75c6" gracePeriod=30 Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.427475 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-1" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="object-server" containerID="cri-o://6f3ce1ef2dba1d61c13484a1020aa353732fd7e4b8aca4d62f081d429dd9cf7f" gracePeriod=30 Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.427461 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-1" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="object-replicator" containerID="cri-o://635ececc7dce091575d667fc55ff70dae5854a71fa610c71180581ff05c44e85" gracePeriod=30 Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.427507 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-1" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="container-updater" containerID="cri-o://4115dfe7381bf3fa3917d022ff124a7aa33326a18d5838b3caf284409aed99a2" gracePeriod=30 Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.427531 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-1" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="rsync" containerID="cri-o://968bf9296a3299fbf29b116fdf1d701bb11bdad0ce42eb97dd829de219f7c9d3" gracePeriod=30 Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.427539 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-1" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="container-auditor" containerID="cri-o://e651e02c4c021413625c666e0f2aae32c58a1f777868324172ec3a75e9fbdd1d" gracePeriod=30 Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.427568 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-1" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="object-expirer" containerID="cri-o://1258864ced31fd4d6f607ce7df12ec446e74a8d5bc0a46cf2122b7b2d5b1972d" gracePeriod=30 Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.427570 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-1" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="container-replicator" containerID="cri-o://f325a942655495f7cfc8dda622634e063c67e9ae302d7bc629c67eafa8aa4902" gracePeriod=30 Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.427600 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-1" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="container-server" containerID="cri-o://c89ce52dcf78c4288953c7557fe606114f20cd3a3351e3c89c2537818ce3be10" gracePeriod=30 Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.427665 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-1" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="account-reaper" containerID="cri-o://fde39a16da2d051fe287b573e0e4cdba6b70fea2e8f7445ef402a9a97f173ca7" gracePeriod=30 Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.427706 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-1" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="account-auditor" containerID="cri-o://1e42ab3da5309b6ad0251fd6b7b50713d2bf82db8ebe73f72da280111db9a63a" gracePeriod=30 Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.427720 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-1" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="account-replicator" containerID="cri-o://c7d049e1848ff61a6aa9a2cd8b145da1ddf8e0d5675e4ddaf1b547116fa455b2" gracePeriod=30 Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.441419 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hk686\" (UniqueName: \"kubernetes.io/projected/6621055b-9adc-4057-bc14-63b4848edbca-kube-api-access-hk686\") pod \"auto-csr-approver-29548258-spxr6\" (UID: \"6621055b-9adc-4057-bc14-63b4848edbca\") " pod="openshift-infra/auto-csr-approver-29548258-spxr6" Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.449681 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-xds8n"] Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.460814 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-xds8n"] Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.477399 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-proxy-76c998454c-9tgn8"] Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.477614 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-proxy-76c998454c-9tgn8" podUID="81fd2fdf-28bf-40a5-92a9-8506339ed373" containerName="proxy-httpd" containerID="cri-o://97c1aa81636e58fd90feba198be51760eed9720497251b34d30f4c9141ea8409" gracePeriod=30 Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.477977 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-proxy-76c998454c-9tgn8" podUID="81fd2fdf-28bf-40a5-92a9-8506339ed373" containerName="proxy-server" containerID="cri-o://d9db52dc67a7e088584da372ecf3e67c4da6b04503b49db9620b4cff5327c002" gracePeriod=30 Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.481657 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548258-spxr6" Mar 07 14:58:00 crc kubenswrapper[4943]: I0307 14:58:00.765395 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6eaa7147-1483-4399-ae55-b0f27c44a0e6" path="/var/lib/kubelet/pods/6eaa7147-1483-4399-ae55-b0f27c44a0e6/volumes" Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.271991 4943 generic.go:334] "Generic (PLEG): container finished" podID="db16093b-eede-415e-8145-56e579a8782a" containerID="968bf9296a3299fbf29b116fdf1d701bb11bdad0ce42eb97dd829de219f7c9d3" exitCode=0 Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.272345 4943 generic.go:334] "Generic (PLEG): container finished" podID="db16093b-eede-415e-8145-56e579a8782a" containerID="1258864ced31fd4d6f607ce7df12ec446e74a8d5bc0a46cf2122b7b2d5b1972d" exitCode=0 Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.272356 4943 generic.go:334] "Generic (PLEG): container finished" podID="db16093b-eede-415e-8145-56e579a8782a" containerID="1676e9d2423174fdbd7bb0a1ca19bd1ce772561c44dd6229f48cf4244b017fd8" exitCode=0 Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.272363 4943 generic.go:334] "Generic (PLEG): container finished" podID="db16093b-eede-415e-8145-56e579a8782a" containerID="8d4253973cf92b8f134d333111c459dd45a28b58115e11f17fba582a8fce9cfd" exitCode=0 Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.272370 4943 generic.go:334] "Generic (PLEG): container finished" podID="db16093b-eede-415e-8145-56e579a8782a" containerID="635ececc7dce091575d667fc55ff70dae5854a71fa610c71180581ff05c44e85" exitCode=0 Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.272375 4943 generic.go:334] "Generic (PLEG): container finished" podID="db16093b-eede-415e-8145-56e579a8782a" containerID="6f3ce1ef2dba1d61c13484a1020aa353732fd7e4b8aca4d62f081d429dd9cf7f" exitCode=0 Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.272384 4943 generic.go:334] "Generic (PLEG): container finished" podID="db16093b-eede-415e-8145-56e579a8782a" containerID="4115dfe7381bf3fa3917d022ff124a7aa33326a18d5838b3caf284409aed99a2" exitCode=0 Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.272391 4943 generic.go:334] "Generic (PLEG): container finished" podID="db16093b-eede-415e-8145-56e579a8782a" containerID="e651e02c4c021413625c666e0f2aae32c58a1f777868324172ec3a75e9fbdd1d" exitCode=0 Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.272397 4943 generic.go:334] "Generic (PLEG): container finished" podID="db16093b-eede-415e-8145-56e579a8782a" containerID="f325a942655495f7cfc8dda622634e063c67e9ae302d7bc629c67eafa8aa4902" exitCode=0 Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.272404 4943 generic.go:334] "Generic (PLEG): container finished" podID="db16093b-eede-415e-8145-56e579a8782a" containerID="c89ce52dcf78c4288953c7557fe606114f20cd3a3351e3c89c2537818ce3be10" exitCode=0 Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.272412 4943 generic.go:334] "Generic (PLEG): container finished" podID="db16093b-eede-415e-8145-56e579a8782a" containerID="fde39a16da2d051fe287b573e0e4cdba6b70fea2e8f7445ef402a9a97f173ca7" exitCode=0 Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.272419 4943 generic.go:334] "Generic (PLEG): container finished" podID="db16093b-eede-415e-8145-56e579a8782a" containerID="1e42ab3da5309b6ad0251fd6b7b50713d2bf82db8ebe73f72da280111db9a63a" exitCode=0 Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.272425 4943 generic.go:334] "Generic (PLEG): container finished" podID="db16093b-eede-415e-8145-56e579a8782a" containerID="c7d049e1848ff61a6aa9a2cd8b145da1ddf8e0d5675e4ddaf1b547116fa455b2" exitCode=0 Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.272432 4943 generic.go:334] "Generic (PLEG): container finished" podID="db16093b-eede-415e-8145-56e579a8782a" containerID="e346ec17aa2d2cdfc931a3fd9158238065a5520b87ac1b27e152b7596b21a812" exitCode=0 Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.272054 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"db16093b-eede-415e-8145-56e579a8782a","Type":"ContainerDied","Data":"968bf9296a3299fbf29b116fdf1d701bb11bdad0ce42eb97dd829de219f7c9d3"} Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.272493 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"db16093b-eede-415e-8145-56e579a8782a","Type":"ContainerDied","Data":"1258864ced31fd4d6f607ce7df12ec446e74a8d5bc0a46cf2122b7b2d5b1972d"} Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.272519 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"db16093b-eede-415e-8145-56e579a8782a","Type":"ContainerDied","Data":"1676e9d2423174fdbd7bb0a1ca19bd1ce772561c44dd6229f48cf4244b017fd8"} Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.272530 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"db16093b-eede-415e-8145-56e579a8782a","Type":"ContainerDied","Data":"8d4253973cf92b8f134d333111c459dd45a28b58115e11f17fba582a8fce9cfd"} Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.272539 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"db16093b-eede-415e-8145-56e579a8782a","Type":"ContainerDied","Data":"635ececc7dce091575d667fc55ff70dae5854a71fa610c71180581ff05c44e85"} Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.272549 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"db16093b-eede-415e-8145-56e579a8782a","Type":"ContainerDied","Data":"6f3ce1ef2dba1d61c13484a1020aa353732fd7e4b8aca4d62f081d429dd9cf7f"} Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.272558 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"db16093b-eede-415e-8145-56e579a8782a","Type":"ContainerDied","Data":"4115dfe7381bf3fa3917d022ff124a7aa33326a18d5838b3caf284409aed99a2"} Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.272566 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"db16093b-eede-415e-8145-56e579a8782a","Type":"ContainerDied","Data":"e651e02c4c021413625c666e0f2aae32c58a1f777868324172ec3a75e9fbdd1d"} Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.272573 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"db16093b-eede-415e-8145-56e579a8782a","Type":"ContainerDied","Data":"f325a942655495f7cfc8dda622634e063c67e9ae302d7bc629c67eafa8aa4902"} Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.272581 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"db16093b-eede-415e-8145-56e579a8782a","Type":"ContainerDied","Data":"c89ce52dcf78c4288953c7557fe606114f20cd3a3351e3c89c2537818ce3be10"} Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.272589 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"db16093b-eede-415e-8145-56e579a8782a","Type":"ContainerDied","Data":"fde39a16da2d051fe287b573e0e4cdba6b70fea2e8f7445ef402a9a97f173ca7"} Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.272605 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"db16093b-eede-415e-8145-56e579a8782a","Type":"ContainerDied","Data":"1e42ab3da5309b6ad0251fd6b7b50713d2bf82db8ebe73f72da280111db9a63a"} Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.272613 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"db16093b-eede-415e-8145-56e579a8782a","Type":"ContainerDied","Data":"c7d049e1848ff61a6aa9a2cd8b145da1ddf8e0d5675e4ddaf1b547116fa455b2"} Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.272623 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"db16093b-eede-415e-8145-56e579a8782a","Type":"ContainerDied","Data":"e346ec17aa2d2cdfc931a3fd9158238065a5520b87ac1b27e152b7596b21a812"} Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.284418 4943 generic.go:334] "Generic (PLEG): container finished" podID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerID="6efa5c1c20f09b9e71b1f4e99886eb7db8e5d70a8d132abf7a9908a54e7dbbea" exitCode=0 Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.284450 4943 generic.go:334] "Generic (PLEG): container finished" podID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerID="26f55f0f6682744f5ed31577037541f7b99894bdd3dd19a114e1590e2b8697ab" exitCode=0 Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.284464 4943 generic.go:334] "Generic (PLEG): container finished" podID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerID="7b896439aae01ff52fa43a69d231483bddd5f7624f25904fc8ae68de8c2dc3cc" exitCode=0 Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.284476 4943 generic.go:334] "Generic (PLEG): container finished" podID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerID="b5daa7f5307510498acd32d81ac9f7c0aab7c6308c53dc697686fd07d1776777" exitCode=0 Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.284485 4943 generic.go:334] "Generic (PLEG): container finished" podID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerID="8a21f246526c030a1c8871a1c9d61670fe1dc6f835254920aaa17f9ba7229d99" exitCode=0 Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.284494 4943 generic.go:334] "Generic (PLEG): container finished" podID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerID="d93dddf647b1cf741991cbef8e4bb89720ea9c094dfde025aa45574b1b46d1b3" exitCode=0 Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.284502 4943 generic.go:334] "Generic (PLEG): container finished" podID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerID="1d86529c5cd723c960db81582958aec90ea60c5ae034f987b60f532622638e91" exitCode=0 Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.284511 4943 generic.go:334] "Generic (PLEG): container finished" podID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerID="a8781652607147e67c9640dde21b8ab12919f2616b8841b8eb11c1a3990f61ff" exitCode=0 Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.284506 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"61bfc108-bcc1-4712-b8c8-8bc58b22777a","Type":"ContainerDied","Data":"6efa5c1c20f09b9e71b1f4e99886eb7db8e5d70a8d132abf7a9908a54e7dbbea"} Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.284562 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"61bfc108-bcc1-4712-b8c8-8bc58b22777a","Type":"ContainerDied","Data":"26f55f0f6682744f5ed31577037541f7b99894bdd3dd19a114e1590e2b8697ab"} Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.284579 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"61bfc108-bcc1-4712-b8c8-8bc58b22777a","Type":"ContainerDied","Data":"7b896439aae01ff52fa43a69d231483bddd5f7624f25904fc8ae68de8c2dc3cc"} Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.284521 4943 generic.go:334] "Generic (PLEG): container finished" podID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerID="6cd56e8740db694809a028cca63c9a4433743dfb0ebe9e04a1967f2b37e9d293" exitCode=0 Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.284592 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"61bfc108-bcc1-4712-b8c8-8bc58b22777a","Type":"ContainerDied","Data":"b5daa7f5307510498acd32d81ac9f7c0aab7c6308c53dc697686fd07d1776777"} Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.284599 4943 generic.go:334] "Generic (PLEG): container finished" podID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerID="aa111fa2f0651801d3a0703bbef1b6650d2d7b0c5e863ca4573074b7061bd9e1" exitCode=0 Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.284605 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"61bfc108-bcc1-4712-b8c8-8bc58b22777a","Type":"ContainerDied","Data":"8a21f246526c030a1c8871a1c9d61670fe1dc6f835254920aaa17f9ba7229d99"} Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.284610 4943 generic.go:334] "Generic (PLEG): container finished" podID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerID="68e7c594bb36cc709f03800c5db0a2d96729018353fac6d60e8cb7a4ba2fa238" exitCode=0 Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.284620 4943 generic.go:334] "Generic (PLEG): container finished" podID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerID="6f41a0b66118e9f7bdbd7717f791fbb3f6762f06eeb7e0aff3405d58daa4958a" exitCode=0 Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.284621 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"61bfc108-bcc1-4712-b8c8-8bc58b22777a","Type":"ContainerDied","Data":"d93dddf647b1cf741991cbef8e4bb89720ea9c094dfde025aa45574b1b46d1b3"} Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.284629 4943 generic.go:334] "Generic (PLEG): container finished" podID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerID="e6a97829dc442d8db3046d52420793bbe248c6e4dacb29a69942321938a60bfa" exitCode=0 Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.284637 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"61bfc108-bcc1-4712-b8c8-8bc58b22777a","Type":"ContainerDied","Data":"1d86529c5cd723c960db81582958aec90ea60c5ae034f987b60f532622638e91"} Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.284641 4943 generic.go:334] "Generic (PLEG): container finished" podID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerID="3e06c0aeda3fbc189f9882999ab5a9183aa33314c68104b0f48b5857654c2239" exitCode=0 Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.284654 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"61bfc108-bcc1-4712-b8c8-8bc58b22777a","Type":"ContainerDied","Data":"a8781652607147e67c9640dde21b8ab12919f2616b8841b8eb11c1a3990f61ff"} Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.284667 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"61bfc108-bcc1-4712-b8c8-8bc58b22777a","Type":"ContainerDied","Data":"6cd56e8740db694809a028cca63c9a4433743dfb0ebe9e04a1967f2b37e9d293"} Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.284679 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"61bfc108-bcc1-4712-b8c8-8bc58b22777a","Type":"ContainerDied","Data":"aa111fa2f0651801d3a0703bbef1b6650d2d7b0c5e863ca4573074b7061bd9e1"} Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.284689 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"61bfc108-bcc1-4712-b8c8-8bc58b22777a","Type":"ContainerDied","Data":"68e7c594bb36cc709f03800c5db0a2d96729018353fac6d60e8cb7a4ba2fa238"} Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.284700 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"61bfc108-bcc1-4712-b8c8-8bc58b22777a","Type":"ContainerDied","Data":"6f41a0b66118e9f7bdbd7717f791fbb3f6762f06eeb7e0aff3405d58daa4958a"} Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.284710 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"61bfc108-bcc1-4712-b8c8-8bc58b22777a","Type":"ContainerDied","Data":"e6a97829dc442d8db3046d52420793bbe248c6e4dacb29a69942321938a60bfa"} Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.284721 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"61bfc108-bcc1-4712-b8c8-8bc58b22777a","Type":"ContainerDied","Data":"3e06c0aeda3fbc189f9882999ab5a9183aa33314c68104b0f48b5857654c2239"} Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.290360 4943 generic.go:334] "Generic (PLEG): container finished" podID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerID="737a49457d09cd8479c49fa391100fe4f3b79bf41392edda0bf94b505382ab04" exitCode=0 Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.290380 4943 generic.go:334] "Generic (PLEG): container finished" podID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerID="2ee77dbb77dbb65d1523838e6afd9d5a7b72613d080864ff79215b4a508ae442" exitCode=0 Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.290389 4943 generic.go:334] "Generic (PLEG): container finished" podID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerID="ba8857ea0c02e4a8248cffb1f9d8e7c2b66181c038e948a8f69e9f91e89979a7" exitCode=0 Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.290397 4943 generic.go:334] "Generic (PLEG): container finished" podID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerID="7b3105463e57fbdbaa2e126058c6aaa054286f515ac013e5b967046d33b29603" exitCode=0 Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.290406 4943 generic.go:334] "Generic (PLEG): container finished" podID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerID="e571f26eb8b825f06454b7946c558c7a1eb1f69258f0d8ada3b654c53c97b55b" exitCode=0 Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.290414 4943 generic.go:334] "Generic (PLEG): container finished" podID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerID="257741f8218f29b0b1e5e8d5ef95ecf3f28bee16cdc292d77de8f3326bbe123d" exitCode=0 Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.290425 4943 generic.go:334] "Generic (PLEG): container finished" podID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerID="0f1f14f4fa5711a1d3b92eef66fde7b260fbee7a7bbeed91f897f960ef2e7c3e" exitCode=0 Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.290433 4943 generic.go:334] "Generic (PLEG): container finished" podID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerID="185a3b1b195164b0391a65fb58c65f24f848fd66987f773e5d4f2031d582ad76" exitCode=0 Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.290442 4943 generic.go:334] "Generic (PLEG): container finished" podID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerID="8a933cae33841023d45b5c9e60e92f252afc264f4c9d875e212fd937dfe8a709" exitCode=0 Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.290453 4943 generic.go:334] "Generic (PLEG): container finished" podID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerID="b1ab2126394629d1a8dbbaa0f7f865780384c107928dc42b81568566f59b1cdd" exitCode=0 Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.290462 4943 generic.go:334] "Generic (PLEG): container finished" podID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerID="466fe0dc2e8ddc693dc18033ff08abb79536de9d9c21b6015040b06e7454b177" exitCode=0 Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.290469 4943 generic.go:334] "Generic (PLEG): container finished" podID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerID="386892a8a101df1e91f2ce3827d6706952226a956fdbb507d91575d4831f42b3" exitCode=0 Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.290476 4943 generic.go:334] "Generic (PLEG): container finished" podID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerID="a36e51a00bdb5a305ca9e42e0cf8e5db2c0802a50081b0774e6d53712e2848da" exitCode=0 Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.290484 4943 generic.go:334] "Generic (PLEG): container finished" podID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerID="378e9e0fc8363d06ebee218138d567cd679e69eab4d30a8c35d8d6ba0c009165" exitCode=0 Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.290434 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"0e83c041-7178-43e4-b05d-e1bcc0ec36bd","Type":"ContainerDied","Data":"737a49457d09cd8479c49fa391100fe4f3b79bf41392edda0bf94b505382ab04"} Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.290541 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"0e83c041-7178-43e4-b05d-e1bcc0ec36bd","Type":"ContainerDied","Data":"2ee77dbb77dbb65d1523838e6afd9d5a7b72613d080864ff79215b4a508ae442"} Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.290553 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"0e83c041-7178-43e4-b05d-e1bcc0ec36bd","Type":"ContainerDied","Data":"ba8857ea0c02e4a8248cffb1f9d8e7c2b66181c038e948a8f69e9f91e89979a7"} Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.290562 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"0e83c041-7178-43e4-b05d-e1bcc0ec36bd","Type":"ContainerDied","Data":"7b3105463e57fbdbaa2e126058c6aaa054286f515ac013e5b967046d33b29603"} Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.290571 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"0e83c041-7178-43e4-b05d-e1bcc0ec36bd","Type":"ContainerDied","Data":"e571f26eb8b825f06454b7946c558c7a1eb1f69258f0d8ada3b654c53c97b55b"} Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.290581 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"0e83c041-7178-43e4-b05d-e1bcc0ec36bd","Type":"ContainerDied","Data":"257741f8218f29b0b1e5e8d5ef95ecf3f28bee16cdc292d77de8f3326bbe123d"} Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.290591 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"0e83c041-7178-43e4-b05d-e1bcc0ec36bd","Type":"ContainerDied","Data":"0f1f14f4fa5711a1d3b92eef66fde7b260fbee7a7bbeed91f897f960ef2e7c3e"} Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.290600 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"0e83c041-7178-43e4-b05d-e1bcc0ec36bd","Type":"ContainerDied","Data":"185a3b1b195164b0391a65fb58c65f24f848fd66987f773e5d4f2031d582ad76"} Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.290609 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"0e83c041-7178-43e4-b05d-e1bcc0ec36bd","Type":"ContainerDied","Data":"8a933cae33841023d45b5c9e60e92f252afc264f4c9d875e212fd937dfe8a709"} Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.290619 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"0e83c041-7178-43e4-b05d-e1bcc0ec36bd","Type":"ContainerDied","Data":"b1ab2126394629d1a8dbbaa0f7f865780384c107928dc42b81568566f59b1cdd"} Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.290628 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"0e83c041-7178-43e4-b05d-e1bcc0ec36bd","Type":"ContainerDied","Data":"466fe0dc2e8ddc693dc18033ff08abb79536de9d9c21b6015040b06e7454b177"} Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.290637 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"0e83c041-7178-43e4-b05d-e1bcc0ec36bd","Type":"ContainerDied","Data":"386892a8a101df1e91f2ce3827d6706952226a956fdbb507d91575d4831f42b3"} Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.290646 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"0e83c041-7178-43e4-b05d-e1bcc0ec36bd","Type":"ContainerDied","Data":"a36e51a00bdb5a305ca9e42e0cf8e5db2c0802a50081b0774e6d53712e2848da"} Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.290655 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"0e83c041-7178-43e4-b05d-e1bcc0ec36bd","Type":"ContainerDied","Data":"378e9e0fc8363d06ebee218138d567cd679e69eab4d30a8c35d8d6ba0c009165"} Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.293058 4943 generic.go:334] "Generic (PLEG): container finished" podID="81fd2fdf-28bf-40a5-92a9-8506339ed373" containerID="d9db52dc67a7e088584da372ecf3e67c4da6b04503b49db9620b4cff5327c002" exitCode=0 Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.293075 4943 generic.go:334] "Generic (PLEG): container finished" podID="81fd2fdf-28bf-40a5-92a9-8506339ed373" containerID="97c1aa81636e58fd90feba198be51760eed9720497251b34d30f4c9141ea8409" exitCode=0 Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.293142 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-proxy-76c998454c-9tgn8" event={"ID":"81fd2fdf-28bf-40a5-92a9-8506339ed373","Type":"ContainerDied","Data":"d9db52dc67a7e088584da372ecf3e67c4da6b04503b49db9620b4cff5327c002"} Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.293187 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-proxy-76c998454c-9tgn8" event={"ID":"81fd2fdf-28bf-40a5-92a9-8506339ed373","Type":"ContainerDied","Data":"97c1aa81636e58fd90feba198be51760eed9720497251b34d30f4c9141ea8409"} Mar 07 14:58:01 crc kubenswrapper[4943]: I0307 14:58:01.962029 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29548258-spxr6"] Mar 07 14:58:01 crc kubenswrapper[4943]: W0307 14:58:01.973778 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6621055b_9adc_4057_bc14_63b4848edbca.slice/crio-78a59550006a5068c77593c75bcc23a65b1a7f906225244e29ea38ee4d084df1 WatchSource:0}: Error finding container 78a59550006a5068c77593c75bcc23a65b1a7f906225244e29ea38ee4d084df1: Status 404 returned error can't find the container with id 78a59550006a5068c77593c75bcc23a65b1a7f906225244e29ea38ee4d084df1 Mar 07 14:58:02 crc kubenswrapper[4943]: I0307 14:58:02.130620 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-dxspq" Mar 07 14:58:02 crc kubenswrapper[4943]: I0307 14:58:02.139034 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-proxy-76c998454c-9tgn8" Mar 07 14:58:02 crc kubenswrapper[4943]: I0307 14:58:02.307679 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2907f386-4042-4031-bbb1-327ac38b4173-dispersionconf\") pod \"2907f386-4042-4031-bbb1-327ac38b4173\" (UID: \"2907f386-4042-4031-bbb1-327ac38b4173\") " Mar 07 14:58:02 crc kubenswrapper[4943]: I0307 14:58:02.307785 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2907f386-4042-4031-bbb1-327ac38b4173-etc-swift\") pod \"2907f386-4042-4031-bbb1-327ac38b4173\" (UID: \"2907f386-4042-4031-bbb1-327ac38b4173\") " Mar 07 14:58:02 crc kubenswrapper[4943]: I0307 14:58:02.307826 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/81fd2fdf-28bf-40a5-92a9-8506339ed373-etc-swift\") pod \"81fd2fdf-28bf-40a5-92a9-8506339ed373\" (UID: \"81fd2fdf-28bf-40a5-92a9-8506339ed373\") " Mar 07 14:58:02 crc kubenswrapper[4943]: I0307 14:58:02.307870 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2907f386-4042-4031-bbb1-327ac38b4173-scripts\") pod \"2907f386-4042-4031-bbb1-327ac38b4173\" (UID: \"2907f386-4042-4031-bbb1-327ac38b4173\") " Mar 07 14:58:02 crc kubenswrapper[4943]: I0307 14:58:02.307914 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/81fd2fdf-28bf-40a5-92a9-8506339ed373-log-httpd\") pod \"81fd2fdf-28bf-40a5-92a9-8506339ed373\" (UID: \"81fd2fdf-28bf-40a5-92a9-8506339ed373\") " Mar 07 14:58:02 crc kubenswrapper[4943]: I0307 14:58:02.307967 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vfnhz\" (UniqueName: \"kubernetes.io/projected/2907f386-4042-4031-bbb1-327ac38b4173-kube-api-access-vfnhz\") pod \"2907f386-4042-4031-bbb1-327ac38b4173\" (UID: \"2907f386-4042-4031-bbb1-327ac38b4173\") " Mar 07 14:58:02 crc kubenswrapper[4943]: I0307 14:58:02.308042 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/81fd2fdf-28bf-40a5-92a9-8506339ed373-run-httpd\") pod \"81fd2fdf-28bf-40a5-92a9-8506339ed373\" (UID: \"81fd2fdf-28bf-40a5-92a9-8506339ed373\") " Mar 07 14:58:02 crc kubenswrapper[4943]: I0307 14:58:02.308107 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81fd2fdf-28bf-40a5-92a9-8506339ed373-config-data\") pod \"81fd2fdf-28bf-40a5-92a9-8506339ed373\" (UID: \"81fd2fdf-28bf-40a5-92a9-8506339ed373\") " Mar 07 14:58:02 crc kubenswrapper[4943]: I0307 14:58:02.308144 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ms495\" (UniqueName: \"kubernetes.io/projected/81fd2fdf-28bf-40a5-92a9-8506339ed373-kube-api-access-ms495\") pod \"81fd2fdf-28bf-40a5-92a9-8506339ed373\" (UID: \"81fd2fdf-28bf-40a5-92a9-8506339ed373\") " Mar 07 14:58:02 crc kubenswrapper[4943]: I0307 14:58:02.308221 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2907f386-4042-4031-bbb1-327ac38b4173-ring-data-devices\") pod \"2907f386-4042-4031-bbb1-327ac38b4173\" (UID: \"2907f386-4042-4031-bbb1-327ac38b4173\") " Mar 07 14:58:02 crc kubenswrapper[4943]: I0307 14:58:02.308252 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2907f386-4042-4031-bbb1-327ac38b4173-swiftconf\") pod \"2907f386-4042-4031-bbb1-327ac38b4173\" (UID: \"2907f386-4042-4031-bbb1-327ac38b4173\") " Mar 07 14:58:02 crc kubenswrapper[4943]: I0307 14:58:02.308597 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/81fd2fdf-28bf-40a5-92a9-8506339ed373-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "81fd2fdf-28bf-40a5-92a9-8506339ed373" (UID: "81fd2fdf-28bf-40a5-92a9-8506339ed373"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:58:02 crc kubenswrapper[4943]: I0307 14:58:02.309033 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2907f386-4042-4031-bbb1-327ac38b4173-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "2907f386-4042-4031-bbb1-327ac38b4173" (UID: "2907f386-4042-4031-bbb1-327ac38b4173"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:58:02 crc kubenswrapper[4943]: I0307 14:58:02.309530 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2907f386-4042-4031-bbb1-327ac38b4173-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "2907f386-4042-4031-bbb1-327ac38b4173" (UID: "2907f386-4042-4031-bbb1-327ac38b4173"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:58:02 crc kubenswrapper[4943]: I0307 14:58:02.309909 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/81fd2fdf-28bf-40a5-92a9-8506339ed373-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "81fd2fdf-28bf-40a5-92a9-8506339ed373" (UID: "81fd2fdf-28bf-40a5-92a9-8506339ed373"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:58:02 crc kubenswrapper[4943]: I0307 14:58:02.314301 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2907f386-4042-4031-bbb1-327ac38b4173-kube-api-access-vfnhz" (OuterVolumeSpecName: "kube-api-access-vfnhz") pod "2907f386-4042-4031-bbb1-327ac38b4173" (UID: "2907f386-4042-4031-bbb1-327ac38b4173"). InnerVolumeSpecName "kube-api-access-vfnhz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:58:02 crc kubenswrapper[4943]: I0307 14:58:02.314549 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81fd2fdf-28bf-40a5-92a9-8506339ed373-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "81fd2fdf-28bf-40a5-92a9-8506339ed373" (UID: "81fd2fdf-28bf-40a5-92a9-8506339ed373"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:58:02 crc kubenswrapper[4943]: I0307 14:58:02.315076 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81fd2fdf-28bf-40a5-92a9-8506339ed373-kube-api-access-ms495" (OuterVolumeSpecName: "kube-api-access-ms495") pod "81fd2fdf-28bf-40a5-92a9-8506339ed373" (UID: "81fd2fdf-28bf-40a5-92a9-8506339ed373"). InnerVolumeSpecName "kube-api-access-ms495". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:58:02 crc kubenswrapper[4943]: I0307 14:58:02.323356 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-proxy-76c998454c-9tgn8" event={"ID":"81fd2fdf-28bf-40a5-92a9-8506339ed373","Type":"ContainerDied","Data":"a01bfe2f9f4f71fa07d531b560b7aa1faea31b5eca72a5c024fad6a7ea4f78e4"} Mar 07 14:58:02 crc kubenswrapper[4943]: I0307 14:58:02.323581 4943 scope.go:117] "RemoveContainer" containerID="d9db52dc67a7e088584da372ecf3e67c4da6b04503b49db9620b4cff5327c002" Mar 07 14:58:02 crc kubenswrapper[4943]: I0307 14:58:02.323819 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-proxy-76c998454c-9tgn8" Mar 07 14:58:02 crc kubenswrapper[4943]: I0307 14:58:02.326227 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-dxspq" Mar 07 14:58:02 crc kubenswrapper[4943]: I0307 14:58:02.330659 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548258-spxr6" event={"ID":"6621055b-9adc-4057-bc14-63b4848edbca","Type":"ContainerStarted","Data":"78a59550006a5068c77593c75bcc23a65b1a7f906225244e29ea38ee4d084df1"} Mar 07 14:58:02 crc kubenswrapper[4943]: I0307 14:58:02.333695 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2907f386-4042-4031-bbb1-327ac38b4173-scripts" (OuterVolumeSpecName: "scripts") pod "2907f386-4042-4031-bbb1-327ac38b4173" (UID: "2907f386-4042-4031-bbb1-327ac38b4173"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:58:02 crc kubenswrapper[4943]: I0307 14:58:02.336619 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2907f386-4042-4031-bbb1-327ac38b4173-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "2907f386-4042-4031-bbb1-327ac38b4173" (UID: "2907f386-4042-4031-bbb1-327ac38b4173"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:58:02 crc kubenswrapper[4943]: I0307 14:58:02.338074 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2907f386-4042-4031-bbb1-327ac38b4173-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "2907f386-4042-4031-bbb1-327ac38b4173" (UID: "2907f386-4042-4031-bbb1-327ac38b4173"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:58:02 crc kubenswrapper[4943]: I0307 14:58:02.369336 4943 scope.go:117] "RemoveContainer" containerID="97c1aa81636e58fd90feba198be51760eed9720497251b34d30f4c9141ea8409" Mar 07 14:58:02 crc kubenswrapper[4943]: I0307 14:58:02.375691 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81fd2fdf-28bf-40a5-92a9-8506339ed373-config-data" (OuterVolumeSpecName: "config-data") pod "81fd2fdf-28bf-40a5-92a9-8506339ed373" (UID: "81fd2fdf-28bf-40a5-92a9-8506339ed373"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:58:02 crc kubenswrapper[4943]: I0307 14:58:02.387782 4943 scope.go:117] "RemoveContainer" containerID="326cea0169c4db33a7d753742d878523652301a137ebd599243c6d0c184e2a3e" Mar 07 14:58:02 crc kubenswrapper[4943]: I0307 14:58:02.409718 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2907f386-4042-4031-bbb1-327ac38b4173-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 14:58:02 crc kubenswrapper[4943]: I0307 14:58:02.409744 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2907f386-4042-4031-bbb1-327ac38b4173-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 14:58:02 crc kubenswrapper[4943]: I0307 14:58:02.409757 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/81fd2fdf-28bf-40a5-92a9-8506339ed373-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 14:58:02 crc kubenswrapper[4943]: I0307 14:58:02.409768 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2907f386-4042-4031-bbb1-327ac38b4173-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 14:58:02 crc kubenswrapper[4943]: I0307 14:58:02.409781 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vfnhz\" (UniqueName: \"kubernetes.io/projected/2907f386-4042-4031-bbb1-327ac38b4173-kube-api-access-vfnhz\") on node \"crc\" DevicePath \"\"" Mar 07 14:58:02 crc kubenswrapper[4943]: I0307 14:58:02.409795 4943 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/81fd2fdf-28bf-40a5-92a9-8506339ed373-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 07 14:58:02 crc kubenswrapper[4943]: I0307 14:58:02.409806 4943 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/81fd2fdf-28bf-40a5-92a9-8506339ed373-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 07 14:58:02 crc kubenswrapper[4943]: I0307 14:58:02.409816 4943 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81fd2fdf-28bf-40a5-92a9-8506339ed373-config-data\") on node \"crc\" DevicePath \"\"" Mar 07 14:58:02 crc kubenswrapper[4943]: I0307 14:58:02.409829 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ms495\" (UniqueName: \"kubernetes.io/projected/81fd2fdf-28bf-40a5-92a9-8506339ed373-kube-api-access-ms495\") on node \"crc\" DevicePath \"\"" Mar 07 14:58:02 crc kubenswrapper[4943]: I0307 14:58:02.409841 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2907f386-4042-4031-bbb1-327ac38b4173-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 14:58:02 crc kubenswrapper[4943]: I0307 14:58:02.409851 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2907f386-4042-4031-bbb1-327ac38b4173-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 14:58:02 crc kubenswrapper[4943]: I0307 14:58:02.675509 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-proxy-76c998454c-9tgn8"] Mar 07 14:58:02 crc kubenswrapper[4943]: I0307 14:58:02.687387 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-proxy-76c998454c-9tgn8"] Mar 07 14:58:02 crc kubenswrapper[4943]: I0307 14:58:02.778896 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2907f386-4042-4031-bbb1-327ac38b4173" path="/var/lib/kubelet/pods/2907f386-4042-4031-bbb1-327ac38b4173/volumes" Mar 07 14:58:02 crc kubenswrapper[4943]: I0307 14:58:02.780687 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81fd2fdf-28bf-40a5-92a9-8506339ed373" path="/var/lib/kubelet/pods/81fd2fdf-28bf-40a5-92a9-8506339ed373/volumes" Mar 07 14:58:03 crc kubenswrapper[4943]: I0307 14:58:03.343167 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548258-spxr6" event={"ID":"6621055b-9adc-4057-bc14-63b4848edbca","Type":"ContainerStarted","Data":"e5dc11f9199090a79eeb09c228f04112c7dc7ee35b2769301fdcf56bf33efdd1"} Mar 07 14:58:04 crc kubenswrapper[4943]: I0307 14:58:04.360696 4943 generic.go:334] "Generic (PLEG): container finished" podID="6621055b-9adc-4057-bc14-63b4848edbca" containerID="e5dc11f9199090a79eeb09c228f04112c7dc7ee35b2769301fdcf56bf33efdd1" exitCode=0 Mar 07 14:58:04 crc kubenswrapper[4943]: I0307 14:58:04.360784 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548258-spxr6" event={"ID":"6621055b-9adc-4057-bc14-63b4848edbca","Type":"ContainerDied","Data":"e5dc11f9199090a79eeb09c228f04112c7dc7ee35b2769301fdcf56bf33efdd1"} Mar 07 14:58:05 crc kubenswrapper[4943]: I0307 14:58:05.702886 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548258-spxr6" Mar 07 14:58:05 crc kubenswrapper[4943]: I0307 14:58:05.760281 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hk686\" (UniqueName: \"kubernetes.io/projected/6621055b-9adc-4057-bc14-63b4848edbca-kube-api-access-hk686\") pod \"6621055b-9adc-4057-bc14-63b4848edbca\" (UID: \"6621055b-9adc-4057-bc14-63b4848edbca\") " Mar 07 14:58:05 crc kubenswrapper[4943]: I0307 14:58:05.769308 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6621055b-9adc-4057-bc14-63b4848edbca-kube-api-access-hk686" (OuterVolumeSpecName: "kube-api-access-hk686") pod "6621055b-9adc-4057-bc14-63b4848edbca" (UID: "6621055b-9adc-4057-bc14-63b4848edbca"). InnerVolumeSpecName "kube-api-access-hk686". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:58:05 crc kubenswrapper[4943]: I0307 14:58:05.862427 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hk686\" (UniqueName: \"kubernetes.io/projected/6621055b-9adc-4057-bc14-63b4848edbca-kube-api-access-hk686\") on node \"crc\" DevicePath \"\"" Mar 07 14:58:06 crc kubenswrapper[4943]: I0307 14:58:06.073908 4943 patch_prober.go:28] interesting pod/machine-config-daemon-cpdmm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 07 14:58:06 crc kubenswrapper[4943]: I0307 14:58:06.074028 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 07 14:58:06 crc kubenswrapper[4943]: I0307 14:58:06.381349 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548258-spxr6" event={"ID":"6621055b-9adc-4057-bc14-63b4848edbca","Type":"ContainerDied","Data":"78a59550006a5068c77593c75bcc23a65b1a7f906225244e29ea38ee4d084df1"} Mar 07 14:58:06 crc kubenswrapper[4943]: I0307 14:58:06.381633 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="78a59550006a5068c77593c75bcc23a65b1a7f906225244e29ea38ee4d084df1" Mar 07 14:58:06 crc kubenswrapper[4943]: I0307 14:58:06.381704 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548258-spxr6" Mar 07 14:58:06 crc kubenswrapper[4943]: I0307 14:58:06.432684 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29548252-brvbf"] Mar 07 14:58:06 crc kubenswrapper[4943]: I0307 14:58:06.444399 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29548252-brvbf"] Mar 07 14:58:06 crc kubenswrapper[4943]: I0307 14:58:06.770442 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba71982c-c6e1-4247-8cf7-28661b6c4a36" path="/var/lib/kubelet/pods/ba71982c-c6e1-4247-8cf7-28661b6c4a36/volumes" Mar 07 14:58:19 crc kubenswrapper[4943]: I0307 14:58:19.821494 4943 scope.go:117] "RemoveContainer" containerID="b554df5bf661f81ff1a5aa8e723d7f0fdf8d057fec6e88cd52bf920ec8b1f872" Mar 07 14:58:30 crc kubenswrapper[4943]: I0307 14:58:30.655083 4943 generic.go:334] "Generic (PLEG): container finished" podID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerID="95bb5217f844e378e57512480c7f0c59e7b85d4265619fb46dcdc132dc57a739" exitCode=137 Mar 07 14:58:30 crc kubenswrapper[4943]: I0307 14:58:30.655310 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"61bfc108-bcc1-4712-b8c8-8bc58b22777a","Type":"ContainerDied","Data":"95bb5217f844e378e57512480c7f0c59e7b85d4265619fb46dcdc132dc57a739"} Mar 07 14:58:30 crc kubenswrapper[4943]: E0307 14:58:30.668232 4943 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddb16093b_eede_415e_8145_56e579a8782a.slice/crio-conmon-a89ba23a18bbf856e369b9997db5b15c2e073e42db3f087a7b8b5524d27f75c6.scope\": RecentStats: unable to find data in memory cache]" Mar 07 14:58:30 crc kubenswrapper[4943]: I0307 14:58:30.669561 4943 generic.go:334] "Generic (PLEG): container finished" podID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerID="7c67a0fd742d59f47a0d1f6a924f182e1e84266f97330d060208d8f777bad883" exitCode=137 Mar 07 14:58:30 crc kubenswrapper[4943]: I0307 14:58:30.669656 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"0e83c041-7178-43e4-b05d-e1bcc0ec36bd","Type":"ContainerDied","Data":"7c67a0fd742d59f47a0d1f6a924f182e1e84266f97330d060208d8f777bad883"} Mar 07 14:58:30 crc kubenswrapper[4943]: I0307 14:58:30.679368 4943 generic.go:334] "Generic (PLEG): container finished" podID="db16093b-eede-415e-8145-56e579a8782a" containerID="a89ba23a18bbf856e369b9997db5b15c2e073e42db3f087a7b8b5524d27f75c6" exitCode=137 Mar 07 14:58:30 crc kubenswrapper[4943]: I0307 14:58:30.679848 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"db16093b-eede-415e-8145-56e579a8782a","Type":"ContainerDied","Data":"a89ba23a18bbf856e369b9997db5b15c2e073e42db3f087a7b8b5524d27f75c6"} Mar 07 14:58:30 crc kubenswrapper[4943]: I0307 14:58:30.863341 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-storage-2" Mar 07 14:58:30 crc kubenswrapper[4943]: I0307 14:58:30.904672 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-storage-0" Mar 07 14:58:30 crc kubenswrapper[4943]: I0307 14:58:30.932635 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-storage-1" Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.060706 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-llm5v\" (UniqueName: \"kubernetes.io/projected/0e83c041-7178-43e4-b05d-e1bcc0ec36bd-kube-api-access-llm5v\") pod \"0e83c041-7178-43e4-b05d-e1bcc0ec36bd\" (UID: \"0e83c041-7178-43e4-b05d-e1bcc0ec36bd\") " Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.060746 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swift\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"0e83c041-7178-43e4-b05d-e1bcc0ec36bd\" (UID: \"0e83c041-7178-43e4-b05d-e1bcc0ec36bd\") " Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.060775 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/db16093b-eede-415e-8145-56e579a8782a-etc-swift\") pod \"db16093b-eede-415e-8145-56e579a8782a\" (UID: \"db16093b-eede-415e-8145-56e579a8782a\") " Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.060798 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/0e83c041-7178-43e4-b05d-e1bcc0ec36bd-cache\") pod \"0e83c041-7178-43e4-b05d-e1bcc0ec36bd\" (UID: \"0e83c041-7178-43e4-b05d-e1bcc0ec36bd\") " Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.060830 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kksv6\" (UniqueName: \"kubernetes.io/projected/61bfc108-bcc1-4712-b8c8-8bc58b22777a-kube-api-access-kksv6\") pod \"61bfc108-bcc1-4712-b8c8-8bc58b22777a\" (UID: \"61bfc108-bcc1-4712-b8c8-8bc58b22777a\") " Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.060843 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swift\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"db16093b-eede-415e-8145-56e579a8782a\" (UID: \"db16093b-eede-415e-8145-56e579a8782a\") " Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.060858 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nknnt\" (UniqueName: \"kubernetes.io/projected/db16093b-eede-415e-8145-56e579a8782a-kube-api-access-nknnt\") pod \"db16093b-eede-415e-8145-56e579a8782a\" (UID: \"db16093b-eede-415e-8145-56e579a8782a\") " Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.060888 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/db16093b-eede-415e-8145-56e579a8782a-cache\") pod \"db16093b-eede-415e-8145-56e579a8782a\" (UID: \"db16093b-eede-415e-8145-56e579a8782a\") " Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.060909 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0e83c041-7178-43e4-b05d-e1bcc0ec36bd-etc-swift\") pod \"0e83c041-7178-43e4-b05d-e1bcc0ec36bd\" (UID: \"0e83c041-7178-43e4-b05d-e1bcc0ec36bd\") " Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.060938 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/61bfc108-bcc1-4712-b8c8-8bc58b22777a-etc-swift\") pod \"61bfc108-bcc1-4712-b8c8-8bc58b22777a\" (UID: \"61bfc108-bcc1-4712-b8c8-8bc58b22777a\") " Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.060977 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/61bfc108-bcc1-4712-b8c8-8bc58b22777a-lock\") pod \"61bfc108-bcc1-4712-b8c8-8bc58b22777a\" (UID: \"61bfc108-bcc1-4712-b8c8-8bc58b22777a\") " Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.061043 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/0e83c041-7178-43e4-b05d-e1bcc0ec36bd-lock\") pod \"0e83c041-7178-43e4-b05d-e1bcc0ec36bd\" (UID: \"0e83c041-7178-43e4-b05d-e1bcc0ec36bd\") " Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.061060 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swift\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"61bfc108-bcc1-4712-b8c8-8bc58b22777a\" (UID: \"61bfc108-bcc1-4712-b8c8-8bc58b22777a\") " Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.061094 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/61bfc108-bcc1-4712-b8c8-8bc58b22777a-cache\") pod \"61bfc108-bcc1-4712-b8c8-8bc58b22777a\" (UID: \"61bfc108-bcc1-4712-b8c8-8bc58b22777a\") " Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.061122 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/db16093b-eede-415e-8145-56e579a8782a-lock\") pod \"db16093b-eede-415e-8145-56e579a8782a\" (UID: \"db16093b-eede-415e-8145-56e579a8782a\") " Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.061523 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db16093b-eede-415e-8145-56e579a8782a-cache" (OuterVolumeSpecName: "cache") pod "db16093b-eede-415e-8145-56e579a8782a" (UID: "db16093b-eede-415e-8145-56e579a8782a"). InnerVolumeSpecName "cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.061855 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db16093b-eede-415e-8145-56e579a8782a-lock" (OuterVolumeSpecName: "lock") pod "db16093b-eede-415e-8145-56e579a8782a" (UID: "db16093b-eede-415e-8145-56e579a8782a"). InnerVolumeSpecName "lock". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.061916 4943 reconciler_common.go:293] "Volume detached for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/db16093b-eede-415e-8145-56e579a8782a-cache\") on node \"crc\" DevicePath \"\"" Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.062968 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e83c041-7178-43e4-b05d-e1bcc0ec36bd-cache" (OuterVolumeSpecName: "cache") pod "0e83c041-7178-43e4-b05d-e1bcc0ec36bd" (UID: "0e83c041-7178-43e4-b05d-e1bcc0ec36bd"). InnerVolumeSpecName "cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.063065 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e83c041-7178-43e4-b05d-e1bcc0ec36bd-lock" (OuterVolumeSpecName: "lock") pod "0e83c041-7178-43e4-b05d-e1bcc0ec36bd" (UID: "0e83c041-7178-43e4-b05d-e1bcc0ec36bd"). InnerVolumeSpecName "lock". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.063354 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61bfc108-bcc1-4712-b8c8-8bc58b22777a-lock" (OuterVolumeSpecName: "lock") pod "61bfc108-bcc1-4712-b8c8-8bc58b22777a" (UID: "61bfc108-bcc1-4712-b8c8-8bc58b22777a"). InnerVolumeSpecName "lock". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.063653 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61bfc108-bcc1-4712-b8c8-8bc58b22777a-cache" (OuterVolumeSpecName: "cache") pod "61bfc108-bcc1-4712-b8c8-8bc58b22777a" (UID: "61bfc108-bcc1-4712-b8c8-8bc58b22777a"). InnerVolumeSpecName "cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.066608 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e83c041-7178-43e4-b05d-e1bcc0ec36bd-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "0e83c041-7178-43e4-b05d-e1bcc0ec36bd" (UID: "0e83c041-7178-43e4-b05d-e1bcc0ec36bd"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.066693 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "swift") pod "61bfc108-bcc1-4712-b8c8-8bc58b22777a" (UID: "61bfc108-bcc1-4712-b8c8-8bc58b22777a"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.066915 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "swift") pod "0e83c041-7178-43e4-b05d-e1bcc0ec36bd" (UID: "0e83c041-7178-43e4-b05d-e1bcc0ec36bd"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.066979 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61bfc108-bcc1-4712-b8c8-8bc58b22777a-kube-api-access-kksv6" (OuterVolumeSpecName: "kube-api-access-kksv6") pod "61bfc108-bcc1-4712-b8c8-8bc58b22777a" (UID: "61bfc108-bcc1-4712-b8c8-8bc58b22777a"). InnerVolumeSpecName "kube-api-access-kksv6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.067097 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db16093b-eede-415e-8145-56e579a8782a-kube-api-access-nknnt" (OuterVolumeSpecName: "kube-api-access-nknnt") pod "db16093b-eede-415e-8145-56e579a8782a" (UID: "db16093b-eede-415e-8145-56e579a8782a"). InnerVolumeSpecName "kube-api-access-nknnt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.067170 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e83c041-7178-43e4-b05d-e1bcc0ec36bd-kube-api-access-llm5v" (OuterVolumeSpecName: "kube-api-access-llm5v") pod "0e83c041-7178-43e4-b05d-e1bcc0ec36bd" (UID: "0e83c041-7178-43e4-b05d-e1bcc0ec36bd"). InnerVolumeSpecName "kube-api-access-llm5v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.067553 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "swift") pod "db16093b-eede-415e-8145-56e579a8782a" (UID: "db16093b-eede-415e-8145-56e579a8782a"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.068065 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db16093b-eede-415e-8145-56e579a8782a-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "db16093b-eede-415e-8145-56e579a8782a" (UID: "db16093b-eede-415e-8145-56e579a8782a"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.069080 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61bfc108-bcc1-4712-b8c8-8bc58b22777a-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "61bfc108-bcc1-4712-b8c8-8bc58b22777a" (UID: "61bfc108-bcc1-4712-b8c8-8bc58b22777a"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.162993 4943 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.163038 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kksv6\" (UniqueName: \"kubernetes.io/projected/61bfc108-bcc1-4712-b8c8-8bc58b22777a-kube-api-access-kksv6\") on node \"crc\" DevicePath \"\"" Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.163054 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nknnt\" (UniqueName: \"kubernetes.io/projected/db16093b-eede-415e-8145-56e579a8782a-kube-api-access-nknnt\") on node \"crc\" DevicePath \"\"" Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.163066 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0e83c041-7178-43e4-b05d-e1bcc0ec36bd-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.163077 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/61bfc108-bcc1-4712-b8c8-8bc58b22777a-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.163090 4943 reconciler_common.go:293] "Volume detached for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/61bfc108-bcc1-4712-b8c8-8bc58b22777a-lock\") on node \"crc\" DevicePath \"\"" Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.163099 4943 reconciler_common.go:293] "Volume detached for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/0e83c041-7178-43e4-b05d-e1bcc0ec36bd-lock\") on node \"crc\" DevicePath \"\"" Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.163118 4943 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.163131 4943 reconciler_common.go:293] "Volume detached for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/61bfc108-bcc1-4712-b8c8-8bc58b22777a-cache\") on node \"crc\" DevicePath \"\"" Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.163141 4943 reconciler_common.go:293] "Volume detached for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/db16093b-eede-415e-8145-56e579a8782a-lock\") on node \"crc\" DevicePath \"\"" Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.163152 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-llm5v\" (UniqueName: \"kubernetes.io/projected/0e83c041-7178-43e4-b05d-e1bcc0ec36bd-kube-api-access-llm5v\") on node \"crc\" DevicePath \"\"" Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.163168 4943 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.163179 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/db16093b-eede-415e-8145-56e579a8782a-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.163189 4943 reconciler_common.go:293] "Volume detached for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/0e83c041-7178-43e4-b05d-e1bcc0ec36bd-cache\") on node \"crc\" DevicePath \"\"" Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.175125 4943 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.181632 4943 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.189167 4943 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.264357 4943 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.264669 4943 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.264687 4943 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.704424 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"db16093b-eede-415e-8145-56e579a8782a","Type":"ContainerDied","Data":"f00d6313816b2a63ab9071197558bf9bdce2ddd970da2a5f725ea0264364e1b6"} Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.704532 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-storage-1" Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.704542 4943 scope.go:117] "RemoveContainer" containerID="a89ba23a18bbf856e369b9997db5b15c2e073e42db3f087a7b8b5524d27f75c6" Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.715810 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"61bfc108-bcc1-4712-b8c8-8bc58b22777a","Type":"ContainerDied","Data":"ebd4939e3816860bb570de03484c98b55a451688036c6832ff420c3f48a74250"} Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.716073 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-storage-0" Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.751364 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"0e83c041-7178-43e4-b05d-e1bcc0ec36bd","Type":"ContainerDied","Data":"9bd03386ba29d23157b54d68484fe1fc385e534892b2e7d7d73856be11bbea58"} Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.751569 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-storage-2" Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.760570 4943 scope.go:117] "RemoveContainer" containerID="968bf9296a3299fbf29b116fdf1d701bb11bdad0ce42eb97dd829de219f7c9d3" Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.809164 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-storage-0"] Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.810290 4943 scope.go:117] "RemoveContainer" containerID="1258864ced31fd4d6f607ce7df12ec446e74a8d5bc0a46cf2122b7b2d5b1972d" Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.822105 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-storage-0"] Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.833316 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-storage-1"] Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.838762 4943 scope.go:117] "RemoveContainer" containerID="1676e9d2423174fdbd7bb0a1ca19bd1ce772561c44dd6229f48cf4244b017fd8" Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.839692 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-storage-1"] Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.846131 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-storage-2"] Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.850299 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-storage-2"] Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.858862 4943 scope.go:117] "RemoveContainer" containerID="8d4253973cf92b8f134d333111c459dd45a28b58115e11f17fba582a8fce9cfd" Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.876526 4943 scope.go:117] "RemoveContainer" containerID="635ececc7dce091575d667fc55ff70dae5854a71fa610c71180581ff05c44e85" Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.894371 4943 scope.go:117] "RemoveContainer" containerID="6f3ce1ef2dba1d61c13484a1020aa353732fd7e4b8aca4d62f081d429dd9cf7f" Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.912115 4943 scope.go:117] "RemoveContainer" containerID="4115dfe7381bf3fa3917d022ff124a7aa33326a18d5838b3caf284409aed99a2" Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.930487 4943 scope.go:117] "RemoveContainer" containerID="e651e02c4c021413625c666e0f2aae32c58a1f777868324172ec3a75e9fbdd1d" Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.951905 4943 scope.go:117] "RemoveContainer" containerID="f325a942655495f7cfc8dda622634e063c67e9ae302d7bc629c67eafa8aa4902" Mar 07 14:58:31 crc kubenswrapper[4943]: I0307 14:58:31.976174 4943 scope.go:117] "RemoveContainer" containerID="c89ce52dcf78c4288953c7557fe606114f20cd3a3351e3c89c2537818ce3be10" Mar 07 14:58:32 crc kubenswrapper[4943]: I0307 14:58:32.003820 4943 scope.go:117] "RemoveContainer" containerID="fde39a16da2d051fe287b573e0e4cdba6b70fea2e8f7445ef402a9a97f173ca7" Mar 07 14:58:32 crc kubenswrapper[4943]: I0307 14:58:32.031726 4943 scope.go:117] "RemoveContainer" containerID="1e42ab3da5309b6ad0251fd6b7b50713d2bf82db8ebe73f72da280111db9a63a" Mar 07 14:58:32 crc kubenswrapper[4943]: I0307 14:58:32.060129 4943 scope.go:117] "RemoveContainer" containerID="c7d049e1848ff61a6aa9a2cd8b145da1ddf8e0d5675e4ddaf1b547116fa455b2" Mar 07 14:58:32 crc kubenswrapper[4943]: I0307 14:58:32.103258 4943 scope.go:117] "RemoveContainer" containerID="e346ec17aa2d2cdfc931a3fd9158238065a5520b87ac1b27e152b7596b21a812" Mar 07 14:58:32 crc kubenswrapper[4943]: I0307 14:58:32.134807 4943 scope.go:117] "RemoveContainer" containerID="95bb5217f844e378e57512480c7f0c59e7b85d4265619fb46dcdc132dc57a739" Mar 07 14:58:32 crc kubenswrapper[4943]: I0307 14:58:32.167101 4943 scope.go:117] "RemoveContainer" containerID="6efa5c1c20f09b9e71b1f4e99886eb7db8e5d70a8d132abf7a9908a54e7dbbea" Mar 07 14:58:32 crc kubenswrapper[4943]: I0307 14:58:32.201115 4943 scope.go:117] "RemoveContainer" containerID="26f55f0f6682744f5ed31577037541f7b99894bdd3dd19a114e1590e2b8697ab" Mar 07 14:58:32 crc kubenswrapper[4943]: I0307 14:58:32.226047 4943 scope.go:117] "RemoveContainer" containerID="7b896439aae01ff52fa43a69d231483bddd5f7624f25904fc8ae68de8c2dc3cc" Mar 07 14:58:32 crc kubenswrapper[4943]: I0307 14:58:32.241790 4943 scope.go:117] "RemoveContainer" containerID="b5daa7f5307510498acd32d81ac9f7c0aab7c6308c53dc697686fd07d1776777" Mar 07 14:58:32 crc kubenswrapper[4943]: I0307 14:58:32.254145 4943 scope.go:117] "RemoveContainer" containerID="8a21f246526c030a1c8871a1c9d61670fe1dc6f835254920aaa17f9ba7229d99" Mar 07 14:58:32 crc kubenswrapper[4943]: I0307 14:58:32.270503 4943 scope.go:117] "RemoveContainer" containerID="d93dddf647b1cf741991cbef8e4bb89720ea9c094dfde025aa45574b1b46d1b3" Mar 07 14:58:32 crc kubenswrapper[4943]: I0307 14:58:32.287077 4943 scope.go:117] "RemoveContainer" containerID="1d86529c5cd723c960db81582958aec90ea60c5ae034f987b60f532622638e91" Mar 07 14:58:32 crc kubenswrapper[4943]: I0307 14:58:32.304948 4943 scope.go:117] "RemoveContainer" containerID="a8781652607147e67c9640dde21b8ab12919f2616b8841b8eb11c1a3990f61ff" Mar 07 14:58:32 crc kubenswrapper[4943]: I0307 14:58:32.334261 4943 scope.go:117] "RemoveContainer" containerID="6cd56e8740db694809a028cca63c9a4433743dfb0ebe9e04a1967f2b37e9d293" Mar 07 14:58:32 crc kubenswrapper[4943]: I0307 14:58:32.360189 4943 scope.go:117] "RemoveContainer" containerID="aa111fa2f0651801d3a0703bbef1b6650d2d7b0c5e863ca4573074b7061bd9e1" Mar 07 14:58:32 crc kubenswrapper[4943]: I0307 14:58:32.374216 4943 scope.go:117] "RemoveContainer" containerID="68e7c594bb36cc709f03800c5db0a2d96729018353fac6d60e8cb7a4ba2fa238" Mar 07 14:58:32 crc kubenswrapper[4943]: I0307 14:58:32.392077 4943 scope.go:117] "RemoveContainer" containerID="6f41a0b66118e9f7bdbd7717f791fbb3f6762f06eeb7e0aff3405d58daa4958a" Mar 07 14:58:32 crc kubenswrapper[4943]: I0307 14:58:32.408303 4943 scope.go:117] "RemoveContainer" containerID="e6a97829dc442d8db3046d52420793bbe248c6e4dacb29a69942321938a60bfa" Mar 07 14:58:32 crc kubenswrapper[4943]: I0307 14:58:32.435768 4943 scope.go:117] "RemoveContainer" containerID="3e06c0aeda3fbc189f9882999ab5a9183aa33314c68104b0f48b5857654c2239" Mar 07 14:58:32 crc kubenswrapper[4943]: I0307 14:58:32.461197 4943 scope.go:117] "RemoveContainer" containerID="7c67a0fd742d59f47a0d1f6a924f182e1e84266f97330d060208d8f777bad883" Mar 07 14:58:32 crc kubenswrapper[4943]: I0307 14:58:32.490988 4943 scope.go:117] "RemoveContainer" containerID="737a49457d09cd8479c49fa391100fe4f3b79bf41392edda0bf94b505382ab04" Mar 07 14:58:32 crc kubenswrapper[4943]: I0307 14:58:32.517708 4943 scope.go:117] "RemoveContainer" containerID="2ee77dbb77dbb65d1523838e6afd9d5a7b72613d080864ff79215b4a508ae442" Mar 07 14:58:32 crc kubenswrapper[4943]: I0307 14:58:32.543882 4943 scope.go:117] "RemoveContainer" containerID="ba8857ea0c02e4a8248cffb1f9d8e7c2b66181c038e948a8f69e9f91e89979a7" Mar 07 14:58:32 crc kubenswrapper[4943]: I0307 14:58:32.576110 4943 scope.go:117] "RemoveContainer" containerID="7b3105463e57fbdbaa2e126058c6aaa054286f515ac013e5b967046d33b29603" Mar 07 14:58:32 crc kubenswrapper[4943]: I0307 14:58:32.609710 4943 scope.go:117] "RemoveContainer" containerID="e571f26eb8b825f06454b7946c558c7a1eb1f69258f0d8ada3b654c53c97b55b" Mar 07 14:58:32 crc kubenswrapper[4943]: I0307 14:58:32.645269 4943 scope.go:117] "RemoveContainer" containerID="257741f8218f29b0b1e5e8d5ef95ecf3f28bee16cdc292d77de8f3326bbe123d" Mar 07 14:58:32 crc kubenswrapper[4943]: I0307 14:58:32.673796 4943 scope.go:117] "RemoveContainer" containerID="0f1f14f4fa5711a1d3b92eef66fde7b260fbee7a7bbeed91f897f960ef2e7c3e" Mar 07 14:58:32 crc kubenswrapper[4943]: I0307 14:58:32.702417 4943 scope.go:117] "RemoveContainer" containerID="185a3b1b195164b0391a65fb58c65f24f848fd66987f773e5d4f2031d582ad76" Mar 07 14:58:32 crc kubenswrapper[4943]: I0307 14:58:32.730223 4943 scope.go:117] "RemoveContainer" containerID="8a933cae33841023d45b5c9e60e92f252afc264f4c9d875e212fd937dfe8a709" Mar 07 14:58:32 crc kubenswrapper[4943]: I0307 14:58:32.767192 4943 scope.go:117] "RemoveContainer" containerID="b1ab2126394629d1a8dbbaa0f7f865780384c107928dc42b81568566f59b1cdd" Mar 07 14:58:32 crc kubenswrapper[4943]: I0307 14:58:32.782653 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" path="/var/lib/kubelet/pods/0e83c041-7178-43e4-b05d-e1bcc0ec36bd/volumes" Mar 07 14:58:32 crc kubenswrapper[4943]: I0307 14:58:32.800838 4943 scope.go:117] "RemoveContainer" containerID="466fe0dc2e8ddc693dc18033ff08abb79536de9d9c21b6015040b06e7454b177" Mar 07 14:58:32 crc kubenswrapper[4943]: I0307 14:58:32.804425 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" path="/var/lib/kubelet/pods/61bfc108-bcc1-4712-b8c8-8bc58b22777a/volumes" Mar 07 14:58:32 crc kubenswrapper[4943]: I0307 14:58:32.808220 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db16093b-eede-415e-8145-56e579a8782a" path="/var/lib/kubelet/pods/db16093b-eede-415e-8145-56e579a8782a/volumes" Mar 07 14:58:32 crc kubenswrapper[4943]: I0307 14:58:32.830218 4943 scope.go:117] "RemoveContainer" containerID="386892a8a101df1e91f2ce3827d6706952226a956fdbb507d91575d4831f42b3" Mar 07 14:58:32 crc kubenswrapper[4943]: I0307 14:58:32.856904 4943 scope.go:117] "RemoveContainer" containerID="a36e51a00bdb5a305ca9e42e0cf8e5db2c0802a50081b0774e6d53712e2848da" Mar 07 14:58:32 crc kubenswrapper[4943]: I0307 14:58:32.881635 4943 scope.go:117] "RemoveContainer" containerID="378e9e0fc8363d06ebee218138d567cd679e69eab4d30a8c35d8d6ba0c009165" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.814766 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-storage-0"] Mar 07 14:58:34 crc kubenswrapper[4943]: E0307 14:58:34.815060 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="account-reaper" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.815072 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="account-reaper" Mar 07 14:58:34 crc kubenswrapper[4943]: E0307 14:58:34.815080 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="container-updater" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.815088 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="container-updater" Mar 07 14:58:34 crc kubenswrapper[4943]: E0307 14:58:34.815100 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="account-replicator" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.815106 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="account-replicator" Mar 07 14:58:34 crc kubenswrapper[4943]: E0307 14:58:34.815115 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="account-auditor" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.815121 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="account-auditor" Mar 07 14:58:34 crc kubenswrapper[4943]: E0307 14:58:34.815127 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="account-server" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.815132 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="account-server" Mar 07 14:58:34 crc kubenswrapper[4943]: E0307 14:58:34.815141 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81fd2fdf-28bf-40a5-92a9-8506339ed373" containerName="proxy-server" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.815147 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="81fd2fdf-28bf-40a5-92a9-8506339ed373" containerName="proxy-server" Mar 07 14:58:34 crc kubenswrapper[4943]: E0307 14:58:34.815153 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="container-server" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.815159 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="container-server" Mar 07 14:58:34 crc kubenswrapper[4943]: E0307 14:58:34.815169 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="account-reaper" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.815175 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="account-reaper" Mar 07 14:58:34 crc kubenswrapper[4943]: E0307 14:58:34.815187 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="account-auditor" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.815192 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="account-auditor" Mar 07 14:58:34 crc kubenswrapper[4943]: E0307 14:58:34.815203 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="object-replicator" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.815209 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="object-replicator" Mar 07 14:58:34 crc kubenswrapper[4943]: E0307 14:58:34.815217 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="container-replicator" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.815222 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="container-replicator" Mar 07 14:58:34 crc kubenswrapper[4943]: E0307 14:58:34.815229 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6621055b-9adc-4057-bc14-63b4848edbca" containerName="oc" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.815235 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="6621055b-9adc-4057-bc14-63b4848edbca" containerName="oc" Mar 07 14:58:34 crc kubenswrapper[4943]: E0307 14:58:34.815244 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="swift-recon-cron" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.815250 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="swift-recon-cron" Mar 07 14:58:34 crc kubenswrapper[4943]: E0307 14:58:34.815260 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="swift-recon-cron" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.815267 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="swift-recon-cron" Mar 07 14:58:34 crc kubenswrapper[4943]: E0307 14:58:34.815276 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="container-replicator" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.815282 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="container-replicator" Mar 07 14:58:34 crc kubenswrapper[4943]: E0307 14:58:34.815292 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="object-replicator" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.815298 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="object-replicator" Mar 07 14:58:34 crc kubenswrapper[4943]: E0307 14:58:34.815308 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="object-auditor" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.815314 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="object-auditor" Mar 07 14:58:34 crc kubenswrapper[4943]: E0307 14:58:34.815321 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="rsync" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.815327 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="rsync" Mar 07 14:58:34 crc kubenswrapper[4943]: E0307 14:58:34.815336 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="object-server" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.815341 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="object-server" Mar 07 14:58:34 crc kubenswrapper[4943]: E0307 14:58:34.815352 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="container-auditor" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.815358 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="container-auditor" Mar 07 14:58:34 crc kubenswrapper[4943]: E0307 14:58:34.815364 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="container-replicator" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.815370 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="container-replicator" Mar 07 14:58:34 crc kubenswrapper[4943]: E0307 14:58:34.815381 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="container-auditor" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.815387 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="container-auditor" Mar 07 14:58:34 crc kubenswrapper[4943]: E0307 14:58:34.815398 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="object-server" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.815404 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="object-server" Mar 07 14:58:34 crc kubenswrapper[4943]: E0307 14:58:34.815414 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="container-server" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.815420 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="container-server" Mar 07 14:58:34 crc kubenswrapper[4943]: E0307 14:58:34.815429 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="object-updater" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.815435 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="object-updater" Mar 07 14:58:34 crc kubenswrapper[4943]: E0307 14:58:34.815444 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="object-replicator" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.815450 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="object-replicator" Mar 07 14:58:34 crc kubenswrapper[4943]: E0307 14:58:34.815461 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="account-replicator" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.815468 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="account-replicator" Mar 07 14:58:34 crc kubenswrapper[4943]: E0307 14:58:34.815479 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="rsync" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.815485 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="rsync" Mar 07 14:58:34 crc kubenswrapper[4943]: E0307 14:58:34.815525 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2907f386-4042-4031-bbb1-327ac38b4173" containerName="swift-ring-rebalance" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.815548 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="2907f386-4042-4031-bbb1-327ac38b4173" containerName="swift-ring-rebalance" Mar 07 14:58:34 crc kubenswrapper[4943]: E0307 14:58:34.815563 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="account-server" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.815572 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="account-server" Mar 07 14:58:34 crc kubenswrapper[4943]: E0307 14:58:34.815580 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="account-reaper" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.815586 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="account-reaper" Mar 07 14:58:34 crc kubenswrapper[4943]: E0307 14:58:34.815593 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="object-auditor" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.815598 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="object-auditor" Mar 07 14:58:34 crc kubenswrapper[4943]: E0307 14:58:34.815608 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="swift-recon-cron" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.815614 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="swift-recon-cron" Mar 07 14:58:34 crc kubenswrapper[4943]: E0307 14:58:34.815621 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="object-expirer" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.815626 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="object-expirer" Mar 07 14:58:34 crc kubenswrapper[4943]: E0307 14:58:34.815635 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="object-expirer" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.815641 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="object-expirer" Mar 07 14:58:34 crc kubenswrapper[4943]: E0307 14:58:34.815647 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="container-auditor" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.815653 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="container-auditor" Mar 07 14:58:34 crc kubenswrapper[4943]: E0307 14:58:34.815660 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="account-replicator" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.815666 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="account-replicator" Mar 07 14:58:34 crc kubenswrapper[4943]: E0307 14:58:34.815672 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="container-server" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.815683 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="container-server" Mar 07 14:58:34 crc kubenswrapper[4943]: E0307 14:58:34.815767 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="object-updater" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.815773 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="object-updater" Mar 07 14:58:34 crc kubenswrapper[4943]: E0307 14:58:34.815786 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="object-auditor" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.815792 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="object-auditor" Mar 07 14:58:34 crc kubenswrapper[4943]: E0307 14:58:34.815800 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="object-updater" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.815805 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="object-updater" Mar 07 14:58:34 crc kubenswrapper[4943]: E0307 14:58:34.815813 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="container-updater" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.815818 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="container-updater" Mar 07 14:58:34 crc kubenswrapper[4943]: E0307 14:58:34.815829 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="account-auditor" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.815836 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="account-auditor" Mar 07 14:58:34 crc kubenswrapper[4943]: E0307 14:58:34.815842 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81fd2fdf-28bf-40a5-92a9-8506339ed373" containerName="proxy-httpd" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.815848 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="81fd2fdf-28bf-40a5-92a9-8506339ed373" containerName="proxy-httpd" Mar 07 14:58:34 crc kubenswrapper[4943]: E0307 14:58:34.815857 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="object-expirer" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.815862 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="object-expirer" Mar 07 14:58:34 crc kubenswrapper[4943]: E0307 14:58:34.815871 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="rsync" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.815877 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="rsync" Mar 07 14:58:34 crc kubenswrapper[4943]: E0307 14:58:34.815890 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="container-updater" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.815896 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="container-updater" Mar 07 14:58:34 crc kubenswrapper[4943]: E0307 14:58:34.815907 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="account-server" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.815912 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="account-server" Mar 07 14:58:34 crc kubenswrapper[4943]: E0307 14:58:34.815918 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="object-server" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.815936 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="object-server" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.816063 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="container-auditor" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.816075 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="object-server" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.816084 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="container-updater" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.816091 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="container-updater" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.816100 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="container-server" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.816107 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="object-updater" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.816114 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="object-server" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.816125 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="container-auditor" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.816135 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="account-server" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.816166 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="object-expirer" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.816176 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="object-updater" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.816183 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="object-replicator" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.816192 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="rsync" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.816201 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="swift-recon-cron" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.816206 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="2907f386-4042-4031-bbb1-327ac38b4173" containerName="swift-ring-rebalance" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.816213 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="81fd2fdf-28bf-40a5-92a9-8506339ed373" containerName="proxy-server" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.816243 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="object-auditor" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.816249 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="object-expirer" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.816256 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="account-reaper" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.816263 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="account-server" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.816271 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="81fd2fdf-28bf-40a5-92a9-8506339ed373" containerName="proxy-httpd" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.816277 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="container-server" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.816283 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="account-replicator" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.816290 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="object-auditor" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.816297 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="account-auditor" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.816322 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="account-reaper" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.816330 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="account-replicator" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.816335 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="object-updater" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.816343 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="account-replicator" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.816351 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="rsync" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.816358 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="container-updater" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.816363 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="object-replicator" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.816370 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="6621055b-9adc-4057-bc14-63b4848edbca" containerName="oc" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.816377 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="container-replicator" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.816403 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="account-auditor" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.816410 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="account-reaper" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.816419 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="swift-recon-cron" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.816429 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="object-replicator" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.816435 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="account-server" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.816445 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="object-expirer" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.816452 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="object-server" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.816479 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="object-auditor" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.816489 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="container-replicator" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.816500 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="61bfc108-bcc1-4712-b8c8-8bc58b22777a" containerName="swift-recon-cron" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.816509 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="container-server" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.816519 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="account-auditor" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.816529 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="container-replicator" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.816559 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="db16093b-eede-415e-8145-56e579a8782a" containerName="container-auditor" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.816568 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e83c041-7178-43e4-b05d-e1bcc0ec36bd" containerName="rsync" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.823843 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-storage-0" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.829962 4943 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"swift-conf" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.830077 4943 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"swift-swift-dockercfg-ndb5p" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.830366 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-files" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.830676 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-storage-config-data" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.889056 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-storage-0"] Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.933142 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8-cache\") pod \"swift-storage-0\" (UID: \"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8\") " pod="swift-kuttl-tests/swift-storage-0" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.933195 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-scpxj\" (UniqueName: \"kubernetes.io/projected/3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8-kube-api-access-scpxj\") pod \"swift-storage-0\" (UID: \"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8\") " pod="swift-kuttl-tests/swift-storage-0" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.933236 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8-lock\") pod \"swift-storage-0\" (UID: \"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8\") " pod="swift-kuttl-tests/swift-storage-0" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.933443 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8-etc-swift\") pod \"swift-storage-0\" (UID: \"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8\") " pod="swift-kuttl-tests/swift-storage-0" Mar 07 14:58:34 crc kubenswrapper[4943]: I0307 14:58:34.933478 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8\") " pod="swift-kuttl-tests/swift-storage-0" Mar 07 14:58:35 crc kubenswrapper[4943]: I0307 14:58:35.035257 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8-etc-swift\") pod \"swift-storage-0\" (UID: \"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8\") " pod="swift-kuttl-tests/swift-storage-0" Mar 07 14:58:35 crc kubenswrapper[4943]: I0307 14:58:35.035340 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8\") " pod="swift-kuttl-tests/swift-storage-0" Mar 07 14:58:35 crc kubenswrapper[4943]: I0307 14:58:35.035493 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8-cache\") pod \"swift-storage-0\" (UID: \"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8\") " pod="swift-kuttl-tests/swift-storage-0" Mar 07 14:58:35 crc kubenswrapper[4943]: E0307 14:58:35.035526 4943 projected.go:288] Couldn't get configMap swift-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Mar 07 14:58:35 crc kubenswrapper[4943]: E0307 14:58:35.035566 4943 projected.go:194] Error preparing data for projected volume etc-swift for pod swift-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Mar 07 14:58:35 crc kubenswrapper[4943]: I0307 14:58:35.035532 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-scpxj\" (UniqueName: \"kubernetes.io/projected/3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8-kube-api-access-scpxj\") pod \"swift-storage-0\" (UID: \"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8\") " pod="swift-kuttl-tests/swift-storage-0" Mar 07 14:58:35 crc kubenswrapper[4943]: E0307 14:58:35.035882 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8-etc-swift podName:3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8 nodeName:}" failed. No retries permitted until 2026-03-07 14:58:35.535668985 +0000 UTC m=+1157.487805523 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8-etc-swift") pod "swift-storage-0" (UID: "3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8") : configmap "swift-ring-files" not found Mar 07 14:58:35 crc kubenswrapper[4943]: I0307 14:58:35.035988 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8-lock\") pod \"swift-storage-0\" (UID: \"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8\") " pod="swift-kuttl-tests/swift-storage-0" Mar 07 14:58:35 crc kubenswrapper[4943]: I0307 14:58:35.036220 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8-cache\") pod \"swift-storage-0\" (UID: \"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8\") " pod="swift-kuttl-tests/swift-storage-0" Mar 07 14:58:35 crc kubenswrapper[4943]: I0307 14:58:35.036227 4943 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8\") device mount path \"/mnt/openstack/pv07\"" pod="swift-kuttl-tests/swift-storage-0" Mar 07 14:58:35 crc kubenswrapper[4943]: I0307 14:58:35.036644 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8-lock\") pod \"swift-storage-0\" (UID: \"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8\") " pod="swift-kuttl-tests/swift-storage-0" Mar 07 14:58:35 crc kubenswrapper[4943]: I0307 14:58:35.074137 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-scpxj\" (UniqueName: \"kubernetes.io/projected/3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8-kube-api-access-scpxj\") pod \"swift-storage-0\" (UID: \"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8\") " pod="swift-kuttl-tests/swift-storage-0" Mar 07 14:58:35 crc kubenswrapper[4943]: I0307 14:58:35.087162 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8\") " pod="swift-kuttl-tests/swift-storage-0" Mar 07 14:58:35 crc kubenswrapper[4943]: I0307 14:58:35.547873 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8-etc-swift\") pod \"swift-storage-0\" (UID: \"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8\") " pod="swift-kuttl-tests/swift-storage-0" Mar 07 14:58:35 crc kubenswrapper[4943]: E0307 14:58:35.548104 4943 projected.go:288] Couldn't get configMap swift-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Mar 07 14:58:35 crc kubenswrapper[4943]: E0307 14:58:35.548146 4943 projected.go:194] Error preparing data for projected volume etc-swift for pod swift-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Mar 07 14:58:35 crc kubenswrapper[4943]: E0307 14:58:35.548227 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8-etc-swift podName:3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8 nodeName:}" failed. No retries permitted until 2026-03-07 14:58:36.548196632 +0000 UTC m=+1158.500333160 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8-etc-swift") pod "swift-storage-0" (UID: "3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8") : configmap "swift-ring-files" not found Mar 07 14:58:36 crc kubenswrapper[4943]: I0307 14:58:36.074286 4943 patch_prober.go:28] interesting pod/machine-config-daemon-cpdmm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 07 14:58:36 crc kubenswrapper[4943]: I0307 14:58:36.074374 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 07 14:58:36 crc kubenswrapper[4943]: I0307 14:58:36.564087 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8-etc-swift\") pod \"swift-storage-0\" (UID: \"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8\") " pod="swift-kuttl-tests/swift-storage-0" Mar 07 14:58:36 crc kubenswrapper[4943]: E0307 14:58:36.564338 4943 projected.go:288] Couldn't get configMap swift-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Mar 07 14:58:36 crc kubenswrapper[4943]: E0307 14:58:36.564392 4943 projected.go:194] Error preparing data for projected volume etc-swift for pod swift-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Mar 07 14:58:36 crc kubenswrapper[4943]: E0307 14:58:36.564498 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8-etc-swift podName:3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8 nodeName:}" failed. No retries permitted until 2026-03-07 14:58:38.564454493 +0000 UTC m=+1160.516591021 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8-etc-swift") pod "swift-storage-0" (UID: "3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8") : configmap "swift-ring-files" not found Mar 07 14:58:38 crc kubenswrapper[4943]: I0307 14:58:38.623641 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8-etc-swift\") pod \"swift-storage-0\" (UID: \"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8\") " pod="swift-kuttl-tests/swift-storage-0" Mar 07 14:58:38 crc kubenswrapper[4943]: E0307 14:58:38.623844 4943 projected.go:288] Couldn't get configMap swift-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Mar 07 14:58:38 crc kubenswrapper[4943]: E0307 14:58:38.624080 4943 projected.go:194] Error preparing data for projected volume etc-swift for pod swift-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Mar 07 14:58:38 crc kubenswrapper[4943]: E0307 14:58:38.624161 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8-etc-swift podName:3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8 nodeName:}" failed. No retries permitted until 2026-03-07 14:58:42.624136777 +0000 UTC m=+1164.576273305 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8-etc-swift") pod "swift-storage-0" (UID: "3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8") : configmap "swift-ring-files" not found Mar 07 14:58:38 crc kubenswrapper[4943]: I0307 14:58:38.693356 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-mvdbn"] Mar 07 14:58:38 crc kubenswrapper[4943]: I0307 14:58:38.695893 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-mvdbn" Mar 07 14:58:38 crc kubenswrapper[4943]: I0307 14:58:38.699900 4943 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"swift-proxy-config-data" Mar 07 14:58:38 crc kubenswrapper[4943]: I0307 14:58:38.700913 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 14:58:38 crc kubenswrapper[4943]: I0307 14:58:38.703023 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 14:58:38 crc kubenswrapper[4943]: I0307 14:58:38.717347 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-mvdbn"] Mar 07 14:58:38 crc kubenswrapper[4943]: I0307 14:58:38.725841 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/24902a59-0296-47f0-b64e-2291d9fe0ee2-ring-data-devices\") pod \"swift-ring-rebalance-mvdbn\" (UID: \"24902a59-0296-47f0-b64e-2291d9fe0ee2\") " pod="swift-kuttl-tests/swift-ring-rebalance-mvdbn" Mar 07 14:58:38 crc kubenswrapper[4943]: I0307 14:58:38.726164 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/24902a59-0296-47f0-b64e-2291d9fe0ee2-scripts\") pod \"swift-ring-rebalance-mvdbn\" (UID: \"24902a59-0296-47f0-b64e-2291d9fe0ee2\") " pod="swift-kuttl-tests/swift-ring-rebalance-mvdbn" Mar 07 14:58:38 crc kubenswrapper[4943]: I0307 14:58:38.726396 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnwdj\" (UniqueName: \"kubernetes.io/projected/24902a59-0296-47f0-b64e-2291d9fe0ee2-kube-api-access-hnwdj\") pod \"swift-ring-rebalance-mvdbn\" (UID: \"24902a59-0296-47f0-b64e-2291d9fe0ee2\") " pod="swift-kuttl-tests/swift-ring-rebalance-mvdbn" Mar 07 14:58:38 crc kubenswrapper[4943]: I0307 14:58:38.726454 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/24902a59-0296-47f0-b64e-2291d9fe0ee2-dispersionconf\") pod \"swift-ring-rebalance-mvdbn\" (UID: \"24902a59-0296-47f0-b64e-2291d9fe0ee2\") " pod="swift-kuttl-tests/swift-ring-rebalance-mvdbn" Mar 07 14:58:38 crc kubenswrapper[4943]: I0307 14:58:38.726552 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/24902a59-0296-47f0-b64e-2291d9fe0ee2-swiftconf\") pod \"swift-ring-rebalance-mvdbn\" (UID: \"24902a59-0296-47f0-b64e-2291d9fe0ee2\") " pod="swift-kuttl-tests/swift-ring-rebalance-mvdbn" Mar 07 14:58:38 crc kubenswrapper[4943]: I0307 14:58:38.726587 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/24902a59-0296-47f0-b64e-2291d9fe0ee2-etc-swift\") pod \"swift-ring-rebalance-mvdbn\" (UID: \"24902a59-0296-47f0-b64e-2291d9fe0ee2\") " pod="swift-kuttl-tests/swift-ring-rebalance-mvdbn" Mar 07 14:58:38 crc kubenswrapper[4943]: I0307 14:58:38.827639 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnwdj\" (UniqueName: \"kubernetes.io/projected/24902a59-0296-47f0-b64e-2291d9fe0ee2-kube-api-access-hnwdj\") pod \"swift-ring-rebalance-mvdbn\" (UID: \"24902a59-0296-47f0-b64e-2291d9fe0ee2\") " pod="swift-kuttl-tests/swift-ring-rebalance-mvdbn" Mar 07 14:58:38 crc kubenswrapper[4943]: I0307 14:58:38.827780 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/24902a59-0296-47f0-b64e-2291d9fe0ee2-dispersionconf\") pod \"swift-ring-rebalance-mvdbn\" (UID: \"24902a59-0296-47f0-b64e-2291d9fe0ee2\") " pod="swift-kuttl-tests/swift-ring-rebalance-mvdbn" Mar 07 14:58:38 crc kubenswrapper[4943]: I0307 14:58:38.827839 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/24902a59-0296-47f0-b64e-2291d9fe0ee2-swiftconf\") pod \"swift-ring-rebalance-mvdbn\" (UID: \"24902a59-0296-47f0-b64e-2291d9fe0ee2\") " pod="swift-kuttl-tests/swift-ring-rebalance-mvdbn" Mar 07 14:58:38 crc kubenswrapper[4943]: I0307 14:58:38.827859 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/24902a59-0296-47f0-b64e-2291d9fe0ee2-etc-swift\") pod \"swift-ring-rebalance-mvdbn\" (UID: \"24902a59-0296-47f0-b64e-2291d9fe0ee2\") " pod="swift-kuttl-tests/swift-ring-rebalance-mvdbn" Mar 07 14:58:38 crc kubenswrapper[4943]: I0307 14:58:38.827957 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/24902a59-0296-47f0-b64e-2291d9fe0ee2-ring-data-devices\") pod \"swift-ring-rebalance-mvdbn\" (UID: \"24902a59-0296-47f0-b64e-2291d9fe0ee2\") " pod="swift-kuttl-tests/swift-ring-rebalance-mvdbn" Mar 07 14:58:38 crc kubenswrapper[4943]: I0307 14:58:38.828006 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/24902a59-0296-47f0-b64e-2291d9fe0ee2-scripts\") pod \"swift-ring-rebalance-mvdbn\" (UID: \"24902a59-0296-47f0-b64e-2291d9fe0ee2\") " pod="swift-kuttl-tests/swift-ring-rebalance-mvdbn" Mar 07 14:58:38 crc kubenswrapper[4943]: I0307 14:58:38.828464 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/24902a59-0296-47f0-b64e-2291d9fe0ee2-etc-swift\") pod \"swift-ring-rebalance-mvdbn\" (UID: \"24902a59-0296-47f0-b64e-2291d9fe0ee2\") " pod="swift-kuttl-tests/swift-ring-rebalance-mvdbn" Mar 07 14:58:38 crc kubenswrapper[4943]: I0307 14:58:38.828713 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/24902a59-0296-47f0-b64e-2291d9fe0ee2-ring-data-devices\") pod \"swift-ring-rebalance-mvdbn\" (UID: \"24902a59-0296-47f0-b64e-2291d9fe0ee2\") " pod="swift-kuttl-tests/swift-ring-rebalance-mvdbn" Mar 07 14:58:38 crc kubenswrapper[4943]: I0307 14:58:38.829237 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/24902a59-0296-47f0-b64e-2291d9fe0ee2-scripts\") pod \"swift-ring-rebalance-mvdbn\" (UID: \"24902a59-0296-47f0-b64e-2291d9fe0ee2\") " pod="swift-kuttl-tests/swift-ring-rebalance-mvdbn" Mar 07 14:58:38 crc kubenswrapper[4943]: I0307 14:58:38.836427 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/24902a59-0296-47f0-b64e-2291d9fe0ee2-dispersionconf\") pod \"swift-ring-rebalance-mvdbn\" (UID: \"24902a59-0296-47f0-b64e-2291d9fe0ee2\") " pod="swift-kuttl-tests/swift-ring-rebalance-mvdbn" Mar 07 14:58:38 crc kubenswrapper[4943]: I0307 14:58:38.836450 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/24902a59-0296-47f0-b64e-2291d9fe0ee2-swiftconf\") pod \"swift-ring-rebalance-mvdbn\" (UID: \"24902a59-0296-47f0-b64e-2291d9fe0ee2\") " pod="swift-kuttl-tests/swift-ring-rebalance-mvdbn" Mar 07 14:58:38 crc kubenswrapper[4943]: I0307 14:58:38.856310 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnwdj\" (UniqueName: \"kubernetes.io/projected/24902a59-0296-47f0-b64e-2291d9fe0ee2-kube-api-access-hnwdj\") pod \"swift-ring-rebalance-mvdbn\" (UID: \"24902a59-0296-47f0-b64e-2291d9fe0ee2\") " pod="swift-kuttl-tests/swift-ring-rebalance-mvdbn" Mar 07 14:58:39 crc kubenswrapper[4943]: I0307 14:58:39.045746 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-mvdbn" Mar 07 14:58:39 crc kubenswrapper[4943]: I0307 14:58:39.322390 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-mvdbn"] Mar 07 14:58:39 crc kubenswrapper[4943]: W0307 14:58:39.326877 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod24902a59_0296_47f0_b64e_2291d9fe0ee2.slice/crio-b862b180a8a758d0013e670b4e8e4a32841bb7660588d32f7e9f9016699c5b15 WatchSource:0}: Error finding container b862b180a8a758d0013e670b4e8e4a32841bb7660588d32f7e9f9016699c5b15: Status 404 returned error can't find the container with id b862b180a8a758d0013e670b4e8e4a32841bb7660588d32f7e9f9016699c5b15 Mar 07 14:58:39 crc kubenswrapper[4943]: I0307 14:58:39.895435 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-mvdbn" event={"ID":"24902a59-0296-47f0-b64e-2291d9fe0ee2","Type":"ContainerStarted","Data":"0894765bb00937a71e749c9f473984ddcb66ad8efbd3f954b5ea0513b71a5a6a"} Mar 07 14:58:39 crc kubenswrapper[4943]: I0307 14:58:39.895851 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-mvdbn" event={"ID":"24902a59-0296-47f0-b64e-2291d9fe0ee2","Type":"ContainerStarted","Data":"b862b180a8a758d0013e670b4e8e4a32841bb7660588d32f7e9f9016699c5b15"} Mar 07 14:58:39 crc kubenswrapper[4943]: I0307 14:58:39.922253 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-mvdbn" podStartSLOduration=1.922224077 podStartE2EDuration="1.922224077s" podCreationTimestamp="2026-03-07 14:58:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:58:39.916247731 +0000 UTC m=+1161.868384239" watchObservedRunningTime="2026-03-07 14:58:39.922224077 +0000 UTC m=+1161.874360605" Mar 07 14:58:42 crc kubenswrapper[4943]: I0307 14:58:42.690650 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8-etc-swift\") pod \"swift-storage-0\" (UID: \"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8\") " pod="swift-kuttl-tests/swift-storage-0" Mar 07 14:58:42 crc kubenswrapper[4943]: E0307 14:58:42.694221 4943 projected.go:288] Couldn't get configMap swift-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Mar 07 14:58:42 crc kubenswrapper[4943]: E0307 14:58:42.694660 4943 projected.go:194] Error preparing data for projected volume etc-swift for pod swift-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Mar 07 14:58:42 crc kubenswrapper[4943]: E0307 14:58:42.694901 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8-etc-swift podName:3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8 nodeName:}" failed. No retries permitted until 2026-03-07 14:58:50.694873147 +0000 UTC m=+1172.647009675 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8-etc-swift") pod "swift-storage-0" (UID: "3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8") : configmap "swift-ring-files" not found Mar 07 14:58:45 crc kubenswrapper[4943]: I0307 14:58:45.941776 4943 generic.go:334] "Generic (PLEG): container finished" podID="24902a59-0296-47f0-b64e-2291d9fe0ee2" containerID="0894765bb00937a71e749c9f473984ddcb66ad8efbd3f954b5ea0513b71a5a6a" exitCode=0 Mar 07 14:58:45 crc kubenswrapper[4943]: I0307 14:58:45.941833 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-mvdbn" event={"ID":"24902a59-0296-47f0-b64e-2291d9fe0ee2","Type":"ContainerDied","Data":"0894765bb00937a71e749c9f473984ddcb66ad8efbd3f954b5ea0513b71a5a6a"} Mar 07 14:58:47 crc kubenswrapper[4943]: I0307 14:58:47.297900 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-mvdbn" Mar 07 14:58:47 crc kubenswrapper[4943]: I0307 14:58:47.464590 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/24902a59-0296-47f0-b64e-2291d9fe0ee2-scripts\") pod \"24902a59-0296-47f0-b64e-2291d9fe0ee2\" (UID: \"24902a59-0296-47f0-b64e-2291d9fe0ee2\") " Mar 07 14:58:47 crc kubenswrapper[4943]: I0307 14:58:47.464674 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/24902a59-0296-47f0-b64e-2291d9fe0ee2-swiftconf\") pod \"24902a59-0296-47f0-b64e-2291d9fe0ee2\" (UID: \"24902a59-0296-47f0-b64e-2291d9fe0ee2\") " Mar 07 14:58:47 crc kubenswrapper[4943]: I0307 14:58:47.464824 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/24902a59-0296-47f0-b64e-2291d9fe0ee2-ring-data-devices\") pod \"24902a59-0296-47f0-b64e-2291d9fe0ee2\" (UID: \"24902a59-0296-47f0-b64e-2291d9fe0ee2\") " Mar 07 14:58:47 crc kubenswrapper[4943]: I0307 14:58:47.464914 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/24902a59-0296-47f0-b64e-2291d9fe0ee2-dispersionconf\") pod \"24902a59-0296-47f0-b64e-2291d9fe0ee2\" (UID: \"24902a59-0296-47f0-b64e-2291d9fe0ee2\") " Mar 07 14:58:47 crc kubenswrapper[4943]: I0307 14:58:47.464983 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/24902a59-0296-47f0-b64e-2291d9fe0ee2-etc-swift\") pod \"24902a59-0296-47f0-b64e-2291d9fe0ee2\" (UID: \"24902a59-0296-47f0-b64e-2291d9fe0ee2\") " Mar 07 14:58:47 crc kubenswrapper[4943]: I0307 14:58:47.465012 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hnwdj\" (UniqueName: \"kubernetes.io/projected/24902a59-0296-47f0-b64e-2291d9fe0ee2-kube-api-access-hnwdj\") pod \"24902a59-0296-47f0-b64e-2291d9fe0ee2\" (UID: \"24902a59-0296-47f0-b64e-2291d9fe0ee2\") " Mar 07 14:58:47 crc kubenswrapper[4943]: I0307 14:58:47.467273 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/24902a59-0296-47f0-b64e-2291d9fe0ee2-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "24902a59-0296-47f0-b64e-2291d9fe0ee2" (UID: "24902a59-0296-47f0-b64e-2291d9fe0ee2"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:58:47 crc kubenswrapper[4943]: I0307 14:58:47.468010 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/24902a59-0296-47f0-b64e-2291d9fe0ee2-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "24902a59-0296-47f0-b64e-2291d9fe0ee2" (UID: "24902a59-0296-47f0-b64e-2291d9fe0ee2"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:58:47 crc kubenswrapper[4943]: I0307 14:58:47.471820 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24902a59-0296-47f0-b64e-2291d9fe0ee2-kube-api-access-hnwdj" (OuterVolumeSpecName: "kube-api-access-hnwdj") pod "24902a59-0296-47f0-b64e-2291d9fe0ee2" (UID: "24902a59-0296-47f0-b64e-2291d9fe0ee2"). InnerVolumeSpecName "kube-api-access-hnwdj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:58:47 crc kubenswrapper[4943]: I0307 14:58:47.476259 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24902a59-0296-47f0-b64e-2291d9fe0ee2-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "24902a59-0296-47f0-b64e-2291d9fe0ee2" (UID: "24902a59-0296-47f0-b64e-2291d9fe0ee2"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:58:47 crc kubenswrapper[4943]: I0307 14:58:47.497436 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24902a59-0296-47f0-b64e-2291d9fe0ee2-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "24902a59-0296-47f0-b64e-2291d9fe0ee2" (UID: "24902a59-0296-47f0-b64e-2291d9fe0ee2"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:58:47 crc kubenswrapper[4943]: I0307 14:58:47.511630 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/24902a59-0296-47f0-b64e-2291d9fe0ee2-scripts" (OuterVolumeSpecName: "scripts") pod "24902a59-0296-47f0-b64e-2291d9fe0ee2" (UID: "24902a59-0296-47f0-b64e-2291d9fe0ee2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:58:47 crc kubenswrapper[4943]: I0307 14:58:47.566382 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/24902a59-0296-47f0-b64e-2291d9fe0ee2-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 14:58:47 crc kubenswrapper[4943]: I0307 14:58:47.566411 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hnwdj\" (UniqueName: \"kubernetes.io/projected/24902a59-0296-47f0-b64e-2291d9fe0ee2-kube-api-access-hnwdj\") on node \"crc\" DevicePath \"\"" Mar 07 14:58:47 crc kubenswrapper[4943]: I0307 14:58:47.566422 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/24902a59-0296-47f0-b64e-2291d9fe0ee2-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 14:58:47 crc kubenswrapper[4943]: I0307 14:58:47.566431 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/24902a59-0296-47f0-b64e-2291d9fe0ee2-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 14:58:47 crc kubenswrapper[4943]: I0307 14:58:47.566440 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/24902a59-0296-47f0-b64e-2291d9fe0ee2-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 14:58:47 crc kubenswrapper[4943]: I0307 14:58:47.566449 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/24902a59-0296-47f0-b64e-2291d9fe0ee2-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 14:58:47 crc kubenswrapper[4943]: I0307 14:58:47.959343 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-mvdbn" event={"ID":"24902a59-0296-47f0-b64e-2291d9fe0ee2","Type":"ContainerDied","Data":"b862b180a8a758d0013e670b4e8e4a32841bb7660588d32f7e9f9016699c5b15"} Mar 07 14:58:47 crc kubenswrapper[4943]: I0307 14:58:47.959728 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b862b180a8a758d0013e670b4e8e4a32841bb7660588d32f7e9f9016699c5b15" Mar 07 14:58:47 crc kubenswrapper[4943]: I0307 14:58:47.959824 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-mvdbn" Mar 07 14:58:50 crc kubenswrapper[4943]: I0307 14:58:50.718175 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8-etc-swift\") pod \"swift-storage-0\" (UID: \"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8\") " pod="swift-kuttl-tests/swift-storage-0" Mar 07 14:58:50 crc kubenswrapper[4943]: I0307 14:58:50.730729 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8-etc-swift\") pod \"swift-storage-0\" (UID: \"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8\") " pod="swift-kuttl-tests/swift-storage-0" Mar 07 14:58:50 crc kubenswrapper[4943]: I0307 14:58:50.781704 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-storage-0" Mar 07 14:58:51 crc kubenswrapper[4943]: I0307 14:58:51.339841 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-storage-0"] Mar 07 14:58:52 crc kubenswrapper[4943]: I0307 14:58:52.000041 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8","Type":"ContainerStarted","Data":"5a68e14eb4c97f4f249f077f9c0d016c858544f0da60e1d6c84ee0b53bc3a259"} Mar 07 14:58:52 crc kubenswrapper[4943]: I0307 14:58:52.000088 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8","Type":"ContainerStarted","Data":"c16bdb03bef6cc9ab7fb55a6953d241e9190e80787f4e3de43c893cd1ed349a2"} Mar 07 14:58:52 crc kubenswrapper[4943]: I0307 14:58:52.000102 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8","Type":"ContainerStarted","Data":"9b6765692b58c5ebd2755fcbcc0bf6fee3ddbe781cfa1b626ad6078346d3e402"} Mar 07 14:58:52 crc kubenswrapper[4943]: I0307 14:58:52.000114 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8","Type":"ContainerStarted","Data":"8ec1d04cd278b86217515269c7955bd492054280abca8f7dd7e6efa559069292"} Mar 07 14:58:53 crc kubenswrapper[4943]: I0307 14:58:53.028939 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8","Type":"ContainerStarted","Data":"f4a58ec1a6b1a5a66faa2ca708935002ee19db5bef2c8c99a431c0bb5ec9eea3"} Mar 07 14:58:53 crc kubenswrapper[4943]: I0307 14:58:53.029212 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8","Type":"ContainerStarted","Data":"4009d66a16af4cf49d7b909280c182c6a2839bf7dc72660a38b3c02320550d9f"} Mar 07 14:58:53 crc kubenswrapper[4943]: I0307 14:58:53.029226 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8","Type":"ContainerStarted","Data":"40df134ae24e3ed4a8fd377fb725953a304574200ac3148e3f9b3b3806e44a47"} Mar 07 14:58:53 crc kubenswrapper[4943]: I0307 14:58:53.029235 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8","Type":"ContainerStarted","Data":"cc374a57378de01b25998555023a72a91a81bfe6f48e0c2d74901d09dc497799"} Mar 07 14:58:53 crc kubenswrapper[4943]: I0307 14:58:53.029244 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8","Type":"ContainerStarted","Data":"a269396243ea039c5a1ff714244738a9af9f115de9be8ca31a6f67a9a2bb2e94"} Mar 07 14:58:53 crc kubenswrapper[4943]: I0307 14:58:53.029254 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8","Type":"ContainerStarted","Data":"059e70bef0f14c03d2e83f29ffb3a0153b9683edc84e88eed7fa99bc0a0d8579"} Mar 07 14:58:54 crc kubenswrapper[4943]: I0307 14:58:54.059477 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8","Type":"ContainerStarted","Data":"a75dfedf2adeb9783f42686241f798122e938a4edd4336b201d2e124ee87e1da"} Mar 07 14:58:54 crc kubenswrapper[4943]: I0307 14:58:54.059785 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8","Type":"ContainerStarted","Data":"a1e3e8dc6afbf3013626e71f36dcde17e61c0ed8650ec0a48d8c2ddccb2aa585"} Mar 07 14:58:54 crc kubenswrapper[4943]: I0307 14:58:54.059797 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8","Type":"ContainerStarted","Data":"06b728b91732e6aa41de478ccef9fb0046a86c5a1dc8baedd3243f9a3c0ed18c"} Mar 07 14:58:54 crc kubenswrapper[4943]: I0307 14:58:54.059804 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8","Type":"ContainerStarted","Data":"b96feec40ee02a1bd0a7a1c8d67abf8300f9b59dbae2f6e54e6ae30df4a3971b"} Mar 07 14:58:54 crc kubenswrapper[4943]: I0307 14:58:54.059812 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8","Type":"ContainerStarted","Data":"f290059bb335b82ed354d4d229fb8530d57e7f29a151602dd18122d411960328"} Mar 07 14:58:54 crc kubenswrapper[4943]: I0307 14:58:54.059820 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8","Type":"ContainerStarted","Data":"a3b3ff008ced9b38eb096a8fe7d155531763f2bdf2e5399ff2af469f95e76448"} Mar 07 14:58:54 crc kubenswrapper[4943]: I0307 14:58:54.059828 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8","Type":"ContainerStarted","Data":"3d606e2c04c5358f4c4404f1af78027ac23f0b1d763a512e4323fdd0240efd19"} Mar 07 14:58:54 crc kubenswrapper[4943]: I0307 14:58:54.112796 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-storage-0" podStartSLOduration=21.112776469 podStartE2EDuration="21.112776469s" podCreationTimestamp="2026-03-07 14:58:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:58:54.100883518 +0000 UTC m=+1176.053020066" watchObservedRunningTime="2026-03-07 14:58:54.112776469 +0000 UTC m=+1176.064912967" Mar 07 14:59:03 crc kubenswrapper[4943]: I0307 14:59:03.655960 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-proxy-646594d5c9-q4vdw"] Mar 07 14:59:03 crc kubenswrapper[4943]: E0307 14:59:03.657086 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24902a59-0296-47f0-b64e-2291d9fe0ee2" containerName="swift-ring-rebalance" Mar 07 14:59:03 crc kubenswrapper[4943]: I0307 14:59:03.657107 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="24902a59-0296-47f0-b64e-2291d9fe0ee2" containerName="swift-ring-rebalance" Mar 07 14:59:03 crc kubenswrapper[4943]: I0307 14:59:03.657422 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="24902a59-0296-47f0-b64e-2291d9fe0ee2" containerName="swift-ring-rebalance" Mar 07 14:59:03 crc kubenswrapper[4943]: I0307 14:59:03.658770 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-proxy-646594d5c9-q4vdw" Mar 07 14:59:03 crc kubenswrapper[4943]: I0307 14:59:03.664237 4943 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"swift-proxy-config-data" Mar 07 14:59:03 crc kubenswrapper[4943]: I0307 14:59:03.680247 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-proxy-646594d5c9-q4vdw"] Mar 07 14:59:03 crc kubenswrapper[4943]: I0307 14:59:03.811618 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wdb4\" (UniqueName: \"kubernetes.io/projected/6d17441e-f6b6-4fab-8239-0409b7bb3910-kube-api-access-9wdb4\") pod \"swift-proxy-646594d5c9-q4vdw\" (UID: \"6d17441e-f6b6-4fab-8239-0409b7bb3910\") " pod="swift-kuttl-tests/swift-proxy-646594d5c9-q4vdw" Mar 07 14:59:03 crc kubenswrapper[4943]: I0307 14:59:03.811689 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6d17441e-f6b6-4fab-8239-0409b7bb3910-log-httpd\") pod \"swift-proxy-646594d5c9-q4vdw\" (UID: \"6d17441e-f6b6-4fab-8239-0409b7bb3910\") " pod="swift-kuttl-tests/swift-proxy-646594d5c9-q4vdw" Mar 07 14:59:03 crc kubenswrapper[4943]: I0307 14:59:03.811744 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6d17441e-f6b6-4fab-8239-0409b7bb3910-etc-swift\") pod \"swift-proxy-646594d5c9-q4vdw\" (UID: \"6d17441e-f6b6-4fab-8239-0409b7bb3910\") " pod="swift-kuttl-tests/swift-proxy-646594d5c9-q4vdw" Mar 07 14:59:03 crc kubenswrapper[4943]: I0307 14:59:03.811881 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d17441e-f6b6-4fab-8239-0409b7bb3910-config-data\") pod \"swift-proxy-646594d5c9-q4vdw\" (UID: \"6d17441e-f6b6-4fab-8239-0409b7bb3910\") " pod="swift-kuttl-tests/swift-proxy-646594d5c9-q4vdw" Mar 07 14:59:03 crc kubenswrapper[4943]: I0307 14:59:03.811918 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6d17441e-f6b6-4fab-8239-0409b7bb3910-run-httpd\") pod \"swift-proxy-646594d5c9-q4vdw\" (UID: \"6d17441e-f6b6-4fab-8239-0409b7bb3910\") " pod="swift-kuttl-tests/swift-proxy-646594d5c9-q4vdw" Mar 07 14:59:03 crc kubenswrapper[4943]: I0307 14:59:03.913331 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wdb4\" (UniqueName: \"kubernetes.io/projected/6d17441e-f6b6-4fab-8239-0409b7bb3910-kube-api-access-9wdb4\") pod \"swift-proxy-646594d5c9-q4vdw\" (UID: \"6d17441e-f6b6-4fab-8239-0409b7bb3910\") " pod="swift-kuttl-tests/swift-proxy-646594d5c9-q4vdw" Mar 07 14:59:03 crc kubenswrapper[4943]: I0307 14:59:03.913477 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6d17441e-f6b6-4fab-8239-0409b7bb3910-log-httpd\") pod \"swift-proxy-646594d5c9-q4vdw\" (UID: \"6d17441e-f6b6-4fab-8239-0409b7bb3910\") " pod="swift-kuttl-tests/swift-proxy-646594d5c9-q4vdw" Mar 07 14:59:03 crc kubenswrapper[4943]: I0307 14:59:03.913601 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6d17441e-f6b6-4fab-8239-0409b7bb3910-etc-swift\") pod \"swift-proxy-646594d5c9-q4vdw\" (UID: \"6d17441e-f6b6-4fab-8239-0409b7bb3910\") " pod="swift-kuttl-tests/swift-proxy-646594d5c9-q4vdw" Mar 07 14:59:03 crc kubenswrapper[4943]: I0307 14:59:03.913689 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d17441e-f6b6-4fab-8239-0409b7bb3910-config-data\") pod \"swift-proxy-646594d5c9-q4vdw\" (UID: \"6d17441e-f6b6-4fab-8239-0409b7bb3910\") " pod="swift-kuttl-tests/swift-proxy-646594d5c9-q4vdw" Mar 07 14:59:03 crc kubenswrapper[4943]: I0307 14:59:03.913730 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6d17441e-f6b6-4fab-8239-0409b7bb3910-run-httpd\") pod \"swift-proxy-646594d5c9-q4vdw\" (UID: \"6d17441e-f6b6-4fab-8239-0409b7bb3910\") " pod="swift-kuttl-tests/swift-proxy-646594d5c9-q4vdw" Mar 07 14:59:03 crc kubenswrapper[4943]: I0307 14:59:03.914248 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6d17441e-f6b6-4fab-8239-0409b7bb3910-log-httpd\") pod \"swift-proxy-646594d5c9-q4vdw\" (UID: \"6d17441e-f6b6-4fab-8239-0409b7bb3910\") " pod="swift-kuttl-tests/swift-proxy-646594d5c9-q4vdw" Mar 07 14:59:03 crc kubenswrapper[4943]: I0307 14:59:03.914619 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6d17441e-f6b6-4fab-8239-0409b7bb3910-run-httpd\") pod \"swift-proxy-646594d5c9-q4vdw\" (UID: \"6d17441e-f6b6-4fab-8239-0409b7bb3910\") " pod="swift-kuttl-tests/swift-proxy-646594d5c9-q4vdw" Mar 07 14:59:03 crc kubenswrapper[4943]: I0307 14:59:03.923643 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d17441e-f6b6-4fab-8239-0409b7bb3910-config-data\") pod \"swift-proxy-646594d5c9-q4vdw\" (UID: \"6d17441e-f6b6-4fab-8239-0409b7bb3910\") " pod="swift-kuttl-tests/swift-proxy-646594d5c9-q4vdw" Mar 07 14:59:03 crc kubenswrapper[4943]: I0307 14:59:03.924186 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6d17441e-f6b6-4fab-8239-0409b7bb3910-etc-swift\") pod \"swift-proxy-646594d5c9-q4vdw\" (UID: \"6d17441e-f6b6-4fab-8239-0409b7bb3910\") " pod="swift-kuttl-tests/swift-proxy-646594d5c9-q4vdw" Mar 07 14:59:03 crc kubenswrapper[4943]: I0307 14:59:03.946646 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wdb4\" (UniqueName: \"kubernetes.io/projected/6d17441e-f6b6-4fab-8239-0409b7bb3910-kube-api-access-9wdb4\") pod \"swift-proxy-646594d5c9-q4vdw\" (UID: \"6d17441e-f6b6-4fab-8239-0409b7bb3910\") " pod="swift-kuttl-tests/swift-proxy-646594d5c9-q4vdw" Mar 07 14:59:03 crc kubenswrapper[4943]: I0307 14:59:03.980963 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-proxy-646594d5c9-q4vdw" Mar 07 14:59:04 crc kubenswrapper[4943]: I0307 14:59:04.266634 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-proxy-646594d5c9-q4vdw"] Mar 07 14:59:04 crc kubenswrapper[4943]: W0307 14:59:04.277137 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6d17441e_f6b6_4fab_8239_0409b7bb3910.slice/crio-5ba7b7da95bb2163a0f0a086f63f0cb2ccc1e9447e39660d8df0ba17604ac53d WatchSource:0}: Error finding container 5ba7b7da95bb2163a0f0a086f63f0cb2ccc1e9447e39660d8df0ba17604ac53d: Status 404 returned error can't find the container with id 5ba7b7da95bb2163a0f0a086f63f0cb2ccc1e9447e39660d8df0ba17604ac53d Mar 07 14:59:05 crc kubenswrapper[4943]: I0307 14:59:05.184998 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-proxy-646594d5c9-q4vdw" event={"ID":"6d17441e-f6b6-4fab-8239-0409b7bb3910","Type":"ContainerStarted","Data":"be38a7d18eaf729bc3bfa860152f8612e983a3ab4366c7882bc771e0466ac733"} Mar 07 14:59:05 crc kubenswrapper[4943]: I0307 14:59:05.185506 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="swift-kuttl-tests/swift-proxy-646594d5c9-q4vdw" Mar 07 14:59:05 crc kubenswrapper[4943]: I0307 14:59:05.185541 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-proxy-646594d5c9-q4vdw" event={"ID":"6d17441e-f6b6-4fab-8239-0409b7bb3910","Type":"ContainerStarted","Data":"ea89ba800a6d23b6b4bc4c690cd8df8bb7d32b6ee9b1f7bebbebd0d0592f47f9"} Mar 07 14:59:05 crc kubenswrapper[4943]: I0307 14:59:05.185568 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-proxy-646594d5c9-q4vdw" event={"ID":"6d17441e-f6b6-4fab-8239-0409b7bb3910","Type":"ContainerStarted","Data":"5ba7b7da95bb2163a0f0a086f63f0cb2ccc1e9447e39660d8df0ba17604ac53d"} Mar 07 14:59:05 crc kubenswrapper[4943]: I0307 14:59:05.220094 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-proxy-646594d5c9-q4vdw" podStartSLOduration=2.220072677 podStartE2EDuration="2.220072677s" podCreationTimestamp="2026-03-07 14:59:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:59:05.211362583 +0000 UTC m=+1187.163499081" watchObservedRunningTime="2026-03-07 14:59:05.220072677 +0000 UTC m=+1187.172209175" Mar 07 14:59:06 crc kubenswrapper[4943]: I0307 14:59:06.074550 4943 patch_prober.go:28] interesting pod/machine-config-daemon-cpdmm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 07 14:59:06 crc kubenswrapper[4943]: I0307 14:59:06.074641 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 07 14:59:06 crc kubenswrapper[4943]: I0307 14:59:06.074703 4943 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" Mar 07 14:59:06 crc kubenswrapper[4943]: I0307 14:59:06.075522 4943 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"808321a0c8648ed7264aba7412d4317b4e5e5302eaf1e96b5220f23b29527cec"} pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 07 14:59:06 crc kubenswrapper[4943]: I0307 14:59:06.075645 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerName="machine-config-daemon" containerID="cri-o://808321a0c8648ed7264aba7412d4317b4e5e5302eaf1e96b5220f23b29527cec" gracePeriod=600 Mar 07 14:59:06 crc kubenswrapper[4943]: I0307 14:59:06.195150 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="swift-kuttl-tests/swift-proxy-646594d5c9-q4vdw" Mar 07 14:59:07 crc kubenswrapper[4943]: I0307 14:59:07.207052 4943 generic.go:334] "Generic (PLEG): container finished" podID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerID="808321a0c8648ed7264aba7412d4317b4e5e5302eaf1e96b5220f23b29527cec" exitCode=0 Mar 07 14:59:07 crc kubenswrapper[4943]: I0307 14:59:07.207152 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" event={"ID":"4c1f18ea-e557-41fc-95dd-ce4e1722a61e","Type":"ContainerDied","Data":"808321a0c8648ed7264aba7412d4317b4e5e5302eaf1e96b5220f23b29527cec"} Mar 07 14:59:07 crc kubenswrapper[4943]: I0307 14:59:07.208067 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" event={"ID":"4c1f18ea-e557-41fc-95dd-ce4e1722a61e","Type":"ContainerStarted","Data":"526f9b1306d1082bdb6e6f98c1f04669533c0a9945daeb9cde6060502da5cecb"} Mar 07 14:59:07 crc kubenswrapper[4943]: I0307 14:59:07.208106 4943 scope.go:117] "RemoveContainer" containerID="fe852e1f10f4ca72f749789ca1e6ab73542eaedb5349732682970babc3f1afd1" Mar 07 14:59:13 crc kubenswrapper[4943]: I0307 14:59:13.986733 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="swift-kuttl-tests/swift-proxy-646594d5c9-q4vdw" Mar 07 14:59:13 crc kubenswrapper[4943]: I0307 14:59:13.994802 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="swift-kuttl-tests/swift-proxy-646594d5c9-q4vdw" Mar 07 14:59:16 crc kubenswrapper[4943]: I0307 14:59:16.315821 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-454h9"] Mar 07 14:59:16 crc kubenswrapper[4943]: I0307 14:59:16.317507 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-454h9" Mar 07 14:59:16 crc kubenswrapper[4943]: I0307 14:59:16.320061 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 14:59:16 crc kubenswrapper[4943]: I0307 14:59:16.321616 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 14:59:16 crc kubenswrapper[4943]: I0307 14:59:16.331361 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-454h9"] Mar 07 14:59:16 crc kubenswrapper[4943]: I0307 14:59:16.415563 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/0e6eb86f-e384-4d31-a64b-7cac66aed4fb-dispersionconf\") pod \"swift-ring-rebalance-debug-454h9\" (UID: \"0e6eb86f-e384-4d31-a64b-7cac66aed4fb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-454h9" Mar 07 14:59:16 crc kubenswrapper[4943]: I0307 14:59:16.415683 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/0e6eb86f-e384-4d31-a64b-7cac66aed4fb-swiftconf\") pod \"swift-ring-rebalance-debug-454h9\" (UID: \"0e6eb86f-e384-4d31-a64b-7cac66aed4fb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-454h9" Mar 07 14:59:16 crc kubenswrapper[4943]: I0307 14:59:16.415767 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0e6eb86f-e384-4d31-a64b-7cac66aed4fb-scripts\") pod \"swift-ring-rebalance-debug-454h9\" (UID: \"0e6eb86f-e384-4d31-a64b-7cac66aed4fb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-454h9" Mar 07 14:59:16 crc kubenswrapper[4943]: I0307 14:59:16.415818 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/0e6eb86f-e384-4d31-a64b-7cac66aed4fb-ring-data-devices\") pod \"swift-ring-rebalance-debug-454h9\" (UID: \"0e6eb86f-e384-4d31-a64b-7cac66aed4fb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-454h9" Mar 07 14:59:16 crc kubenswrapper[4943]: I0307 14:59:16.415862 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9k7nk\" (UniqueName: \"kubernetes.io/projected/0e6eb86f-e384-4d31-a64b-7cac66aed4fb-kube-api-access-9k7nk\") pod \"swift-ring-rebalance-debug-454h9\" (UID: \"0e6eb86f-e384-4d31-a64b-7cac66aed4fb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-454h9" Mar 07 14:59:16 crc kubenswrapper[4943]: I0307 14:59:16.416128 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/0e6eb86f-e384-4d31-a64b-7cac66aed4fb-etc-swift\") pod \"swift-ring-rebalance-debug-454h9\" (UID: \"0e6eb86f-e384-4d31-a64b-7cac66aed4fb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-454h9" Mar 07 14:59:16 crc kubenswrapper[4943]: I0307 14:59:16.517486 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/0e6eb86f-e384-4d31-a64b-7cac66aed4fb-dispersionconf\") pod \"swift-ring-rebalance-debug-454h9\" (UID: \"0e6eb86f-e384-4d31-a64b-7cac66aed4fb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-454h9" Mar 07 14:59:16 crc kubenswrapper[4943]: I0307 14:59:16.517614 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/0e6eb86f-e384-4d31-a64b-7cac66aed4fb-swiftconf\") pod \"swift-ring-rebalance-debug-454h9\" (UID: \"0e6eb86f-e384-4d31-a64b-7cac66aed4fb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-454h9" Mar 07 14:59:16 crc kubenswrapper[4943]: I0307 14:59:16.517682 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0e6eb86f-e384-4d31-a64b-7cac66aed4fb-scripts\") pod \"swift-ring-rebalance-debug-454h9\" (UID: \"0e6eb86f-e384-4d31-a64b-7cac66aed4fb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-454h9" Mar 07 14:59:16 crc kubenswrapper[4943]: I0307 14:59:16.517751 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/0e6eb86f-e384-4d31-a64b-7cac66aed4fb-ring-data-devices\") pod \"swift-ring-rebalance-debug-454h9\" (UID: \"0e6eb86f-e384-4d31-a64b-7cac66aed4fb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-454h9" Mar 07 14:59:16 crc kubenswrapper[4943]: I0307 14:59:16.517807 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9k7nk\" (UniqueName: \"kubernetes.io/projected/0e6eb86f-e384-4d31-a64b-7cac66aed4fb-kube-api-access-9k7nk\") pod \"swift-ring-rebalance-debug-454h9\" (UID: \"0e6eb86f-e384-4d31-a64b-7cac66aed4fb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-454h9" Mar 07 14:59:16 crc kubenswrapper[4943]: I0307 14:59:16.517870 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/0e6eb86f-e384-4d31-a64b-7cac66aed4fb-etc-swift\") pod \"swift-ring-rebalance-debug-454h9\" (UID: \"0e6eb86f-e384-4d31-a64b-7cac66aed4fb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-454h9" Mar 07 14:59:16 crc kubenswrapper[4943]: I0307 14:59:16.518978 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/0e6eb86f-e384-4d31-a64b-7cac66aed4fb-etc-swift\") pod \"swift-ring-rebalance-debug-454h9\" (UID: \"0e6eb86f-e384-4d31-a64b-7cac66aed4fb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-454h9" Mar 07 14:59:16 crc kubenswrapper[4943]: I0307 14:59:16.519443 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/0e6eb86f-e384-4d31-a64b-7cac66aed4fb-ring-data-devices\") pod \"swift-ring-rebalance-debug-454h9\" (UID: \"0e6eb86f-e384-4d31-a64b-7cac66aed4fb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-454h9" Mar 07 14:59:16 crc kubenswrapper[4943]: I0307 14:59:16.519459 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0e6eb86f-e384-4d31-a64b-7cac66aed4fb-scripts\") pod \"swift-ring-rebalance-debug-454h9\" (UID: \"0e6eb86f-e384-4d31-a64b-7cac66aed4fb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-454h9" Mar 07 14:59:16 crc kubenswrapper[4943]: I0307 14:59:16.528458 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/0e6eb86f-e384-4d31-a64b-7cac66aed4fb-swiftconf\") pod \"swift-ring-rebalance-debug-454h9\" (UID: \"0e6eb86f-e384-4d31-a64b-7cac66aed4fb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-454h9" Mar 07 14:59:16 crc kubenswrapper[4943]: I0307 14:59:16.528726 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/0e6eb86f-e384-4d31-a64b-7cac66aed4fb-dispersionconf\") pod \"swift-ring-rebalance-debug-454h9\" (UID: \"0e6eb86f-e384-4d31-a64b-7cac66aed4fb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-454h9" Mar 07 14:59:16 crc kubenswrapper[4943]: I0307 14:59:16.546607 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9k7nk\" (UniqueName: \"kubernetes.io/projected/0e6eb86f-e384-4d31-a64b-7cac66aed4fb-kube-api-access-9k7nk\") pod \"swift-ring-rebalance-debug-454h9\" (UID: \"0e6eb86f-e384-4d31-a64b-7cac66aed4fb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-454h9" Mar 07 14:59:16 crc kubenswrapper[4943]: I0307 14:59:16.654412 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-454h9" Mar 07 14:59:17 crc kubenswrapper[4943]: I0307 14:59:17.151355 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-454h9"] Mar 07 14:59:17 crc kubenswrapper[4943]: W0307 14:59:17.162022 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0e6eb86f_e384_4d31_a64b_7cac66aed4fb.slice/crio-98a77db97dbd8bc7b12cfcef23a0f339ac028f668fdac92d9844bb6c196aab1c WatchSource:0}: Error finding container 98a77db97dbd8bc7b12cfcef23a0f339ac028f668fdac92d9844bb6c196aab1c: Status 404 returned error can't find the container with id 98a77db97dbd8bc7b12cfcef23a0f339ac028f668fdac92d9844bb6c196aab1c Mar 07 14:59:17 crc kubenswrapper[4943]: I0307 14:59:17.302099 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-454h9" event={"ID":"0e6eb86f-e384-4d31-a64b-7cac66aed4fb","Type":"ContainerStarted","Data":"98a77db97dbd8bc7b12cfcef23a0f339ac028f668fdac92d9844bb6c196aab1c"} Mar 07 14:59:18 crc kubenswrapper[4943]: I0307 14:59:18.310700 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-454h9" event={"ID":"0e6eb86f-e384-4d31-a64b-7cac66aed4fb","Type":"ContainerStarted","Data":"c31fb8459a7f0666328afbbc11aa2081c9308b186d597d14883f1cc883104844"} Mar 07 14:59:18 crc kubenswrapper[4943]: I0307 14:59:18.335249 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-454h9" podStartSLOduration=2.335196729 podStartE2EDuration="2.335196729s" podCreationTimestamp="2026-03-07 14:59:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:59:18.33318822 +0000 UTC m=+1200.285324778" watchObservedRunningTime="2026-03-07 14:59:18.335196729 +0000 UTC m=+1200.287333277" Mar 07 14:59:20 crc kubenswrapper[4943]: I0307 14:59:20.332386 4943 generic.go:334] "Generic (PLEG): container finished" podID="0e6eb86f-e384-4d31-a64b-7cac66aed4fb" containerID="c31fb8459a7f0666328afbbc11aa2081c9308b186d597d14883f1cc883104844" exitCode=0 Mar 07 14:59:20 crc kubenswrapper[4943]: I0307 14:59:20.332477 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-454h9" event={"ID":"0e6eb86f-e384-4d31-a64b-7cac66aed4fb","Type":"ContainerDied","Data":"c31fb8459a7f0666328afbbc11aa2081c9308b186d597d14883f1cc883104844"} Mar 07 14:59:21 crc kubenswrapper[4943]: I0307 14:59:21.684896 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-454h9" Mar 07 14:59:21 crc kubenswrapper[4943]: I0307 14:59:21.726923 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-454h9"] Mar 07 14:59:21 crc kubenswrapper[4943]: I0307 14:59:21.732791 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-454h9"] Mar 07 14:59:21 crc kubenswrapper[4943]: I0307 14:59:21.818133 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9k7nk\" (UniqueName: \"kubernetes.io/projected/0e6eb86f-e384-4d31-a64b-7cac66aed4fb-kube-api-access-9k7nk\") pod \"0e6eb86f-e384-4d31-a64b-7cac66aed4fb\" (UID: \"0e6eb86f-e384-4d31-a64b-7cac66aed4fb\") " Mar 07 14:59:21 crc kubenswrapper[4943]: I0307 14:59:21.818191 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/0e6eb86f-e384-4d31-a64b-7cac66aed4fb-swiftconf\") pod \"0e6eb86f-e384-4d31-a64b-7cac66aed4fb\" (UID: \"0e6eb86f-e384-4d31-a64b-7cac66aed4fb\") " Mar 07 14:59:21 crc kubenswrapper[4943]: I0307 14:59:21.818231 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/0e6eb86f-e384-4d31-a64b-7cac66aed4fb-ring-data-devices\") pod \"0e6eb86f-e384-4d31-a64b-7cac66aed4fb\" (UID: \"0e6eb86f-e384-4d31-a64b-7cac66aed4fb\") " Mar 07 14:59:21 crc kubenswrapper[4943]: I0307 14:59:21.818271 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0e6eb86f-e384-4d31-a64b-7cac66aed4fb-scripts\") pod \"0e6eb86f-e384-4d31-a64b-7cac66aed4fb\" (UID: \"0e6eb86f-e384-4d31-a64b-7cac66aed4fb\") " Mar 07 14:59:21 crc kubenswrapper[4943]: I0307 14:59:21.818308 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/0e6eb86f-e384-4d31-a64b-7cac66aed4fb-etc-swift\") pod \"0e6eb86f-e384-4d31-a64b-7cac66aed4fb\" (UID: \"0e6eb86f-e384-4d31-a64b-7cac66aed4fb\") " Mar 07 14:59:21 crc kubenswrapper[4943]: I0307 14:59:21.818339 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/0e6eb86f-e384-4d31-a64b-7cac66aed4fb-dispersionconf\") pod \"0e6eb86f-e384-4d31-a64b-7cac66aed4fb\" (UID: \"0e6eb86f-e384-4d31-a64b-7cac66aed4fb\") " Mar 07 14:59:21 crc kubenswrapper[4943]: I0307 14:59:21.818790 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e6eb86f-e384-4d31-a64b-7cac66aed4fb-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "0e6eb86f-e384-4d31-a64b-7cac66aed4fb" (UID: "0e6eb86f-e384-4d31-a64b-7cac66aed4fb"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:59:21 crc kubenswrapper[4943]: I0307 14:59:21.819884 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e6eb86f-e384-4d31-a64b-7cac66aed4fb-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "0e6eb86f-e384-4d31-a64b-7cac66aed4fb" (UID: "0e6eb86f-e384-4d31-a64b-7cac66aed4fb"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:59:21 crc kubenswrapper[4943]: I0307 14:59:21.818874 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/0e6eb86f-e384-4d31-a64b-7cac66aed4fb-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 14:59:21 crc kubenswrapper[4943]: I0307 14:59:21.826693 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e6eb86f-e384-4d31-a64b-7cac66aed4fb-kube-api-access-9k7nk" (OuterVolumeSpecName: "kube-api-access-9k7nk") pod "0e6eb86f-e384-4d31-a64b-7cac66aed4fb" (UID: "0e6eb86f-e384-4d31-a64b-7cac66aed4fb"). InnerVolumeSpecName "kube-api-access-9k7nk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:59:21 crc kubenswrapper[4943]: I0307 14:59:21.839949 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e6eb86f-e384-4d31-a64b-7cac66aed4fb-scripts" (OuterVolumeSpecName: "scripts") pod "0e6eb86f-e384-4d31-a64b-7cac66aed4fb" (UID: "0e6eb86f-e384-4d31-a64b-7cac66aed4fb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:59:21 crc kubenswrapper[4943]: I0307 14:59:21.842901 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e6eb86f-e384-4d31-a64b-7cac66aed4fb-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "0e6eb86f-e384-4d31-a64b-7cac66aed4fb" (UID: "0e6eb86f-e384-4d31-a64b-7cac66aed4fb"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:59:21 crc kubenswrapper[4943]: I0307 14:59:21.858644 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e6eb86f-e384-4d31-a64b-7cac66aed4fb-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "0e6eb86f-e384-4d31-a64b-7cac66aed4fb" (UID: "0e6eb86f-e384-4d31-a64b-7cac66aed4fb"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:59:21 crc kubenswrapper[4943]: I0307 14:59:21.866496 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-889rr"] Mar 07 14:59:21 crc kubenswrapper[4943]: E0307 14:59:21.866824 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e6eb86f-e384-4d31-a64b-7cac66aed4fb" containerName="swift-ring-rebalance" Mar 07 14:59:21 crc kubenswrapper[4943]: I0307 14:59:21.866843 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e6eb86f-e384-4d31-a64b-7cac66aed4fb" containerName="swift-ring-rebalance" Mar 07 14:59:21 crc kubenswrapper[4943]: I0307 14:59:21.866986 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e6eb86f-e384-4d31-a64b-7cac66aed4fb" containerName="swift-ring-rebalance" Mar 07 14:59:21 crc kubenswrapper[4943]: I0307 14:59:21.869077 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-889rr" Mar 07 14:59:21 crc kubenswrapper[4943]: I0307 14:59:21.884173 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-889rr"] Mar 07 14:59:21 crc kubenswrapper[4943]: I0307 14:59:21.920701 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9k7nk\" (UniqueName: \"kubernetes.io/projected/0e6eb86f-e384-4d31-a64b-7cac66aed4fb-kube-api-access-9k7nk\") on node \"crc\" DevicePath \"\"" Mar 07 14:59:21 crc kubenswrapper[4943]: I0307 14:59:21.920731 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/0e6eb86f-e384-4d31-a64b-7cac66aed4fb-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 14:59:21 crc kubenswrapper[4943]: I0307 14:59:21.920742 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0e6eb86f-e384-4d31-a64b-7cac66aed4fb-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 14:59:21 crc kubenswrapper[4943]: I0307 14:59:21.920752 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/0e6eb86f-e384-4d31-a64b-7cac66aed4fb-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 14:59:21 crc kubenswrapper[4943]: I0307 14:59:21.920761 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/0e6eb86f-e384-4d31-a64b-7cac66aed4fb-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 14:59:22 crc kubenswrapper[4943]: I0307 14:59:22.021964 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/870c9817-5966-4f86-82e0-b22084e7ce1d-etc-swift\") pod \"swift-ring-rebalance-debug-889rr\" (UID: \"870c9817-5966-4f86-82e0-b22084e7ce1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-889rr" Mar 07 14:59:22 crc kubenswrapper[4943]: I0307 14:59:22.022230 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnt7h\" (UniqueName: \"kubernetes.io/projected/870c9817-5966-4f86-82e0-b22084e7ce1d-kube-api-access-qnt7h\") pod \"swift-ring-rebalance-debug-889rr\" (UID: \"870c9817-5966-4f86-82e0-b22084e7ce1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-889rr" Mar 07 14:59:22 crc kubenswrapper[4943]: I0307 14:59:22.022326 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/870c9817-5966-4f86-82e0-b22084e7ce1d-scripts\") pod \"swift-ring-rebalance-debug-889rr\" (UID: \"870c9817-5966-4f86-82e0-b22084e7ce1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-889rr" Mar 07 14:59:22 crc kubenswrapper[4943]: I0307 14:59:22.022356 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/870c9817-5966-4f86-82e0-b22084e7ce1d-ring-data-devices\") pod \"swift-ring-rebalance-debug-889rr\" (UID: \"870c9817-5966-4f86-82e0-b22084e7ce1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-889rr" Mar 07 14:59:22 crc kubenswrapper[4943]: I0307 14:59:22.022425 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/870c9817-5966-4f86-82e0-b22084e7ce1d-dispersionconf\") pod \"swift-ring-rebalance-debug-889rr\" (UID: \"870c9817-5966-4f86-82e0-b22084e7ce1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-889rr" Mar 07 14:59:22 crc kubenswrapper[4943]: I0307 14:59:22.022487 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/870c9817-5966-4f86-82e0-b22084e7ce1d-swiftconf\") pod \"swift-ring-rebalance-debug-889rr\" (UID: \"870c9817-5966-4f86-82e0-b22084e7ce1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-889rr" Mar 07 14:59:22 crc kubenswrapper[4943]: I0307 14:59:22.123529 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/870c9817-5966-4f86-82e0-b22084e7ce1d-etc-swift\") pod \"swift-ring-rebalance-debug-889rr\" (UID: \"870c9817-5966-4f86-82e0-b22084e7ce1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-889rr" Mar 07 14:59:22 crc kubenswrapper[4943]: I0307 14:59:22.123729 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnt7h\" (UniqueName: \"kubernetes.io/projected/870c9817-5966-4f86-82e0-b22084e7ce1d-kube-api-access-qnt7h\") pod \"swift-ring-rebalance-debug-889rr\" (UID: \"870c9817-5966-4f86-82e0-b22084e7ce1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-889rr" Mar 07 14:59:22 crc kubenswrapper[4943]: I0307 14:59:22.123799 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/870c9817-5966-4f86-82e0-b22084e7ce1d-scripts\") pod \"swift-ring-rebalance-debug-889rr\" (UID: \"870c9817-5966-4f86-82e0-b22084e7ce1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-889rr" Mar 07 14:59:22 crc kubenswrapper[4943]: I0307 14:59:22.123835 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/870c9817-5966-4f86-82e0-b22084e7ce1d-ring-data-devices\") pod \"swift-ring-rebalance-debug-889rr\" (UID: \"870c9817-5966-4f86-82e0-b22084e7ce1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-889rr" Mar 07 14:59:22 crc kubenswrapper[4943]: I0307 14:59:22.123871 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/870c9817-5966-4f86-82e0-b22084e7ce1d-dispersionconf\") pod \"swift-ring-rebalance-debug-889rr\" (UID: \"870c9817-5966-4f86-82e0-b22084e7ce1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-889rr" Mar 07 14:59:22 crc kubenswrapper[4943]: I0307 14:59:22.123921 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/870c9817-5966-4f86-82e0-b22084e7ce1d-swiftconf\") pod \"swift-ring-rebalance-debug-889rr\" (UID: \"870c9817-5966-4f86-82e0-b22084e7ce1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-889rr" Mar 07 14:59:22 crc kubenswrapper[4943]: I0307 14:59:22.125320 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/870c9817-5966-4f86-82e0-b22084e7ce1d-ring-data-devices\") pod \"swift-ring-rebalance-debug-889rr\" (UID: \"870c9817-5966-4f86-82e0-b22084e7ce1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-889rr" Mar 07 14:59:22 crc kubenswrapper[4943]: I0307 14:59:22.125353 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/870c9817-5966-4f86-82e0-b22084e7ce1d-etc-swift\") pod \"swift-ring-rebalance-debug-889rr\" (UID: \"870c9817-5966-4f86-82e0-b22084e7ce1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-889rr" Mar 07 14:59:22 crc kubenswrapper[4943]: I0307 14:59:22.125500 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/870c9817-5966-4f86-82e0-b22084e7ce1d-scripts\") pod \"swift-ring-rebalance-debug-889rr\" (UID: \"870c9817-5966-4f86-82e0-b22084e7ce1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-889rr" Mar 07 14:59:22 crc kubenswrapper[4943]: I0307 14:59:22.128691 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/870c9817-5966-4f86-82e0-b22084e7ce1d-dispersionconf\") pod \"swift-ring-rebalance-debug-889rr\" (UID: \"870c9817-5966-4f86-82e0-b22084e7ce1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-889rr" Mar 07 14:59:22 crc kubenswrapper[4943]: I0307 14:59:22.129567 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/870c9817-5966-4f86-82e0-b22084e7ce1d-swiftconf\") pod \"swift-ring-rebalance-debug-889rr\" (UID: \"870c9817-5966-4f86-82e0-b22084e7ce1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-889rr" Mar 07 14:59:22 crc kubenswrapper[4943]: I0307 14:59:22.154525 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnt7h\" (UniqueName: \"kubernetes.io/projected/870c9817-5966-4f86-82e0-b22084e7ce1d-kube-api-access-qnt7h\") pod \"swift-ring-rebalance-debug-889rr\" (UID: \"870c9817-5966-4f86-82e0-b22084e7ce1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-889rr" Mar 07 14:59:22 crc kubenswrapper[4943]: I0307 14:59:22.232161 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-889rr" Mar 07 14:59:22 crc kubenswrapper[4943]: I0307 14:59:22.350700 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="98a77db97dbd8bc7b12cfcef23a0f339ac028f668fdac92d9844bb6c196aab1c" Mar 07 14:59:22 crc kubenswrapper[4943]: I0307 14:59:22.350815 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-454h9" Mar 07 14:59:22 crc kubenswrapper[4943]: I0307 14:59:22.771211 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e6eb86f-e384-4d31-a64b-7cac66aed4fb" path="/var/lib/kubelet/pods/0e6eb86f-e384-4d31-a64b-7cac66aed4fb/volumes" Mar 07 14:59:22 crc kubenswrapper[4943]: I0307 14:59:22.798357 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-889rr"] Mar 07 14:59:23 crc kubenswrapper[4943]: I0307 14:59:23.359773 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-889rr" event={"ID":"870c9817-5966-4f86-82e0-b22084e7ce1d","Type":"ContainerStarted","Data":"3249b974d903bb8bc3583a79f36e4779b8523fc5199d38645e56bcc5bcbec1d1"} Mar 07 14:59:23 crc kubenswrapper[4943]: I0307 14:59:23.360243 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-889rr" event={"ID":"870c9817-5966-4f86-82e0-b22084e7ce1d","Type":"ContainerStarted","Data":"cf3965cb5887e19f1c24f8aeb22528dbb681e913193faf94c835846d597dc5f1"} Mar 07 14:59:23 crc kubenswrapper[4943]: I0307 14:59:23.391235 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-889rr" podStartSLOduration=2.3911984889999998 podStartE2EDuration="2.391198489s" podCreationTimestamp="2026-03-07 14:59:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 14:59:23.38796919 +0000 UTC m=+1205.340105718" watchObservedRunningTime="2026-03-07 14:59:23.391198489 +0000 UTC m=+1205.343335037" Mar 07 14:59:25 crc kubenswrapper[4943]: I0307 14:59:25.380748 4943 generic.go:334] "Generic (PLEG): container finished" podID="870c9817-5966-4f86-82e0-b22084e7ce1d" containerID="3249b974d903bb8bc3583a79f36e4779b8523fc5199d38645e56bcc5bcbec1d1" exitCode=0 Mar 07 14:59:25 crc kubenswrapper[4943]: I0307 14:59:25.380883 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-889rr" event={"ID":"870c9817-5966-4f86-82e0-b22084e7ce1d","Type":"ContainerDied","Data":"3249b974d903bb8bc3583a79f36e4779b8523fc5199d38645e56bcc5bcbec1d1"} Mar 07 14:59:26 crc kubenswrapper[4943]: I0307 14:59:26.796902 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-889rr" Mar 07 14:59:26 crc kubenswrapper[4943]: I0307 14:59:26.849648 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-889rr"] Mar 07 14:59:26 crc kubenswrapper[4943]: I0307 14:59:26.850107 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-889rr"] Mar 07 14:59:26 crc kubenswrapper[4943]: I0307 14:59:26.911800 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/870c9817-5966-4f86-82e0-b22084e7ce1d-swiftconf\") pod \"870c9817-5966-4f86-82e0-b22084e7ce1d\" (UID: \"870c9817-5966-4f86-82e0-b22084e7ce1d\") " Mar 07 14:59:26 crc kubenswrapper[4943]: I0307 14:59:26.911873 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/870c9817-5966-4f86-82e0-b22084e7ce1d-ring-data-devices\") pod \"870c9817-5966-4f86-82e0-b22084e7ce1d\" (UID: \"870c9817-5966-4f86-82e0-b22084e7ce1d\") " Mar 07 14:59:26 crc kubenswrapper[4943]: I0307 14:59:26.912071 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/870c9817-5966-4f86-82e0-b22084e7ce1d-scripts\") pod \"870c9817-5966-4f86-82e0-b22084e7ce1d\" (UID: \"870c9817-5966-4f86-82e0-b22084e7ce1d\") " Mar 07 14:59:26 crc kubenswrapper[4943]: I0307 14:59:26.912710 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/870c9817-5966-4f86-82e0-b22084e7ce1d-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "870c9817-5966-4f86-82e0-b22084e7ce1d" (UID: "870c9817-5966-4f86-82e0-b22084e7ce1d"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:59:26 crc kubenswrapper[4943]: I0307 14:59:26.912787 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qnt7h\" (UniqueName: \"kubernetes.io/projected/870c9817-5966-4f86-82e0-b22084e7ce1d-kube-api-access-qnt7h\") pod \"870c9817-5966-4f86-82e0-b22084e7ce1d\" (UID: \"870c9817-5966-4f86-82e0-b22084e7ce1d\") " Mar 07 14:59:26 crc kubenswrapper[4943]: I0307 14:59:26.913077 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/870c9817-5966-4f86-82e0-b22084e7ce1d-dispersionconf\") pod \"870c9817-5966-4f86-82e0-b22084e7ce1d\" (UID: \"870c9817-5966-4f86-82e0-b22084e7ce1d\") " Mar 07 14:59:26 crc kubenswrapper[4943]: I0307 14:59:26.913166 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/870c9817-5966-4f86-82e0-b22084e7ce1d-etc-swift\") pod \"870c9817-5966-4f86-82e0-b22084e7ce1d\" (UID: \"870c9817-5966-4f86-82e0-b22084e7ce1d\") " Mar 07 14:59:26 crc kubenswrapper[4943]: I0307 14:59:26.913806 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/870c9817-5966-4f86-82e0-b22084e7ce1d-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 14:59:26 crc kubenswrapper[4943]: I0307 14:59:26.914622 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/870c9817-5966-4f86-82e0-b22084e7ce1d-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "870c9817-5966-4f86-82e0-b22084e7ce1d" (UID: "870c9817-5966-4f86-82e0-b22084e7ce1d"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 14:59:26 crc kubenswrapper[4943]: I0307 14:59:26.924294 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/870c9817-5966-4f86-82e0-b22084e7ce1d-kube-api-access-qnt7h" (OuterVolumeSpecName: "kube-api-access-qnt7h") pod "870c9817-5966-4f86-82e0-b22084e7ce1d" (UID: "870c9817-5966-4f86-82e0-b22084e7ce1d"). InnerVolumeSpecName "kube-api-access-qnt7h". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 14:59:26 crc kubenswrapper[4943]: I0307 14:59:26.938458 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/870c9817-5966-4f86-82e0-b22084e7ce1d-scripts" (OuterVolumeSpecName: "scripts") pod "870c9817-5966-4f86-82e0-b22084e7ce1d" (UID: "870c9817-5966-4f86-82e0-b22084e7ce1d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 14:59:26 crc kubenswrapper[4943]: I0307 14:59:26.944372 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/870c9817-5966-4f86-82e0-b22084e7ce1d-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "870c9817-5966-4f86-82e0-b22084e7ce1d" (UID: "870c9817-5966-4f86-82e0-b22084e7ce1d"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:59:26 crc kubenswrapper[4943]: I0307 14:59:26.953259 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/870c9817-5966-4f86-82e0-b22084e7ce1d-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "870c9817-5966-4f86-82e0-b22084e7ce1d" (UID: "870c9817-5966-4f86-82e0-b22084e7ce1d"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 14:59:27 crc kubenswrapper[4943]: I0307 14:59:27.017406 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/870c9817-5966-4f86-82e0-b22084e7ce1d-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 14:59:27 crc kubenswrapper[4943]: I0307 14:59:27.017474 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qnt7h\" (UniqueName: \"kubernetes.io/projected/870c9817-5966-4f86-82e0-b22084e7ce1d-kube-api-access-qnt7h\") on node \"crc\" DevicePath \"\"" Mar 07 14:59:27 crc kubenswrapper[4943]: I0307 14:59:27.017506 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/870c9817-5966-4f86-82e0-b22084e7ce1d-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 14:59:27 crc kubenswrapper[4943]: I0307 14:59:27.017534 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/870c9817-5966-4f86-82e0-b22084e7ce1d-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 14:59:27 crc kubenswrapper[4943]: I0307 14:59:27.017561 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/870c9817-5966-4f86-82e0-b22084e7ce1d-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 14:59:27 crc kubenswrapper[4943]: I0307 14:59:27.401531 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cf3965cb5887e19f1c24f8aeb22528dbb681e913193faf94c835846d597dc5f1" Mar 07 14:59:27 crc kubenswrapper[4943]: I0307 14:59:27.401629 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-889rr" Mar 07 14:59:28 crc kubenswrapper[4943]: I0307 14:59:28.768339 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="870c9817-5966-4f86-82e0-b22084e7ce1d" path="/var/lib/kubelet/pods/870c9817-5966-4f86-82e0-b22084e7ce1d/volumes" Mar 07 15:00:00 crc kubenswrapper[4943]: I0307 15:00:00.142379 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29548260-ljrss"] Mar 07 15:00:00 crc kubenswrapper[4943]: E0307 15:00:00.144786 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="870c9817-5966-4f86-82e0-b22084e7ce1d" containerName="swift-ring-rebalance" Mar 07 15:00:00 crc kubenswrapper[4943]: I0307 15:00:00.144926 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="870c9817-5966-4f86-82e0-b22084e7ce1d" containerName="swift-ring-rebalance" Mar 07 15:00:00 crc kubenswrapper[4943]: I0307 15:00:00.145317 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="870c9817-5966-4f86-82e0-b22084e7ce1d" containerName="swift-ring-rebalance" Mar 07 15:00:00 crc kubenswrapper[4943]: I0307 15:00:00.146218 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548260-ljrss" Mar 07 15:00:00 crc kubenswrapper[4943]: I0307 15:00:00.149493 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 07 15:00:00 crc kubenswrapper[4943]: I0307 15:00:00.149536 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29548260-2c2jw"] Mar 07 15:00:00 crc kubenswrapper[4943]: I0307 15:00:00.149535 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-qqlk5" Mar 07 15:00:00 crc kubenswrapper[4943]: I0307 15:00:00.150287 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 07 15:00:00 crc kubenswrapper[4943]: I0307 15:00:00.151109 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29548260-2c2jw" Mar 07 15:00:00 crc kubenswrapper[4943]: I0307 15:00:00.154774 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 07 15:00:00 crc kubenswrapper[4943]: I0307 15:00:00.155916 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 07 15:00:00 crc kubenswrapper[4943]: I0307 15:00:00.161228 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29548260-ljrss"] Mar 07 15:00:00 crc kubenswrapper[4943]: I0307 15:00:00.167952 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29548260-2c2jw"] Mar 07 15:00:00 crc kubenswrapper[4943]: I0307 15:00:00.277673 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1cf0ea1a-8473-4c48-b3d3-796d4c6d6de5-secret-volume\") pod \"collect-profiles-29548260-2c2jw\" (UID: \"1cf0ea1a-8473-4c48-b3d3-796d4c6d6de5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29548260-2c2jw" Mar 07 15:00:00 crc kubenswrapper[4943]: I0307 15:00:00.277889 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qr77k\" (UniqueName: \"kubernetes.io/projected/601960fc-b3bc-4a62-9843-fd22ba6a1850-kube-api-access-qr77k\") pod \"auto-csr-approver-29548260-ljrss\" (UID: \"601960fc-b3bc-4a62-9843-fd22ba6a1850\") " pod="openshift-infra/auto-csr-approver-29548260-ljrss" Mar 07 15:00:00 crc kubenswrapper[4943]: I0307 15:00:00.278066 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4tvzd\" (UniqueName: \"kubernetes.io/projected/1cf0ea1a-8473-4c48-b3d3-796d4c6d6de5-kube-api-access-4tvzd\") pod \"collect-profiles-29548260-2c2jw\" (UID: \"1cf0ea1a-8473-4c48-b3d3-796d4c6d6de5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29548260-2c2jw" Mar 07 15:00:00 crc kubenswrapper[4943]: I0307 15:00:00.278169 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1cf0ea1a-8473-4c48-b3d3-796d4c6d6de5-config-volume\") pod \"collect-profiles-29548260-2c2jw\" (UID: \"1cf0ea1a-8473-4c48-b3d3-796d4c6d6de5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29548260-2c2jw" Mar 07 15:00:00 crc kubenswrapper[4943]: I0307 15:00:00.379352 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1cf0ea1a-8473-4c48-b3d3-796d4c6d6de5-config-volume\") pod \"collect-profiles-29548260-2c2jw\" (UID: \"1cf0ea1a-8473-4c48-b3d3-796d4c6d6de5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29548260-2c2jw" Mar 07 15:00:00 crc kubenswrapper[4943]: I0307 15:00:00.379472 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1cf0ea1a-8473-4c48-b3d3-796d4c6d6de5-secret-volume\") pod \"collect-profiles-29548260-2c2jw\" (UID: \"1cf0ea1a-8473-4c48-b3d3-796d4c6d6de5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29548260-2c2jw" Mar 07 15:00:00 crc kubenswrapper[4943]: I0307 15:00:00.379517 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qr77k\" (UniqueName: \"kubernetes.io/projected/601960fc-b3bc-4a62-9843-fd22ba6a1850-kube-api-access-qr77k\") pod \"auto-csr-approver-29548260-ljrss\" (UID: \"601960fc-b3bc-4a62-9843-fd22ba6a1850\") " pod="openshift-infra/auto-csr-approver-29548260-ljrss" Mar 07 15:00:00 crc kubenswrapper[4943]: I0307 15:00:00.379561 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4tvzd\" (UniqueName: \"kubernetes.io/projected/1cf0ea1a-8473-4c48-b3d3-796d4c6d6de5-kube-api-access-4tvzd\") pod \"collect-profiles-29548260-2c2jw\" (UID: \"1cf0ea1a-8473-4c48-b3d3-796d4c6d6de5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29548260-2c2jw" Mar 07 15:00:00 crc kubenswrapper[4943]: I0307 15:00:00.380871 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1cf0ea1a-8473-4c48-b3d3-796d4c6d6de5-config-volume\") pod \"collect-profiles-29548260-2c2jw\" (UID: \"1cf0ea1a-8473-4c48-b3d3-796d4c6d6de5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29548260-2c2jw" Mar 07 15:00:00 crc kubenswrapper[4943]: I0307 15:00:00.391440 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1cf0ea1a-8473-4c48-b3d3-796d4c6d6de5-secret-volume\") pod \"collect-profiles-29548260-2c2jw\" (UID: \"1cf0ea1a-8473-4c48-b3d3-796d4c6d6de5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29548260-2c2jw" Mar 07 15:00:00 crc kubenswrapper[4943]: I0307 15:00:00.412300 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qr77k\" (UniqueName: \"kubernetes.io/projected/601960fc-b3bc-4a62-9843-fd22ba6a1850-kube-api-access-qr77k\") pod \"auto-csr-approver-29548260-ljrss\" (UID: \"601960fc-b3bc-4a62-9843-fd22ba6a1850\") " pod="openshift-infra/auto-csr-approver-29548260-ljrss" Mar 07 15:00:00 crc kubenswrapper[4943]: I0307 15:00:00.413717 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4tvzd\" (UniqueName: \"kubernetes.io/projected/1cf0ea1a-8473-4c48-b3d3-796d4c6d6de5-kube-api-access-4tvzd\") pod \"collect-profiles-29548260-2c2jw\" (UID: \"1cf0ea1a-8473-4c48-b3d3-796d4c6d6de5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29548260-2c2jw" Mar 07 15:00:00 crc kubenswrapper[4943]: I0307 15:00:00.482417 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548260-ljrss" Mar 07 15:00:00 crc kubenswrapper[4943]: I0307 15:00:00.503706 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29548260-2c2jw" Mar 07 15:00:00 crc kubenswrapper[4943]: I0307 15:00:00.826499 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29548260-2c2jw"] Mar 07 15:00:00 crc kubenswrapper[4943]: I0307 15:00:00.961224 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29548260-ljrss"] Mar 07 15:00:00 crc kubenswrapper[4943]: I0307 15:00:00.983179 4943 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 07 15:00:01 crc kubenswrapper[4943]: I0307 15:00:01.723985 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548260-ljrss" event={"ID":"601960fc-b3bc-4a62-9843-fd22ba6a1850","Type":"ContainerStarted","Data":"325eceb155a0687c1e4430e9a97de5267730e13904fc6c16a071bceb3afb08fb"} Mar 07 15:00:01 crc kubenswrapper[4943]: I0307 15:00:01.728355 4943 generic.go:334] "Generic (PLEG): container finished" podID="1cf0ea1a-8473-4c48-b3d3-796d4c6d6de5" containerID="9e16c555a9bac90433d1f8250f3e5fc922f180a770a4f148e9cdd10cd37d9b4c" exitCode=0 Mar 07 15:00:01 crc kubenswrapper[4943]: I0307 15:00:01.728402 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29548260-2c2jw" event={"ID":"1cf0ea1a-8473-4c48-b3d3-796d4c6d6de5","Type":"ContainerDied","Data":"9e16c555a9bac90433d1f8250f3e5fc922f180a770a4f148e9cdd10cd37d9b4c"} Mar 07 15:00:01 crc kubenswrapper[4943]: I0307 15:00:01.728429 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29548260-2c2jw" event={"ID":"1cf0ea1a-8473-4c48-b3d3-796d4c6d6de5","Type":"ContainerStarted","Data":"4d92b1c93c7a691b486d5274a24ab8cdd201e08cd7e72aeb4d73b31223721203"} Mar 07 15:00:03 crc kubenswrapper[4943]: I0307 15:00:03.066213 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29548260-2c2jw" Mar 07 15:00:03 crc kubenswrapper[4943]: I0307 15:00:03.254513 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1cf0ea1a-8473-4c48-b3d3-796d4c6d6de5-config-volume\") pod \"1cf0ea1a-8473-4c48-b3d3-796d4c6d6de5\" (UID: \"1cf0ea1a-8473-4c48-b3d3-796d4c6d6de5\") " Mar 07 15:00:03 crc kubenswrapper[4943]: I0307 15:00:03.255065 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4tvzd\" (UniqueName: \"kubernetes.io/projected/1cf0ea1a-8473-4c48-b3d3-796d4c6d6de5-kube-api-access-4tvzd\") pod \"1cf0ea1a-8473-4c48-b3d3-796d4c6d6de5\" (UID: \"1cf0ea1a-8473-4c48-b3d3-796d4c6d6de5\") " Mar 07 15:00:03 crc kubenswrapper[4943]: I0307 15:00:03.255156 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1cf0ea1a-8473-4c48-b3d3-796d4c6d6de5-secret-volume\") pod \"1cf0ea1a-8473-4c48-b3d3-796d4c6d6de5\" (UID: \"1cf0ea1a-8473-4c48-b3d3-796d4c6d6de5\") " Mar 07 15:00:03 crc kubenswrapper[4943]: I0307 15:00:03.255979 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1cf0ea1a-8473-4c48-b3d3-796d4c6d6de5-config-volume" (OuterVolumeSpecName: "config-volume") pod "1cf0ea1a-8473-4c48-b3d3-796d4c6d6de5" (UID: "1cf0ea1a-8473-4c48-b3d3-796d4c6d6de5"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:00:03 crc kubenswrapper[4943]: I0307 15:00:03.266106 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1cf0ea1a-8473-4c48-b3d3-796d4c6d6de5-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1cf0ea1a-8473-4c48-b3d3-796d4c6d6de5" (UID: "1cf0ea1a-8473-4c48-b3d3-796d4c6d6de5"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:00:03 crc kubenswrapper[4943]: I0307 15:00:03.266233 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1cf0ea1a-8473-4c48-b3d3-796d4c6d6de5-kube-api-access-4tvzd" (OuterVolumeSpecName: "kube-api-access-4tvzd") pod "1cf0ea1a-8473-4c48-b3d3-796d4c6d6de5" (UID: "1cf0ea1a-8473-4c48-b3d3-796d4c6d6de5"). InnerVolumeSpecName "kube-api-access-4tvzd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:00:03 crc kubenswrapper[4943]: I0307 15:00:03.357159 4943 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1cf0ea1a-8473-4c48-b3d3-796d4c6d6de5-config-volume\") on node \"crc\" DevicePath \"\"" Mar 07 15:00:03 crc kubenswrapper[4943]: I0307 15:00:03.357238 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4tvzd\" (UniqueName: \"kubernetes.io/projected/1cf0ea1a-8473-4c48-b3d3-796d4c6d6de5-kube-api-access-4tvzd\") on node \"crc\" DevicePath \"\"" Mar 07 15:00:03 crc kubenswrapper[4943]: I0307 15:00:03.357260 4943 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1cf0ea1a-8473-4c48-b3d3-796d4c6d6de5-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 07 15:00:03 crc kubenswrapper[4943]: I0307 15:00:03.749510 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29548260-2c2jw" event={"ID":"1cf0ea1a-8473-4c48-b3d3-796d4c6d6de5","Type":"ContainerDied","Data":"4d92b1c93c7a691b486d5274a24ab8cdd201e08cd7e72aeb4d73b31223721203"} Mar 07 15:00:03 crc kubenswrapper[4943]: I0307 15:00:03.749549 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4d92b1c93c7a691b486d5274a24ab8cdd201e08cd7e72aeb4d73b31223721203" Mar 07 15:00:03 crc kubenswrapper[4943]: I0307 15:00:03.749641 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29548260-2c2jw" Mar 07 15:00:04 crc kubenswrapper[4943]: I0307 15:00:04.762871 4943 generic.go:334] "Generic (PLEG): container finished" podID="601960fc-b3bc-4a62-9843-fd22ba6a1850" containerID="e10b1432f5f9b11449cf70b1dcf772f9002fc6d36fe8ecbcfc87b27610e44016" exitCode=0 Mar 07 15:00:04 crc kubenswrapper[4943]: I0307 15:00:04.770332 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548260-ljrss" event={"ID":"601960fc-b3bc-4a62-9843-fd22ba6a1850","Type":"ContainerDied","Data":"e10b1432f5f9b11449cf70b1dcf772f9002fc6d36fe8ecbcfc87b27610e44016"} Mar 07 15:00:06 crc kubenswrapper[4943]: I0307 15:00:06.127582 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548260-ljrss" Mar 07 15:00:06 crc kubenswrapper[4943]: I0307 15:00:06.304455 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qr77k\" (UniqueName: \"kubernetes.io/projected/601960fc-b3bc-4a62-9843-fd22ba6a1850-kube-api-access-qr77k\") pod \"601960fc-b3bc-4a62-9843-fd22ba6a1850\" (UID: \"601960fc-b3bc-4a62-9843-fd22ba6a1850\") " Mar 07 15:00:06 crc kubenswrapper[4943]: I0307 15:00:06.312326 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/601960fc-b3bc-4a62-9843-fd22ba6a1850-kube-api-access-qr77k" (OuterVolumeSpecName: "kube-api-access-qr77k") pod "601960fc-b3bc-4a62-9843-fd22ba6a1850" (UID: "601960fc-b3bc-4a62-9843-fd22ba6a1850"). InnerVolumeSpecName "kube-api-access-qr77k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:00:06 crc kubenswrapper[4943]: I0307 15:00:06.406026 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qr77k\" (UniqueName: \"kubernetes.io/projected/601960fc-b3bc-4a62-9843-fd22ba6a1850-kube-api-access-qr77k\") on node \"crc\" DevicePath \"\"" Mar 07 15:00:06 crc kubenswrapper[4943]: I0307 15:00:06.787465 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548260-ljrss" event={"ID":"601960fc-b3bc-4a62-9843-fd22ba6a1850","Type":"ContainerDied","Data":"325eceb155a0687c1e4430e9a97de5267730e13904fc6c16a071bceb3afb08fb"} Mar 07 15:00:06 crc kubenswrapper[4943]: I0307 15:00:06.787823 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="325eceb155a0687c1e4430e9a97de5267730e13904fc6c16a071bceb3afb08fb" Mar 07 15:00:06 crc kubenswrapper[4943]: I0307 15:00:06.787587 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548260-ljrss" Mar 07 15:00:07 crc kubenswrapper[4943]: I0307 15:00:07.207739 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29548254-zpwrq"] Mar 07 15:00:07 crc kubenswrapper[4943]: I0307 15:00:07.214274 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29548254-zpwrq"] Mar 07 15:00:08 crc kubenswrapper[4943]: I0307 15:00:08.772840 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81098c4a-ecc7-4fff-a29f-f31365f42a62" path="/var/lib/kubelet/pods/81098c4a-ecc7-4fff-a29f-f31365f42a62/volumes" Mar 07 15:00:20 crc kubenswrapper[4943]: I0307 15:00:20.018579 4943 scope.go:117] "RemoveContainer" containerID="09ac5748f6e67412ea6c6600ed76341524eb969dba0d55c3ffac79f034e4588a" Mar 07 15:01:00 crc kubenswrapper[4943]: I0307 15:01:00.161561 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/keystone-cron-29548261-jgm2m"] Mar 07 15:01:00 crc kubenswrapper[4943]: E0307 15:01:00.183313 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="601960fc-b3bc-4a62-9843-fd22ba6a1850" containerName="oc" Mar 07 15:01:00 crc kubenswrapper[4943]: I0307 15:01:00.183335 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="601960fc-b3bc-4a62-9843-fd22ba6a1850" containerName="oc" Mar 07 15:01:00 crc kubenswrapper[4943]: E0307 15:01:00.183366 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1cf0ea1a-8473-4c48-b3d3-796d4c6d6de5" containerName="collect-profiles" Mar 07 15:01:00 crc kubenswrapper[4943]: I0307 15:01:00.183374 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="1cf0ea1a-8473-4c48-b3d3-796d4c6d6de5" containerName="collect-profiles" Mar 07 15:01:00 crc kubenswrapper[4943]: I0307 15:01:00.183521 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="1cf0ea1a-8473-4c48-b3d3-796d4c6d6de5" containerName="collect-profiles" Mar 07 15:01:00 crc kubenswrapper[4943]: I0307 15:01:00.183539 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="601960fc-b3bc-4a62-9843-fd22ba6a1850" containerName="oc" Mar 07 15:01:00 crc kubenswrapper[4943]: I0307 15:01:00.183996 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/keystone-cron-29548261-jgm2m"] Mar 07 15:01:00 crc kubenswrapper[4943]: I0307 15:01:00.184079 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/keystone-cron-29548261-jgm2m" Mar 07 15:01:00 crc kubenswrapper[4943]: I0307 15:01:00.270966 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d1b17bd5-bcfb-49b7-9a5a-9d718d864740-fernet-keys\") pod \"keystone-cron-29548261-jgm2m\" (UID: \"d1b17bd5-bcfb-49b7-9a5a-9d718d864740\") " pod="swift-kuttl-tests/keystone-cron-29548261-jgm2m" Mar 07 15:01:00 crc kubenswrapper[4943]: I0307 15:01:00.271029 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zr6lz\" (UniqueName: \"kubernetes.io/projected/d1b17bd5-bcfb-49b7-9a5a-9d718d864740-kube-api-access-zr6lz\") pod \"keystone-cron-29548261-jgm2m\" (UID: \"d1b17bd5-bcfb-49b7-9a5a-9d718d864740\") " pod="swift-kuttl-tests/keystone-cron-29548261-jgm2m" Mar 07 15:01:00 crc kubenswrapper[4943]: I0307 15:01:00.271070 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1b17bd5-bcfb-49b7-9a5a-9d718d864740-config-data\") pod \"keystone-cron-29548261-jgm2m\" (UID: \"d1b17bd5-bcfb-49b7-9a5a-9d718d864740\") " pod="swift-kuttl-tests/keystone-cron-29548261-jgm2m" Mar 07 15:01:00 crc kubenswrapper[4943]: I0307 15:01:00.372508 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1b17bd5-bcfb-49b7-9a5a-9d718d864740-config-data\") pod \"keystone-cron-29548261-jgm2m\" (UID: \"d1b17bd5-bcfb-49b7-9a5a-9d718d864740\") " pod="swift-kuttl-tests/keystone-cron-29548261-jgm2m" Mar 07 15:01:00 crc kubenswrapper[4943]: I0307 15:01:00.372609 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d1b17bd5-bcfb-49b7-9a5a-9d718d864740-fernet-keys\") pod \"keystone-cron-29548261-jgm2m\" (UID: \"d1b17bd5-bcfb-49b7-9a5a-9d718d864740\") " pod="swift-kuttl-tests/keystone-cron-29548261-jgm2m" Mar 07 15:01:00 crc kubenswrapper[4943]: I0307 15:01:00.372641 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zr6lz\" (UniqueName: \"kubernetes.io/projected/d1b17bd5-bcfb-49b7-9a5a-9d718d864740-kube-api-access-zr6lz\") pod \"keystone-cron-29548261-jgm2m\" (UID: \"d1b17bd5-bcfb-49b7-9a5a-9d718d864740\") " pod="swift-kuttl-tests/keystone-cron-29548261-jgm2m" Mar 07 15:01:00 crc kubenswrapper[4943]: I0307 15:01:00.398641 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zr6lz\" (UniqueName: \"kubernetes.io/projected/d1b17bd5-bcfb-49b7-9a5a-9d718d864740-kube-api-access-zr6lz\") pod \"keystone-cron-29548261-jgm2m\" (UID: \"d1b17bd5-bcfb-49b7-9a5a-9d718d864740\") " pod="swift-kuttl-tests/keystone-cron-29548261-jgm2m" Mar 07 15:01:00 crc kubenswrapper[4943]: I0307 15:01:00.402118 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d1b17bd5-bcfb-49b7-9a5a-9d718d864740-fernet-keys\") pod \"keystone-cron-29548261-jgm2m\" (UID: \"d1b17bd5-bcfb-49b7-9a5a-9d718d864740\") " pod="swift-kuttl-tests/keystone-cron-29548261-jgm2m" Mar 07 15:01:00 crc kubenswrapper[4943]: I0307 15:01:00.431706 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1b17bd5-bcfb-49b7-9a5a-9d718d864740-config-data\") pod \"keystone-cron-29548261-jgm2m\" (UID: \"d1b17bd5-bcfb-49b7-9a5a-9d718d864740\") " pod="swift-kuttl-tests/keystone-cron-29548261-jgm2m" Mar 07 15:01:00 crc kubenswrapper[4943]: I0307 15:01:00.527180 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/keystone-cron-29548261-jgm2m" Mar 07 15:01:00 crc kubenswrapper[4943]: I0307 15:01:00.989071 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/keystone-cron-29548261-jgm2m"] Mar 07 15:01:00 crc kubenswrapper[4943]: W0307 15:01:00.997827 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd1b17bd5_bcfb_49b7_9a5a_9d718d864740.slice/crio-d34b7b9ecc3017df3d9c7fcaa72d201f68651a9053673c3b3d9a55a6ab567873 WatchSource:0}: Error finding container d34b7b9ecc3017df3d9c7fcaa72d201f68651a9053673c3b3d9a55a6ab567873: Status 404 returned error can't find the container with id d34b7b9ecc3017df3d9c7fcaa72d201f68651a9053673c3b3d9a55a6ab567873 Mar 07 15:01:01 crc kubenswrapper[4943]: I0307 15:01:01.370883 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/keystone-cron-29548261-jgm2m" event={"ID":"d1b17bd5-bcfb-49b7-9a5a-9d718d864740","Type":"ContainerStarted","Data":"99ef3eccca56493debd297b4512c7a73510ea06b89a8fc2a2b4832df6d55daae"} Mar 07 15:01:01 crc kubenswrapper[4943]: I0307 15:01:01.372871 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/keystone-cron-29548261-jgm2m" event={"ID":"d1b17bd5-bcfb-49b7-9a5a-9d718d864740","Type":"ContainerStarted","Data":"d34b7b9ecc3017df3d9c7fcaa72d201f68651a9053673c3b3d9a55a6ab567873"} Mar 07 15:01:01 crc kubenswrapper[4943]: I0307 15:01:01.391531 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/keystone-cron-29548261-jgm2m" podStartSLOduration=1.391516612 podStartE2EDuration="1.391516612s" podCreationTimestamp="2026-03-07 15:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:01:01.389649206 +0000 UTC m=+1303.341785704" watchObservedRunningTime="2026-03-07 15:01:01.391516612 +0000 UTC m=+1303.343653110" Mar 07 15:01:03 crc kubenswrapper[4943]: I0307 15:01:03.394351 4943 generic.go:334] "Generic (PLEG): container finished" podID="d1b17bd5-bcfb-49b7-9a5a-9d718d864740" containerID="99ef3eccca56493debd297b4512c7a73510ea06b89a8fc2a2b4832df6d55daae" exitCode=0 Mar 07 15:01:03 crc kubenswrapper[4943]: I0307 15:01:03.394636 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/keystone-cron-29548261-jgm2m" event={"ID":"d1b17bd5-bcfb-49b7-9a5a-9d718d864740","Type":"ContainerDied","Data":"99ef3eccca56493debd297b4512c7a73510ea06b89a8fc2a2b4832df6d55daae"} Mar 07 15:01:04 crc kubenswrapper[4943]: I0307 15:01:04.799829 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/keystone-cron-29548261-jgm2m" Mar 07 15:01:04 crc kubenswrapper[4943]: I0307 15:01:04.842155 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d1b17bd5-bcfb-49b7-9a5a-9d718d864740-fernet-keys\") pod \"d1b17bd5-bcfb-49b7-9a5a-9d718d864740\" (UID: \"d1b17bd5-bcfb-49b7-9a5a-9d718d864740\") " Mar 07 15:01:04 crc kubenswrapper[4943]: I0307 15:01:04.842248 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1b17bd5-bcfb-49b7-9a5a-9d718d864740-config-data\") pod \"d1b17bd5-bcfb-49b7-9a5a-9d718d864740\" (UID: \"d1b17bd5-bcfb-49b7-9a5a-9d718d864740\") " Mar 07 15:01:04 crc kubenswrapper[4943]: I0307 15:01:04.842333 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zr6lz\" (UniqueName: \"kubernetes.io/projected/d1b17bd5-bcfb-49b7-9a5a-9d718d864740-kube-api-access-zr6lz\") pod \"d1b17bd5-bcfb-49b7-9a5a-9d718d864740\" (UID: \"d1b17bd5-bcfb-49b7-9a5a-9d718d864740\") " Mar 07 15:01:04 crc kubenswrapper[4943]: I0307 15:01:04.848697 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1b17bd5-bcfb-49b7-9a5a-9d718d864740-kube-api-access-zr6lz" (OuterVolumeSpecName: "kube-api-access-zr6lz") pod "d1b17bd5-bcfb-49b7-9a5a-9d718d864740" (UID: "d1b17bd5-bcfb-49b7-9a5a-9d718d864740"). InnerVolumeSpecName "kube-api-access-zr6lz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:01:04 crc kubenswrapper[4943]: I0307 15:01:04.850192 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1b17bd5-bcfb-49b7-9a5a-9d718d864740-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "d1b17bd5-bcfb-49b7-9a5a-9d718d864740" (UID: "d1b17bd5-bcfb-49b7-9a5a-9d718d864740"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:01:04 crc kubenswrapper[4943]: I0307 15:01:04.907453 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1b17bd5-bcfb-49b7-9a5a-9d718d864740-config-data" (OuterVolumeSpecName: "config-data") pod "d1b17bd5-bcfb-49b7-9a5a-9d718d864740" (UID: "d1b17bd5-bcfb-49b7-9a5a-9d718d864740"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:01:04 crc kubenswrapper[4943]: I0307 15:01:04.943787 4943 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d1b17bd5-bcfb-49b7-9a5a-9d718d864740-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 07 15:01:04 crc kubenswrapper[4943]: I0307 15:01:04.943822 4943 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1b17bd5-bcfb-49b7-9a5a-9d718d864740-config-data\") on node \"crc\" DevicePath \"\"" Mar 07 15:01:04 crc kubenswrapper[4943]: I0307 15:01:04.943832 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zr6lz\" (UniqueName: \"kubernetes.io/projected/d1b17bd5-bcfb-49b7-9a5a-9d718d864740-kube-api-access-zr6lz\") on node \"crc\" DevicePath \"\"" Mar 07 15:01:05 crc kubenswrapper[4943]: I0307 15:01:05.427683 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/keystone-cron-29548261-jgm2m" event={"ID":"d1b17bd5-bcfb-49b7-9a5a-9d718d864740","Type":"ContainerDied","Data":"d34b7b9ecc3017df3d9c7fcaa72d201f68651a9053673c3b3d9a55a6ab567873"} Mar 07 15:01:05 crc kubenswrapper[4943]: I0307 15:01:05.427723 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d34b7b9ecc3017df3d9c7fcaa72d201f68651a9053673c3b3d9a55a6ab567873" Mar 07 15:01:05 crc kubenswrapper[4943]: I0307 15:01:05.427726 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/keystone-cron-29548261-jgm2m" Mar 07 15:01:06 crc kubenswrapper[4943]: I0307 15:01:06.074179 4943 patch_prober.go:28] interesting pod/machine-config-daemon-cpdmm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 07 15:01:06 crc kubenswrapper[4943]: I0307 15:01:06.074265 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 07 15:01:36 crc kubenswrapper[4943]: I0307 15:01:36.074357 4943 patch_prober.go:28] interesting pod/machine-config-daemon-cpdmm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 07 15:01:36 crc kubenswrapper[4943]: I0307 15:01:36.075121 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 07 15:02:00 crc kubenswrapper[4943]: I0307 15:02:00.143287 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29548262-k2c8b"] Mar 07 15:02:00 crc kubenswrapper[4943]: E0307 15:02:00.143993 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1b17bd5-bcfb-49b7-9a5a-9d718d864740" containerName="keystone-cron" Mar 07 15:02:00 crc kubenswrapper[4943]: I0307 15:02:00.144005 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1b17bd5-bcfb-49b7-9a5a-9d718d864740" containerName="keystone-cron" Mar 07 15:02:00 crc kubenswrapper[4943]: I0307 15:02:00.144136 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1b17bd5-bcfb-49b7-9a5a-9d718d864740" containerName="keystone-cron" Mar 07 15:02:00 crc kubenswrapper[4943]: I0307 15:02:00.144557 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548262-k2c8b" Mar 07 15:02:00 crc kubenswrapper[4943]: I0307 15:02:00.147587 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-qqlk5" Mar 07 15:02:00 crc kubenswrapper[4943]: I0307 15:02:00.147653 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 07 15:02:00 crc kubenswrapper[4943]: I0307 15:02:00.147702 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 07 15:02:00 crc kubenswrapper[4943]: I0307 15:02:00.163209 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29548262-k2c8b"] Mar 07 15:02:00 crc kubenswrapper[4943]: I0307 15:02:00.241760 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqll6\" (UniqueName: \"kubernetes.io/projected/feed891c-4b59-472b-9cce-89d1c90652a5-kube-api-access-sqll6\") pod \"auto-csr-approver-29548262-k2c8b\" (UID: \"feed891c-4b59-472b-9cce-89d1c90652a5\") " pod="openshift-infra/auto-csr-approver-29548262-k2c8b" Mar 07 15:02:00 crc kubenswrapper[4943]: I0307 15:02:00.343648 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqll6\" (UniqueName: \"kubernetes.io/projected/feed891c-4b59-472b-9cce-89d1c90652a5-kube-api-access-sqll6\") pod \"auto-csr-approver-29548262-k2c8b\" (UID: \"feed891c-4b59-472b-9cce-89d1c90652a5\") " pod="openshift-infra/auto-csr-approver-29548262-k2c8b" Mar 07 15:02:00 crc kubenswrapper[4943]: I0307 15:02:00.368218 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqll6\" (UniqueName: \"kubernetes.io/projected/feed891c-4b59-472b-9cce-89d1c90652a5-kube-api-access-sqll6\") pod \"auto-csr-approver-29548262-k2c8b\" (UID: \"feed891c-4b59-472b-9cce-89d1c90652a5\") " pod="openshift-infra/auto-csr-approver-29548262-k2c8b" Mar 07 15:02:00 crc kubenswrapper[4943]: I0307 15:02:00.501515 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548262-k2c8b" Mar 07 15:02:00 crc kubenswrapper[4943]: W0307 15:02:00.872222 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfeed891c_4b59_472b_9cce_89d1c90652a5.slice/crio-42ccc5fe54c7f21c106d10871ec0662682f65e39be4ca64bbc9de54a0d517d74 WatchSource:0}: Error finding container 42ccc5fe54c7f21c106d10871ec0662682f65e39be4ca64bbc9de54a0d517d74: Status 404 returned error can't find the container with id 42ccc5fe54c7f21c106d10871ec0662682f65e39be4ca64bbc9de54a0d517d74 Mar 07 15:02:00 crc kubenswrapper[4943]: I0307 15:02:00.873446 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29548262-k2c8b"] Mar 07 15:02:00 crc kubenswrapper[4943]: I0307 15:02:00.973433 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548262-k2c8b" event={"ID":"feed891c-4b59-472b-9cce-89d1c90652a5","Type":"ContainerStarted","Data":"42ccc5fe54c7f21c106d10871ec0662682f65e39be4ca64bbc9de54a0d517d74"} Mar 07 15:02:02 crc kubenswrapper[4943]: I0307 15:02:02.994739 4943 generic.go:334] "Generic (PLEG): container finished" podID="feed891c-4b59-472b-9cce-89d1c90652a5" containerID="d445bd4612bc3ec557b286d279d0eea9cd49a0656c2b26d29b28f1675b62d5bb" exitCode=0 Mar 07 15:02:02 crc kubenswrapper[4943]: I0307 15:02:02.994817 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548262-k2c8b" event={"ID":"feed891c-4b59-472b-9cce-89d1c90652a5","Type":"ContainerDied","Data":"d445bd4612bc3ec557b286d279d0eea9cd49a0656c2b26d29b28f1675b62d5bb"} Mar 07 15:02:04 crc kubenswrapper[4943]: I0307 15:02:04.338575 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548262-k2c8b" Mar 07 15:02:04 crc kubenswrapper[4943]: I0307 15:02:04.412999 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sqll6\" (UniqueName: \"kubernetes.io/projected/feed891c-4b59-472b-9cce-89d1c90652a5-kube-api-access-sqll6\") pod \"feed891c-4b59-472b-9cce-89d1c90652a5\" (UID: \"feed891c-4b59-472b-9cce-89d1c90652a5\") " Mar 07 15:02:04 crc kubenswrapper[4943]: I0307 15:02:04.423326 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/feed891c-4b59-472b-9cce-89d1c90652a5-kube-api-access-sqll6" (OuterVolumeSpecName: "kube-api-access-sqll6") pod "feed891c-4b59-472b-9cce-89d1c90652a5" (UID: "feed891c-4b59-472b-9cce-89d1c90652a5"). InnerVolumeSpecName "kube-api-access-sqll6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:02:04 crc kubenswrapper[4943]: I0307 15:02:04.514662 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sqll6\" (UniqueName: \"kubernetes.io/projected/feed891c-4b59-472b-9cce-89d1c90652a5-kube-api-access-sqll6\") on node \"crc\" DevicePath \"\"" Mar 07 15:02:05 crc kubenswrapper[4943]: I0307 15:02:05.015733 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548262-k2c8b" event={"ID":"feed891c-4b59-472b-9cce-89d1c90652a5","Type":"ContainerDied","Data":"42ccc5fe54c7f21c106d10871ec0662682f65e39be4ca64bbc9de54a0d517d74"} Mar 07 15:02:05 crc kubenswrapper[4943]: I0307 15:02:05.015794 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="42ccc5fe54c7f21c106d10871ec0662682f65e39be4ca64bbc9de54a0d517d74" Mar 07 15:02:05 crc kubenswrapper[4943]: I0307 15:02:05.016342 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548262-k2c8b" Mar 07 15:02:05 crc kubenswrapper[4943]: I0307 15:02:05.431701 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29548256-q9db4"] Mar 07 15:02:05 crc kubenswrapper[4943]: I0307 15:02:05.441600 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29548256-q9db4"] Mar 07 15:02:06 crc kubenswrapper[4943]: I0307 15:02:06.075127 4943 patch_prober.go:28] interesting pod/machine-config-daemon-cpdmm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 07 15:02:06 crc kubenswrapper[4943]: I0307 15:02:06.075248 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 07 15:02:06 crc kubenswrapper[4943]: I0307 15:02:06.075617 4943 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" Mar 07 15:02:06 crc kubenswrapper[4943]: I0307 15:02:06.076917 4943 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"526f9b1306d1082bdb6e6f98c1f04669533c0a9945daeb9cde6060502da5cecb"} pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 07 15:02:06 crc kubenswrapper[4943]: I0307 15:02:06.077107 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerName="machine-config-daemon" containerID="cri-o://526f9b1306d1082bdb6e6f98c1f04669533c0a9945daeb9cde6060502da5cecb" gracePeriod=600 Mar 07 15:02:06 crc kubenswrapper[4943]: I0307 15:02:06.768822 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a5f15db-2baf-441d-b3c7-d93e0779d268" path="/var/lib/kubelet/pods/8a5f15db-2baf-441d-b3c7-d93e0779d268/volumes" Mar 07 15:02:07 crc kubenswrapper[4943]: I0307 15:02:07.042167 4943 generic.go:334] "Generic (PLEG): container finished" podID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerID="526f9b1306d1082bdb6e6f98c1f04669533c0a9945daeb9cde6060502da5cecb" exitCode=0 Mar 07 15:02:07 crc kubenswrapper[4943]: I0307 15:02:07.042243 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" event={"ID":"4c1f18ea-e557-41fc-95dd-ce4e1722a61e","Type":"ContainerDied","Data":"526f9b1306d1082bdb6e6f98c1f04669533c0a9945daeb9cde6060502da5cecb"} Mar 07 15:02:07 crc kubenswrapper[4943]: I0307 15:02:07.042313 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" event={"ID":"4c1f18ea-e557-41fc-95dd-ce4e1722a61e","Type":"ContainerStarted","Data":"26f35e1f941cca0c34e3650d7aa047834e769dd76d2a61c3a39cad5bb2d090ae"} Mar 07 15:02:07 crc kubenswrapper[4943]: I0307 15:02:07.042337 4943 scope.go:117] "RemoveContainer" containerID="808321a0c8648ed7264aba7412d4317b4e5e5302eaf1e96b5220f23b29527cec" Mar 07 15:02:20 crc kubenswrapper[4943]: I0307 15:02:20.110826 4943 scope.go:117] "RemoveContainer" containerID="950a2ae469245e87ccf033b6bcaab52a5596412a5f98a2ad425c32993ae4aec0" Mar 07 15:02:54 crc kubenswrapper[4943]: E0307 15:02:54.223977 4943 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.75:45786->38.102.83.75:36531: write tcp 38.102.83.75:45786->38.102.83.75:36531: write: broken pipe Mar 07 15:03:00 crc kubenswrapper[4943]: E0307 15:03:00.587576 4943 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.75:56804->38.102.83.75:36531: write tcp 38.102.83.75:56804->38.102.83.75:36531: write: connection reset by peer Mar 07 15:03:20 crc kubenswrapper[4943]: I0307 15:03:20.162223 4943 scope.go:117] "RemoveContainer" containerID="462f33b8284a4b8f9e8effbfe35a02663da8eb314ff89b451e7780bf15a3a839" Mar 07 15:04:00 crc kubenswrapper[4943]: I0307 15:04:00.142333 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29548264-qrpfx"] Mar 07 15:04:00 crc kubenswrapper[4943]: E0307 15:04:00.142979 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="feed891c-4b59-472b-9cce-89d1c90652a5" containerName="oc" Mar 07 15:04:00 crc kubenswrapper[4943]: I0307 15:04:00.142990 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="feed891c-4b59-472b-9cce-89d1c90652a5" containerName="oc" Mar 07 15:04:00 crc kubenswrapper[4943]: I0307 15:04:00.143123 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="feed891c-4b59-472b-9cce-89d1c90652a5" containerName="oc" Mar 07 15:04:00 crc kubenswrapper[4943]: I0307 15:04:00.143566 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548264-qrpfx" Mar 07 15:04:00 crc kubenswrapper[4943]: I0307 15:04:00.149634 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 07 15:04:00 crc kubenswrapper[4943]: I0307 15:04:00.149996 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 07 15:04:00 crc kubenswrapper[4943]: I0307 15:04:00.150205 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-qqlk5" Mar 07 15:04:00 crc kubenswrapper[4943]: I0307 15:04:00.158092 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29548264-qrpfx"] Mar 07 15:04:00 crc kubenswrapper[4943]: I0307 15:04:00.270704 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6twvc\" (UniqueName: \"kubernetes.io/projected/0838f7e8-2cbe-4b70-a399-ff1200dc1693-kube-api-access-6twvc\") pod \"auto-csr-approver-29548264-qrpfx\" (UID: \"0838f7e8-2cbe-4b70-a399-ff1200dc1693\") " pod="openshift-infra/auto-csr-approver-29548264-qrpfx" Mar 07 15:04:00 crc kubenswrapper[4943]: I0307 15:04:00.372450 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6twvc\" (UniqueName: \"kubernetes.io/projected/0838f7e8-2cbe-4b70-a399-ff1200dc1693-kube-api-access-6twvc\") pod \"auto-csr-approver-29548264-qrpfx\" (UID: \"0838f7e8-2cbe-4b70-a399-ff1200dc1693\") " pod="openshift-infra/auto-csr-approver-29548264-qrpfx" Mar 07 15:04:00 crc kubenswrapper[4943]: I0307 15:04:00.401600 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6twvc\" (UniqueName: \"kubernetes.io/projected/0838f7e8-2cbe-4b70-a399-ff1200dc1693-kube-api-access-6twvc\") pod \"auto-csr-approver-29548264-qrpfx\" (UID: \"0838f7e8-2cbe-4b70-a399-ff1200dc1693\") " pod="openshift-infra/auto-csr-approver-29548264-qrpfx" Mar 07 15:04:00 crc kubenswrapper[4943]: I0307 15:04:00.463396 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548264-qrpfx" Mar 07 15:04:01 crc kubenswrapper[4943]: I0307 15:04:01.005310 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29548264-qrpfx"] Mar 07 15:04:01 crc kubenswrapper[4943]: I0307 15:04:01.176686 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548264-qrpfx" event={"ID":"0838f7e8-2cbe-4b70-a399-ff1200dc1693","Type":"ContainerStarted","Data":"81641a423db6ae52bde42156915e5451fecd5da355d9b70d7c9265e453912e58"} Mar 07 15:04:02 crc kubenswrapper[4943]: I0307 15:04:02.186241 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548264-qrpfx" event={"ID":"0838f7e8-2cbe-4b70-a399-ff1200dc1693","Type":"ContainerStarted","Data":"4ed71004898eb8cd3a17623fb532ec001eedf0b02e3f35dee7ff33da101e22db"} Mar 07 15:04:02 crc kubenswrapper[4943]: I0307 15:04:02.218591 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29548264-qrpfx" podStartSLOduration=1.383060501 podStartE2EDuration="2.218563855s" podCreationTimestamp="2026-03-07 15:04:00 +0000 UTC" firstStartedPulling="2026-03-07 15:04:01.0192427 +0000 UTC m=+1482.971379198" lastFinishedPulling="2026-03-07 15:04:01.854746014 +0000 UTC m=+1483.806882552" observedRunningTime="2026-03-07 15:04:02.213220655 +0000 UTC m=+1484.165357193" watchObservedRunningTime="2026-03-07 15:04:02.218563855 +0000 UTC m=+1484.170700383" Mar 07 15:04:03 crc kubenswrapper[4943]: I0307 15:04:03.197657 4943 generic.go:334] "Generic (PLEG): container finished" podID="0838f7e8-2cbe-4b70-a399-ff1200dc1693" containerID="4ed71004898eb8cd3a17623fb532ec001eedf0b02e3f35dee7ff33da101e22db" exitCode=0 Mar 07 15:04:03 crc kubenswrapper[4943]: I0307 15:04:03.197720 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548264-qrpfx" event={"ID":"0838f7e8-2cbe-4b70-a399-ff1200dc1693","Type":"ContainerDied","Data":"4ed71004898eb8cd3a17623fb532ec001eedf0b02e3f35dee7ff33da101e22db"} Mar 07 15:04:04 crc kubenswrapper[4943]: I0307 15:04:04.567315 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548264-qrpfx" Mar 07 15:04:04 crc kubenswrapper[4943]: I0307 15:04:04.735669 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6twvc\" (UniqueName: \"kubernetes.io/projected/0838f7e8-2cbe-4b70-a399-ff1200dc1693-kube-api-access-6twvc\") pod \"0838f7e8-2cbe-4b70-a399-ff1200dc1693\" (UID: \"0838f7e8-2cbe-4b70-a399-ff1200dc1693\") " Mar 07 15:04:04 crc kubenswrapper[4943]: I0307 15:04:04.746554 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0838f7e8-2cbe-4b70-a399-ff1200dc1693-kube-api-access-6twvc" (OuterVolumeSpecName: "kube-api-access-6twvc") pod "0838f7e8-2cbe-4b70-a399-ff1200dc1693" (UID: "0838f7e8-2cbe-4b70-a399-ff1200dc1693"). InnerVolumeSpecName "kube-api-access-6twvc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:04:04 crc kubenswrapper[4943]: I0307 15:04:04.838097 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6twvc\" (UniqueName: \"kubernetes.io/projected/0838f7e8-2cbe-4b70-a399-ff1200dc1693-kube-api-access-6twvc\") on node \"crc\" DevicePath \"\"" Mar 07 15:04:05 crc kubenswrapper[4943]: I0307 15:04:05.216429 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548264-qrpfx" event={"ID":"0838f7e8-2cbe-4b70-a399-ff1200dc1693","Type":"ContainerDied","Data":"81641a423db6ae52bde42156915e5451fecd5da355d9b70d7c9265e453912e58"} Mar 07 15:04:05 crc kubenswrapper[4943]: I0307 15:04:05.216691 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="81641a423db6ae52bde42156915e5451fecd5da355d9b70d7c9265e453912e58" Mar 07 15:04:05 crc kubenswrapper[4943]: I0307 15:04:05.216493 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548264-qrpfx" Mar 07 15:04:05 crc kubenswrapper[4943]: I0307 15:04:05.274860 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29548258-spxr6"] Mar 07 15:04:05 crc kubenswrapper[4943]: I0307 15:04:05.279970 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29548258-spxr6"] Mar 07 15:04:06 crc kubenswrapper[4943]: I0307 15:04:06.074166 4943 patch_prober.go:28] interesting pod/machine-config-daemon-cpdmm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 07 15:04:06 crc kubenswrapper[4943]: I0307 15:04:06.074226 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 07 15:04:06 crc kubenswrapper[4943]: I0307 15:04:06.768262 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6621055b-9adc-4057-bc14-63b4848edbca" path="/var/lib/kubelet/pods/6621055b-9adc-4057-bc14-63b4848edbca/volumes" Mar 07 15:04:09 crc kubenswrapper[4943]: I0307 15:04:09.068452 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/root-account-create-update-qkv6t"] Mar 07 15:04:09 crc kubenswrapper[4943]: I0307 15:04:09.076147 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/root-account-create-update-qkv6t"] Mar 07 15:04:10 crc kubenswrapper[4943]: I0307 15:04:10.769236 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21a7b33d-af38-44df-aca9-a602922a101f" path="/var/lib/kubelet/pods/21a7b33d-af38-44df-aca9-a602922a101f/volumes" Mar 07 15:04:20 crc kubenswrapper[4943]: I0307 15:04:20.245444 4943 scope.go:117] "RemoveContainer" containerID="e5dc11f9199090a79eeb09c228f04112c7dc7ee35b2769301fdcf56bf33efdd1" Mar 07 15:04:20 crc kubenswrapper[4943]: I0307 15:04:20.297398 4943 scope.go:117] "RemoveContainer" containerID="9aeef101e414d6daf2c8c8b10d657d3c5e34ec1c372f98d99be7fc3609de1900" Mar 07 15:04:20 crc kubenswrapper[4943]: I0307 15:04:20.323047 4943 scope.go:117] "RemoveContainer" containerID="7122613220e2db80a53f9f6dce66813c8891cd55870903238eb8dae123d96919" Mar 07 15:04:20 crc kubenswrapper[4943]: I0307 15:04:20.354420 4943 scope.go:117] "RemoveContainer" containerID="39de95413cade901e51f50b2d8f9abb6980ab8da312acae8fa47accbc002bb30" Mar 07 15:04:26 crc kubenswrapper[4943]: I0307 15:04:26.137252 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zh5q5"] Mar 07 15:04:26 crc kubenswrapper[4943]: E0307 15:04:26.137981 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0838f7e8-2cbe-4b70-a399-ff1200dc1693" containerName="oc" Mar 07 15:04:26 crc kubenswrapper[4943]: I0307 15:04:26.137991 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="0838f7e8-2cbe-4b70-a399-ff1200dc1693" containerName="oc" Mar 07 15:04:26 crc kubenswrapper[4943]: I0307 15:04:26.138125 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="0838f7e8-2cbe-4b70-a399-ff1200dc1693" containerName="oc" Mar 07 15:04:26 crc kubenswrapper[4943]: I0307 15:04:26.139002 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zh5q5" Mar 07 15:04:26 crc kubenswrapper[4943]: I0307 15:04:26.167354 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zh5q5"] Mar 07 15:04:26 crc kubenswrapper[4943]: I0307 15:04:26.281207 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7r6g\" (UniqueName: \"kubernetes.io/projected/5a86aa17-23ed-495c-83bf-a3188e63fbf6-kube-api-access-v7r6g\") pod \"redhat-marketplace-zh5q5\" (UID: \"5a86aa17-23ed-495c-83bf-a3188e63fbf6\") " pod="openshift-marketplace/redhat-marketplace-zh5q5" Mar 07 15:04:26 crc kubenswrapper[4943]: I0307 15:04:26.281500 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5a86aa17-23ed-495c-83bf-a3188e63fbf6-utilities\") pod \"redhat-marketplace-zh5q5\" (UID: \"5a86aa17-23ed-495c-83bf-a3188e63fbf6\") " pod="openshift-marketplace/redhat-marketplace-zh5q5" Mar 07 15:04:26 crc kubenswrapper[4943]: I0307 15:04:26.281593 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5a86aa17-23ed-495c-83bf-a3188e63fbf6-catalog-content\") pod \"redhat-marketplace-zh5q5\" (UID: \"5a86aa17-23ed-495c-83bf-a3188e63fbf6\") " pod="openshift-marketplace/redhat-marketplace-zh5q5" Mar 07 15:04:26 crc kubenswrapper[4943]: I0307 15:04:26.383022 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5a86aa17-23ed-495c-83bf-a3188e63fbf6-catalog-content\") pod \"redhat-marketplace-zh5q5\" (UID: \"5a86aa17-23ed-495c-83bf-a3188e63fbf6\") " pod="openshift-marketplace/redhat-marketplace-zh5q5" Mar 07 15:04:26 crc kubenswrapper[4943]: I0307 15:04:26.383112 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7r6g\" (UniqueName: \"kubernetes.io/projected/5a86aa17-23ed-495c-83bf-a3188e63fbf6-kube-api-access-v7r6g\") pod \"redhat-marketplace-zh5q5\" (UID: \"5a86aa17-23ed-495c-83bf-a3188e63fbf6\") " pod="openshift-marketplace/redhat-marketplace-zh5q5" Mar 07 15:04:26 crc kubenswrapper[4943]: I0307 15:04:26.383184 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5a86aa17-23ed-495c-83bf-a3188e63fbf6-utilities\") pod \"redhat-marketplace-zh5q5\" (UID: \"5a86aa17-23ed-495c-83bf-a3188e63fbf6\") " pod="openshift-marketplace/redhat-marketplace-zh5q5" Mar 07 15:04:26 crc kubenswrapper[4943]: I0307 15:04:26.383791 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5a86aa17-23ed-495c-83bf-a3188e63fbf6-utilities\") pod \"redhat-marketplace-zh5q5\" (UID: \"5a86aa17-23ed-495c-83bf-a3188e63fbf6\") " pod="openshift-marketplace/redhat-marketplace-zh5q5" Mar 07 15:04:26 crc kubenswrapper[4943]: I0307 15:04:26.383886 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5a86aa17-23ed-495c-83bf-a3188e63fbf6-catalog-content\") pod \"redhat-marketplace-zh5q5\" (UID: \"5a86aa17-23ed-495c-83bf-a3188e63fbf6\") " pod="openshift-marketplace/redhat-marketplace-zh5q5" Mar 07 15:04:26 crc kubenswrapper[4943]: I0307 15:04:26.405323 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7r6g\" (UniqueName: \"kubernetes.io/projected/5a86aa17-23ed-495c-83bf-a3188e63fbf6-kube-api-access-v7r6g\") pod \"redhat-marketplace-zh5q5\" (UID: \"5a86aa17-23ed-495c-83bf-a3188e63fbf6\") " pod="openshift-marketplace/redhat-marketplace-zh5q5" Mar 07 15:04:26 crc kubenswrapper[4943]: I0307 15:04:26.453838 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zh5q5" Mar 07 15:04:26 crc kubenswrapper[4943]: I0307 15:04:26.891600 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zh5q5"] Mar 07 15:04:27 crc kubenswrapper[4943]: I0307 15:04:27.442412 4943 generic.go:334] "Generic (PLEG): container finished" podID="5a86aa17-23ed-495c-83bf-a3188e63fbf6" containerID="732f95f512133b675348db781f2b5261b2a52d3e23fd32457b5369f41d15b8a9" exitCode=0 Mar 07 15:04:27 crc kubenswrapper[4943]: I0307 15:04:27.442526 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zh5q5" event={"ID":"5a86aa17-23ed-495c-83bf-a3188e63fbf6","Type":"ContainerDied","Data":"732f95f512133b675348db781f2b5261b2a52d3e23fd32457b5369f41d15b8a9"} Mar 07 15:04:27 crc kubenswrapper[4943]: I0307 15:04:27.442863 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zh5q5" event={"ID":"5a86aa17-23ed-495c-83bf-a3188e63fbf6","Type":"ContainerStarted","Data":"bdb60c7da25d0ee5ec5d42b800f403ed19993c27b6acd301508a1d758b8e05c7"} Mar 07 15:04:29 crc kubenswrapper[4943]: I0307 15:04:29.461121 4943 generic.go:334] "Generic (PLEG): container finished" podID="5a86aa17-23ed-495c-83bf-a3188e63fbf6" containerID="4e7cf0d746045a135c40c9957b20468368ac440749a790d9e17bd28ecaf2461e" exitCode=0 Mar 07 15:04:29 crc kubenswrapper[4943]: I0307 15:04:29.461276 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zh5q5" event={"ID":"5a86aa17-23ed-495c-83bf-a3188e63fbf6","Type":"ContainerDied","Data":"4e7cf0d746045a135c40c9957b20468368ac440749a790d9e17bd28ecaf2461e"} Mar 07 15:04:30 crc kubenswrapper[4943]: I0307 15:04:30.474115 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zh5q5" event={"ID":"5a86aa17-23ed-495c-83bf-a3188e63fbf6","Type":"ContainerStarted","Data":"fb5fd14fcad86887e6572546247087cbaff8a5d139b62ce3c8a9132b1227c77b"} Mar 07 15:04:30 crc kubenswrapper[4943]: I0307 15:04:30.499310 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-zh5q5" podStartSLOduration=2.022363542 podStartE2EDuration="4.499278132s" podCreationTimestamp="2026-03-07 15:04:26 +0000 UTC" firstStartedPulling="2026-03-07 15:04:27.445882573 +0000 UTC m=+1509.398019091" lastFinishedPulling="2026-03-07 15:04:29.922797143 +0000 UTC m=+1511.874933681" observedRunningTime="2026-03-07 15:04:30.494745571 +0000 UTC m=+1512.446882099" watchObservedRunningTime="2026-03-07 15:04:30.499278132 +0000 UTC m=+1512.451414670" Mar 07 15:04:31 crc kubenswrapper[4943]: E0307 15:04:31.367006 4943 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.75:56846->38.102.83.75:36531: write tcp 38.102.83.75:56846->38.102.83.75:36531: write: broken pipe Mar 07 15:04:36 crc kubenswrapper[4943]: I0307 15:04:36.074291 4943 patch_prober.go:28] interesting pod/machine-config-daemon-cpdmm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 07 15:04:36 crc kubenswrapper[4943]: I0307 15:04:36.075133 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 07 15:04:36 crc kubenswrapper[4943]: I0307 15:04:36.454710 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-zh5q5" Mar 07 15:04:36 crc kubenswrapper[4943]: I0307 15:04:36.454769 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-zh5q5" Mar 07 15:04:36 crc kubenswrapper[4943]: I0307 15:04:36.535389 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-zh5q5" Mar 07 15:04:36 crc kubenswrapper[4943]: I0307 15:04:36.606676 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-zh5q5" Mar 07 15:04:36 crc kubenswrapper[4943]: I0307 15:04:36.786892 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zh5q5"] Mar 07 15:04:38 crc kubenswrapper[4943]: I0307 15:04:38.553085 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-zh5q5" podUID="5a86aa17-23ed-495c-83bf-a3188e63fbf6" containerName="registry-server" containerID="cri-o://fb5fd14fcad86887e6572546247087cbaff8a5d139b62ce3c8a9132b1227c77b" gracePeriod=2 Mar 07 15:04:39 crc kubenswrapper[4943]: I0307 15:04:39.141709 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zh5q5" Mar 07 15:04:39 crc kubenswrapper[4943]: I0307 15:04:39.207251 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v7r6g\" (UniqueName: \"kubernetes.io/projected/5a86aa17-23ed-495c-83bf-a3188e63fbf6-kube-api-access-v7r6g\") pod \"5a86aa17-23ed-495c-83bf-a3188e63fbf6\" (UID: \"5a86aa17-23ed-495c-83bf-a3188e63fbf6\") " Mar 07 15:04:39 crc kubenswrapper[4943]: I0307 15:04:39.207516 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5a86aa17-23ed-495c-83bf-a3188e63fbf6-utilities\") pod \"5a86aa17-23ed-495c-83bf-a3188e63fbf6\" (UID: \"5a86aa17-23ed-495c-83bf-a3188e63fbf6\") " Mar 07 15:04:39 crc kubenswrapper[4943]: I0307 15:04:39.207606 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5a86aa17-23ed-495c-83bf-a3188e63fbf6-catalog-content\") pod \"5a86aa17-23ed-495c-83bf-a3188e63fbf6\" (UID: \"5a86aa17-23ed-495c-83bf-a3188e63fbf6\") " Mar 07 15:04:39 crc kubenswrapper[4943]: I0307 15:04:39.208370 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5a86aa17-23ed-495c-83bf-a3188e63fbf6-utilities" (OuterVolumeSpecName: "utilities") pod "5a86aa17-23ed-495c-83bf-a3188e63fbf6" (UID: "5a86aa17-23ed-495c-83bf-a3188e63fbf6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:04:39 crc kubenswrapper[4943]: I0307 15:04:39.212336 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a86aa17-23ed-495c-83bf-a3188e63fbf6-kube-api-access-v7r6g" (OuterVolumeSpecName: "kube-api-access-v7r6g") pod "5a86aa17-23ed-495c-83bf-a3188e63fbf6" (UID: "5a86aa17-23ed-495c-83bf-a3188e63fbf6"). InnerVolumeSpecName "kube-api-access-v7r6g". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:04:39 crc kubenswrapper[4943]: I0307 15:04:39.241263 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5a86aa17-23ed-495c-83bf-a3188e63fbf6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5a86aa17-23ed-495c-83bf-a3188e63fbf6" (UID: "5a86aa17-23ed-495c-83bf-a3188e63fbf6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:04:39 crc kubenswrapper[4943]: I0307 15:04:39.309787 4943 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5a86aa17-23ed-495c-83bf-a3188e63fbf6-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 07 15:04:39 crc kubenswrapper[4943]: I0307 15:04:39.309828 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v7r6g\" (UniqueName: \"kubernetes.io/projected/5a86aa17-23ed-495c-83bf-a3188e63fbf6-kube-api-access-v7r6g\") on node \"crc\" DevicePath \"\"" Mar 07 15:04:39 crc kubenswrapper[4943]: I0307 15:04:39.309842 4943 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5a86aa17-23ed-495c-83bf-a3188e63fbf6-utilities\") on node \"crc\" DevicePath \"\"" Mar 07 15:04:39 crc kubenswrapper[4943]: I0307 15:04:39.596790 4943 generic.go:334] "Generic (PLEG): container finished" podID="5a86aa17-23ed-495c-83bf-a3188e63fbf6" containerID="fb5fd14fcad86887e6572546247087cbaff8a5d139b62ce3c8a9132b1227c77b" exitCode=0 Mar 07 15:04:39 crc kubenswrapper[4943]: I0307 15:04:39.597240 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zh5q5" event={"ID":"5a86aa17-23ed-495c-83bf-a3188e63fbf6","Type":"ContainerDied","Data":"fb5fd14fcad86887e6572546247087cbaff8a5d139b62ce3c8a9132b1227c77b"} Mar 07 15:04:39 crc kubenswrapper[4943]: I0307 15:04:39.597278 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zh5q5" event={"ID":"5a86aa17-23ed-495c-83bf-a3188e63fbf6","Type":"ContainerDied","Data":"bdb60c7da25d0ee5ec5d42b800f403ed19993c27b6acd301508a1d758b8e05c7"} Mar 07 15:04:39 crc kubenswrapper[4943]: I0307 15:04:39.597304 4943 scope.go:117] "RemoveContainer" containerID="fb5fd14fcad86887e6572546247087cbaff8a5d139b62ce3c8a9132b1227c77b" Mar 07 15:04:39 crc kubenswrapper[4943]: I0307 15:04:39.597501 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zh5q5" Mar 07 15:04:39 crc kubenswrapper[4943]: I0307 15:04:39.622915 4943 scope.go:117] "RemoveContainer" containerID="4e7cf0d746045a135c40c9957b20468368ac440749a790d9e17bd28ecaf2461e" Mar 07 15:04:39 crc kubenswrapper[4943]: I0307 15:04:39.657458 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zh5q5"] Mar 07 15:04:39 crc kubenswrapper[4943]: I0307 15:04:39.659878 4943 scope.go:117] "RemoveContainer" containerID="732f95f512133b675348db781f2b5261b2a52d3e23fd32457b5369f41d15b8a9" Mar 07 15:04:39 crc kubenswrapper[4943]: I0307 15:04:39.665551 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-zh5q5"] Mar 07 15:04:39 crc kubenswrapper[4943]: I0307 15:04:39.690093 4943 scope.go:117] "RemoveContainer" containerID="fb5fd14fcad86887e6572546247087cbaff8a5d139b62ce3c8a9132b1227c77b" Mar 07 15:04:39 crc kubenswrapper[4943]: E0307 15:04:39.692816 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb5fd14fcad86887e6572546247087cbaff8a5d139b62ce3c8a9132b1227c77b\": container with ID starting with fb5fd14fcad86887e6572546247087cbaff8a5d139b62ce3c8a9132b1227c77b not found: ID does not exist" containerID="fb5fd14fcad86887e6572546247087cbaff8a5d139b62ce3c8a9132b1227c77b" Mar 07 15:04:39 crc kubenswrapper[4943]: I0307 15:04:39.692857 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb5fd14fcad86887e6572546247087cbaff8a5d139b62ce3c8a9132b1227c77b"} err="failed to get container status \"fb5fd14fcad86887e6572546247087cbaff8a5d139b62ce3c8a9132b1227c77b\": rpc error: code = NotFound desc = could not find container \"fb5fd14fcad86887e6572546247087cbaff8a5d139b62ce3c8a9132b1227c77b\": container with ID starting with fb5fd14fcad86887e6572546247087cbaff8a5d139b62ce3c8a9132b1227c77b not found: ID does not exist" Mar 07 15:04:39 crc kubenswrapper[4943]: I0307 15:04:39.692883 4943 scope.go:117] "RemoveContainer" containerID="4e7cf0d746045a135c40c9957b20468368ac440749a790d9e17bd28ecaf2461e" Mar 07 15:04:39 crc kubenswrapper[4943]: E0307 15:04:39.703822 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e7cf0d746045a135c40c9957b20468368ac440749a790d9e17bd28ecaf2461e\": container with ID starting with 4e7cf0d746045a135c40c9957b20468368ac440749a790d9e17bd28ecaf2461e not found: ID does not exist" containerID="4e7cf0d746045a135c40c9957b20468368ac440749a790d9e17bd28ecaf2461e" Mar 07 15:04:39 crc kubenswrapper[4943]: I0307 15:04:39.703884 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e7cf0d746045a135c40c9957b20468368ac440749a790d9e17bd28ecaf2461e"} err="failed to get container status \"4e7cf0d746045a135c40c9957b20468368ac440749a790d9e17bd28ecaf2461e\": rpc error: code = NotFound desc = could not find container \"4e7cf0d746045a135c40c9957b20468368ac440749a790d9e17bd28ecaf2461e\": container with ID starting with 4e7cf0d746045a135c40c9957b20468368ac440749a790d9e17bd28ecaf2461e not found: ID does not exist" Mar 07 15:04:39 crc kubenswrapper[4943]: I0307 15:04:39.703947 4943 scope.go:117] "RemoveContainer" containerID="732f95f512133b675348db781f2b5261b2a52d3e23fd32457b5369f41d15b8a9" Mar 07 15:04:39 crc kubenswrapper[4943]: E0307 15:04:39.707188 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"732f95f512133b675348db781f2b5261b2a52d3e23fd32457b5369f41d15b8a9\": container with ID starting with 732f95f512133b675348db781f2b5261b2a52d3e23fd32457b5369f41d15b8a9 not found: ID does not exist" containerID="732f95f512133b675348db781f2b5261b2a52d3e23fd32457b5369f41d15b8a9" Mar 07 15:04:39 crc kubenswrapper[4943]: I0307 15:04:39.707282 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"732f95f512133b675348db781f2b5261b2a52d3e23fd32457b5369f41d15b8a9"} err="failed to get container status \"732f95f512133b675348db781f2b5261b2a52d3e23fd32457b5369f41d15b8a9\": rpc error: code = NotFound desc = could not find container \"732f95f512133b675348db781f2b5261b2a52d3e23fd32457b5369f41d15b8a9\": container with ID starting with 732f95f512133b675348db781f2b5261b2a52d3e23fd32457b5369f41d15b8a9 not found: ID does not exist" Mar 07 15:04:40 crc kubenswrapper[4943]: I0307 15:04:40.773055 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a86aa17-23ed-495c-83bf-a3188e63fbf6" path="/var/lib/kubelet/pods/5a86aa17-23ed-495c-83bf-a3188e63fbf6/volumes" Mar 07 15:04:46 crc kubenswrapper[4943]: E0307 15:04:46.695772 4943 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.75:43458->38.102.83.75:36531: write tcp 38.102.83.75:43458->38.102.83.75:36531: write: broken pipe Mar 07 15:05:06 crc kubenswrapper[4943]: I0307 15:05:06.073725 4943 patch_prober.go:28] interesting pod/machine-config-daemon-cpdmm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 07 15:05:06 crc kubenswrapper[4943]: I0307 15:05:06.074491 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 07 15:05:06 crc kubenswrapper[4943]: I0307 15:05:06.074576 4943 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" Mar 07 15:05:06 crc kubenswrapper[4943]: I0307 15:05:06.075334 4943 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"26f35e1f941cca0c34e3650d7aa047834e769dd76d2a61c3a39cad5bb2d090ae"} pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 07 15:05:06 crc kubenswrapper[4943]: I0307 15:05:06.075427 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerName="machine-config-daemon" containerID="cri-o://26f35e1f941cca0c34e3650d7aa047834e769dd76d2a61c3a39cad5bb2d090ae" gracePeriod=600 Mar 07 15:05:06 crc kubenswrapper[4943]: E0307 15:05:06.204261 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cpdmm_openshift-machine-config-operator(4c1f18ea-e557-41fc-95dd-ce4e1722a61e)\"" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" Mar 07 15:05:06 crc kubenswrapper[4943]: I0307 15:05:06.882435 4943 generic.go:334] "Generic (PLEG): container finished" podID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerID="26f35e1f941cca0c34e3650d7aa047834e769dd76d2a61c3a39cad5bb2d090ae" exitCode=0 Mar 07 15:05:06 crc kubenswrapper[4943]: I0307 15:05:06.882502 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" event={"ID":"4c1f18ea-e557-41fc-95dd-ce4e1722a61e","Type":"ContainerDied","Data":"26f35e1f941cca0c34e3650d7aa047834e769dd76d2a61c3a39cad5bb2d090ae"} Mar 07 15:05:06 crc kubenswrapper[4943]: I0307 15:05:06.882549 4943 scope.go:117] "RemoveContainer" containerID="526f9b1306d1082bdb6e6f98c1f04669533c0a9945daeb9cde6060502da5cecb" Mar 07 15:05:06 crc kubenswrapper[4943]: I0307 15:05:06.883396 4943 scope.go:117] "RemoveContainer" containerID="26f35e1f941cca0c34e3650d7aa047834e769dd76d2a61c3a39cad5bb2d090ae" Mar 07 15:05:06 crc kubenswrapper[4943]: E0307 15:05:06.883818 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cpdmm_openshift-machine-config-operator(4c1f18ea-e557-41fc-95dd-ce4e1722a61e)\"" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" Mar 07 15:05:17 crc kubenswrapper[4943]: I0307 15:05:17.094075 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-q2jd7"] Mar 07 15:05:17 crc kubenswrapper[4943]: E0307 15:05:17.094881 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a86aa17-23ed-495c-83bf-a3188e63fbf6" containerName="extract-utilities" Mar 07 15:05:17 crc kubenswrapper[4943]: I0307 15:05:17.094895 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a86aa17-23ed-495c-83bf-a3188e63fbf6" containerName="extract-utilities" Mar 07 15:05:17 crc kubenswrapper[4943]: E0307 15:05:17.094910 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a86aa17-23ed-495c-83bf-a3188e63fbf6" containerName="extract-content" Mar 07 15:05:17 crc kubenswrapper[4943]: I0307 15:05:17.094916 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a86aa17-23ed-495c-83bf-a3188e63fbf6" containerName="extract-content" Mar 07 15:05:17 crc kubenswrapper[4943]: E0307 15:05:17.094931 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a86aa17-23ed-495c-83bf-a3188e63fbf6" containerName="registry-server" Mar 07 15:05:17 crc kubenswrapper[4943]: I0307 15:05:17.094951 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a86aa17-23ed-495c-83bf-a3188e63fbf6" containerName="registry-server" Mar 07 15:05:17 crc kubenswrapper[4943]: I0307 15:05:17.095100 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a86aa17-23ed-495c-83bf-a3188e63fbf6" containerName="registry-server" Mar 07 15:05:17 crc kubenswrapper[4943]: I0307 15:05:17.096032 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q2jd7" Mar 07 15:05:17 crc kubenswrapper[4943]: I0307 15:05:17.110055 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-q2jd7"] Mar 07 15:05:17 crc kubenswrapper[4943]: I0307 15:05:17.281472 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5s56x\" (UniqueName: \"kubernetes.io/projected/3f2c31eb-4758-44de-8ebc-f76e877cc899-kube-api-access-5s56x\") pod \"community-operators-q2jd7\" (UID: \"3f2c31eb-4758-44de-8ebc-f76e877cc899\") " pod="openshift-marketplace/community-operators-q2jd7" Mar 07 15:05:17 crc kubenswrapper[4943]: I0307 15:05:17.281524 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f2c31eb-4758-44de-8ebc-f76e877cc899-utilities\") pod \"community-operators-q2jd7\" (UID: \"3f2c31eb-4758-44de-8ebc-f76e877cc899\") " pod="openshift-marketplace/community-operators-q2jd7" Mar 07 15:05:17 crc kubenswrapper[4943]: I0307 15:05:17.281912 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f2c31eb-4758-44de-8ebc-f76e877cc899-catalog-content\") pod \"community-operators-q2jd7\" (UID: \"3f2c31eb-4758-44de-8ebc-f76e877cc899\") " pod="openshift-marketplace/community-operators-q2jd7" Mar 07 15:05:17 crc kubenswrapper[4943]: I0307 15:05:17.383584 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f2c31eb-4758-44de-8ebc-f76e877cc899-catalog-content\") pod \"community-operators-q2jd7\" (UID: \"3f2c31eb-4758-44de-8ebc-f76e877cc899\") " pod="openshift-marketplace/community-operators-q2jd7" Mar 07 15:05:17 crc kubenswrapper[4943]: I0307 15:05:17.383711 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5s56x\" (UniqueName: \"kubernetes.io/projected/3f2c31eb-4758-44de-8ebc-f76e877cc899-kube-api-access-5s56x\") pod \"community-operators-q2jd7\" (UID: \"3f2c31eb-4758-44de-8ebc-f76e877cc899\") " pod="openshift-marketplace/community-operators-q2jd7" Mar 07 15:05:17 crc kubenswrapper[4943]: I0307 15:05:17.383744 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f2c31eb-4758-44de-8ebc-f76e877cc899-utilities\") pod \"community-operators-q2jd7\" (UID: \"3f2c31eb-4758-44de-8ebc-f76e877cc899\") " pod="openshift-marketplace/community-operators-q2jd7" Mar 07 15:05:17 crc kubenswrapper[4943]: I0307 15:05:17.384562 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f2c31eb-4758-44de-8ebc-f76e877cc899-utilities\") pod \"community-operators-q2jd7\" (UID: \"3f2c31eb-4758-44de-8ebc-f76e877cc899\") " pod="openshift-marketplace/community-operators-q2jd7" Mar 07 15:05:17 crc kubenswrapper[4943]: I0307 15:05:17.384562 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f2c31eb-4758-44de-8ebc-f76e877cc899-catalog-content\") pod \"community-operators-q2jd7\" (UID: \"3f2c31eb-4758-44de-8ebc-f76e877cc899\") " pod="openshift-marketplace/community-operators-q2jd7" Mar 07 15:05:17 crc kubenswrapper[4943]: I0307 15:05:17.410287 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5s56x\" (UniqueName: \"kubernetes.io/projected/3f2c31eb-4758-44de-8ebc-f76e877cc899-kube-api-access-5s56x\") pod \"community-operators-q2jd7\" (UID: \"3f2c31eb-4758-44de-8ebc-f76e877cc899\") " pod="openshift-marketplace/community-operators-q2jd7" Mar 07 15:05:17 crc kubenswrapper[4943]: I0307 15:05:17.416008 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q2jd7" Mar 07 15:05:17 crc kubenswrapper[4943]: I0307 15:05:17.935272 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-q2jd7"] Mar 07 15:05:17 crc kubenswrapper[4943]: I0307 15:05:17.995435 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q2jd7" event={"ID":"3f2c31eb-4758-44de-8ebc-f76e877cc899","Type":"ContainerStarted","Data":"13f48bd8e702e3e61d4b4a85f6b2087c0380d2afc1db23299d74f81b5bf88604"} Mar 07 15:05:19 crc kubenswrapper[4943]: I0307 15:05:19.002507 4943 generic.go:334] "Generic (PLEG): container finished" podID="3f2c31eb-4758-44de-8ebc-f76e877cc899" containerID="df67ca685d907dbb5eb1e9a16add75f6bf16a76fb6303f0829069fd8057266d6" exitCode=0 Mar 07 15:05:19 crc kubenswrapper[4943]: I0307 15:05:19.002555 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q2jd7" event={"ID":"3f2c31eb-4758-44de-8ebc-f76e877cc899","Type":"ContainerDied","Data":"df67ca685d907dbb5eb1e9a16add75f6bf16a76fb6303f0829069fd8057266d6"} Mar 07 15:05:19 crc kubenswrapper[4943]: I0307 15:05:19.005092 4943 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 07 15:05:20 crc kubenswrapper[4943]: I0307 15:05:20.012307 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q2jd7" event={"ID":"3f2c31eb-4758-44de-8ebc-f76e877cc899","Type":"ContainerStarted","Data":"8b4e8434294e0cea6704dc61a671a566faad06ccb1e051768608b920b1442f70"} Mar 07 15:05:20 crc kubenswrapper[4943]: I0307 15:05:20.442422 4943 scope.go:117] "RemoveContainer" containerID="c31fb8459a7f0666328afbbc11aa2081c9308b186d597d14883f1cc883104844" Mar 07 15:05:21 crc kubenswrapper[4943]: I0307 15:05:21.027164 4943 generic.go:334] "Generic (PLEG): container finished" podID="3f2c31eb-4758-44de-8ebc-f76e877cc899" containerID="8b4e8434294e0cea6704dc61a671a566faad06ccb1e051768608b920b1442f70" exitCode=0 Mar 07 15:05:21 crc kubenswrapper[4943]: I0307 15:05:21.028411 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q2jd7" event={"ID":"3f2c31eb-4758-44de-8ebc-f76e877cc899","Type":"ContainerDied","Data":"8b4e8434294e0cea6704dc61a671a566faad06ccb1e051768608b920b1442f70"} Mar 07 15:05:21 crc kubenswrapper[4943]: I0307 15:05:21.792542 4943 scope.go:117] "RemoveContainer" containerID="26f35e1f941cca0c34e3650d7aa047834e769dd76d2a61c3a39cad5bb2d090ae" Mar 07 15:05:21 crc kubenswrapper[4943]: E0307 15:05:21.793507 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cpdmm_openshift-machine-config-operator(4c1f18ea-e557-41fc-95dd-ce4e1722a61e)\"" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" Mar 07 15:05:22 crc kubenswrapper[4943]: I0307 15:05:22.067431 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q2jd7" event={"ID":"3f2c31eb-4758-44de-8ebc-f76e877cc899","Type":"ContainerStarted","Data":"cb51832e6b7fbb91639a50a7f088545f8ffac9372194e7a9ce6023df452f65a1"} Mar 07 15:05:22 crc kubenswrapper[4943]: I0307 15:05:22.098773 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-q2jd7" podStartSLOduration=2.673800497 podStartE2EDuration="5.098754556s" podCreationTimestamp="2026-03-07 15:05:17 +0000 UTC" firstStartedPulling="2026-03-07 15:05:19.004726943 +0000 UTC m=+1560.956863451" lastFinishedPulling="2026-03-07 15:05:21.429680972 +0000 UTC m=+1563.381817510" observedRunningTime="2026-03-07 15:05:22.096751567 +0000 UTC m=+1564.048888115" watchObservedRunningTime="2026-03-07 15:05:22.098754556 +0000 UTC m=+1564.050891064" Mar 07 15:05:27 crc kubenswrapper[4943]: I0307 15:05:27.417123 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-q2jd7" Mar 07 15:05:27 crc kubenswrapper[4943]: I0307 15:05:27.417666 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-q2jd7" Mar 07 15:05:27 crc kubenswrapper[4943]: I0307 15:05:27.492964 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-q2jd7" Mar 07 15:05:28 crc kubenswrapper[4943]: I0307 15:05:28.190032 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-q2jd7" Mar 07 15:05:28 crc kubenswrapper[4943]: I0307 15:05:28.250246 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-q2jd7"] Mar 07 15:05:30 crc kubenswrapper[4943]: I0307 15:05:30.143947 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-q2jd7" podUID="3f2c31eb-4758-44de-8ebc-f76e877cc899" containerName="registry-server" containerID="cri-o://cb51832e6b7fbb91639a50a7f088545f8ffac9372194e7a9ce6023df452f65a1" gracePeriod=2 Mar 07 15:05:30 crc kubenswrapper[4943]: I0307 15:05:30.165759 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9rcxq"] Mar 07 15:05:30 crc kubenswrapper[4943]: I0307 15:05:30.168545 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9rcxq" Mar 07 15:05:30 crc kubenswrapper[4943]: I0307 15:05:30.184118 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9rcxq"] Mar 07 15:05:30 crc kubenswrapper[4943]: I0307 15:05:30.334970 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwgnv\" (UniqueName: \"kubernetes.io/projected/28c32ced-1924-491a-add9-496101b1c366-kube-api-access-zwgnv\") pod \"certified-operators-9rcxq\" (UID: \"28c32ced-1924-491a-add9-496101b1c366\") " pod="openshift-marketplace/certified-operators-9rcxq" Mar 07 15:05:30 crc kubenswrapper[4943]: I0307 15:05:30.335100 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28c32ced-1924-491a-add9-496101b1c366-utilities\") pod \"certified-operators-9rcxq\" (UID: \"28c32ced-1924-491a-add9-496101b1c366\") " pod="openshift-marketplace/certified-operators-9rcxq" Mar 07 15:05:30 crc kubenswrapper[4943]: I0307 15:05:30.335135 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28c32ced-1924-491a-add9-496101b1c366-catalog-content\") pod \"certified-operators-9rcxq\" (UID: \"28c32ced-1924-491a-add9-496101b1c366\") " pod="openshift-marketplace/certified-operators-9rcxq" Mar 07 15:05:30 crc kubenswrapper[4943]: I0307 15:05:30.437704 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28c32ced-1924-491a-add9-496101b1c366-utilities\") pod \"certified-operators-9rcxq\" (UID: \"28c32ced-1924-491a-add9-496101b1c366\") " pod="openshift-marketplace/certified-operators-9rcxq" Mar 07 15:05:30 crc kubenswrapper[4943]: I0307 15:05:30.438170 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28c32ced-1924-491a-add9-496101b1c366-utilities\") pod \"certified-operators-9rcxq\" (UID: \"28c32ced-1924-491a-add9-496101b1c366\") " pod="openshift-marketplace/certified-operators-9rcxq" Mar 07 15:05:30 crc kubenswrapper[4943]: I0307 15:05:30.438302 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28c32ced-1924-491a-add9-496101b1c366-catalog-content\") pod \"certified-operators-9rcxq\" (UID: \"28c32ced-1924-491a-add9-496101b1c366\") " pod="openshift-marketplace/certified-operators-9rcxq" Mar 07 15:05:30 crc kubenswrapper[4943]: I0307 15:05:30.438428 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwgnv\" (UniqueName: \"kubernetes.io/projected/28c32ced-1924-491a-add9-496101b1c366-kube-api-access-zwgnv\") pod \"certified-operators-9rcxq\" (UID: \"28c32ced-1924-491a-add9-496101b1c366\") " pod="openshift-marketplace/certified-operators-9rcxq" Mar 07 15:05:30 crc kubenswrapper[4943]: I0307 15:05:30.439087 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28c32ced-1924-491a-add9-496101b1c366-catalog-content\") pod \"certified-operators-9rcxq\" (UID: \"28c32ced-1924-491a-add9-496101b1c366\") " pod="openshift-marketplace/certified-operators-9rcxq" Mar 07 15:05:30 crc kubenswrapper[4943]: I0307 15:05:30.460960 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwgnv\" (UniqueName: \"kubernetes.io/projected/28c32ced-1924-491a-add9-496101b1c366-kube-api-access-zwgnv\") pod \"certified-operators-9rcxq\" (UID: \"28c32ced-1924-491a-add9-496101b1c366\") " pod="openshift-marketplace/certified-operators-9rcxq" Mar 07 15:05:30 crc kubenswrapper[4943]: I0307 15:05:30.506720 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9rcxq" Mar 07 15:05:30 crc kubenswrapper[4943]: I0307 15:05:30.623233 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q2jd7" Mar 07 15:05:30 crc kubenswrapper[4943]: I0307 15:05:30.743740 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5s56x\" (UniqueName: \"kubernetes.io/projected/3f2c31eb-4758-44de-8ebc-f76e877cc899-kube-api-access-5s56x\") pod \"3f2c31eb-4758-44de-8ebc-f76e877cc899\" (UID: \"3f2c31eb-4758-44de-8ebc-f76e877cc899\") " Mar 07 15:05:30 crc kubenswrapper[4943]: I0307 15:05:30.743792 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f2c31eb-4758-44de-8ebc-f76e877cc899-catalog-content\") pod \"3f2c31eb-4758-44de-8ebc-f76e877cc899\" (UID: \"3f2c31eb-4758-44de-8ebc-f76e877cc899\") " Mar 07 15:05:30 crc kubenswrapper[4943]: I0307 15:05:30.743905 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f2c31eb-4758-44de-8ebc-f76e877cc899-utilities\") pod \"3f2c31eb-4758-44de-8ebc-f76e877cc899\" (UID: \"3f2c31eb-4758-44de-8ebc-f76e877cc899\") " Mar 07 15:05:30 crc kubenswrapper[4943]: I0307 15:05:30.744747 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f2c31eb-4758-44de-8ebc-f76e877cc899-utilities" (OuterVolumeSpecName: "utilities") pod "3f2c31eb-4758-44de-8ebc-f76e877cc899" (UID: "3f2c31eb-4758-44de-8ebc-f76e877cc899"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:05:30 crc kubenswrapper[4943]: I0307 15:05:30.748372 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f2c31eb-4758-44de-8ebc-f76e877cc899-kube-api-access-5s56x" (OuterVolumeSpecName: "kube-api-access-5s56x") pod "3f2c31eb-4758-44de-8ebc-f76e877cc899" (UID: "3f2c31eb-4758-44de-8ebc-f76e877cc899"). InnerVolumeSpecName "kube-api-access-5s56x". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:05:30 crc kubenswrapper[4943]: I0307 15:05:30.788730 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f2c31eb-4758-44de-8ebc-f76e877cc899-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3f2c31eb-4758-44de-8ebc-f76e877cc899" (UID: "3f2c31eb-4758-44de-8ebc-f76e877cc899"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:05:30 crc kubenswrapper[4943]: I0307 15:05:30.844954 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5s56x\" (UniqueName: \"kubernetes.io/projected/3f2c31eb-4758-44de-8ebc-f76e877cc899-kube-api-access-5s56x\") on node \"crc\" DevicePath \"\"" Mar 07 15:05:30 crc kubenswrapper[4943]: I0307 15:05:30.844987 4943 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f2c31eb-4758-44de-8ebc-f76e877cc899-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 07 15:05:30 crc kubenswrapper[4943]: I0307 15:05:30.844996 4943 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f2c31eb-4758-44de-8ebc-f76e877cc899-utilities\") on node \"crc\" DevicePath \"\"" Mar 07 15:05:30 crc kubenswrapper[4943]: I0307 15:05:30.934246 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9rcxq"] Mar 07 15:05:31 crc kubenswrapper[4943]: I0307 15:05:31.151277 4943 generic.go:334] "Generic (PLEG): container finished" podID="28c32ced-1924-491a-add9-496101b1c366" containerID="6e7492e44c9a99e0bfa5adab9201d8384619d6eff67e925effa806f9c17bfec7" exitCode=0 Mar 07 15:05:31 crc kubenswrapper[4943]: I0307 15:05:31.151841 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9rcxq" event={"ID":"28c32ced-1924-491a-add9-496101b1c366","Type":"ContainerDied","Data":"6e7492e44c9a99e0bfa5adab9201d8384619d6eff67e925effa806f9c17bfec7"} Mar 07 15:05:31 crc kubenswrapper[4943]: I0307 15:05:31.151868 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9rcxq" event={"ID":"28c32ced-1924-491a-add9-496101b1c366","Type":"ContainerStarted","Data":"65b94629bf32afbddbe9af4e7eda555bfcf3c9b8a9642438a0cac07f0c004c97"} Mar 07 15:05:31 crc kubenswrapper[4943]: I0307 15:05:31.154111 4943 generic.go:334] "Generic (PLEG): container finished" podID="3f2c31eb-4758-44de-8ebc-f76e877cc899" containerID="cb51832e6b7fbb91639a50a7f088545f8ffac9372194e7a9ce6023df452f65a1" exitCode=0 Mar 07 15:05:31 crc kubenswrapper[4943]: I0307 15:05:31.154183 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q2jd7" Mar 07 15:05:31 crc kubenswrapper[4943]: I0307 15:05:31.154161 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q2jd7" event={"ID":"3f2c31eb-4758-44de-8ebc-f76e877cc899","Type":"ContainerDied","Data":"cb51832e6b7fbb91639a50a7f088545f8ffac9372194e7a9ce6023df452f65a1"} Mar 07 15:05:31 crc kubenswrapper[4943]: I0307 15:05:31.154318 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q2jd7" event={"ID":"3f2c31eb-4758-44de-8ebc-f76e877cc899","Type":"ContainerDied","Data":"13f48bd8e702e3e61d4b4a85f6b2087c0380d2afc1db23299d74f81b5bf88604"} Mar 07 15:05:31 crc kubenswrapper[4943]: I0307 15:05:31.154341 4943 scope.go:117] "RemoveContainer" containerID="cb51832e6b7fbb91639a50a7f088545f8ffac9372194e7a9ce6023df452f65a1" Mar 07 15:05:31 crc kubenswrapper[4943]: I0307 15:05:31.197204 4943 scope.go:117] "RemoveContainer" containerID="8b4e8434294e0cea6704dc61a671a566faad06ccb1e051768608b920b1442f70" Mar 07 15:05:31 crc kubenswrapper[4943]: I0307 15:05:31.205686 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-q2jd7"] Mar 07 15:05:31 crc kubenswrapper[4943]: I0307 15:05:31.210816 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-q2jd7"] Mar 07 15:05:31 crc kubenswrapper[4943]: I0307 15:05:31.218977 4943 scope.go:117] "RemoveContainer" containerID="df67ca685d907dbb5eb1e9a16add75f6bf16a76fb6303f0829069fd8057266d6" Mar 07 15:05:31 crc kubenswrapper[4943]: I0307 15:05:31.235516 4943 scope.go:117] "RemoveContainer" containerID="cb51832e6b7fbb91639a50a7f088545f8ffac9372194e7a9ce6023df452f65a1" Mar 07 15:05:31 crc kubenswrapper[4943]: E0307 15:05:31.235869 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb51832e6b7fbb91639a50a7f088545f8ffac9372194e7a9ce6023df452f65a1\": container with ID starting with cb51832e6b7fbb91639a50a7f088545f8ffac9372194e7a9ce6023df452f65a1 not found: ID does not exist" containerID="cb51832e6b7fbb91639a50a7f088545f8ffac9372194e7a9ce6023df452f65a1" Mar 07 15:05:31 crc kubenswrapper[4943]: I0307 15:05:31.235898 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb51832e6b7fbb91639a50a7f088545f8ffac9372194e7a9ce6023df452f65a1"} err="failed to get container status \"cb51832e6b7fbb91639a50a7f088545f8ffac9372194e7a9ce6023df452f65a1\": rpc error: code = NotFound desc = could not find container \"cb51832e6b7fbb91639a50a7f088545f8ffac9372194e7a9ce6023df452f65a1\": container with ID starting with cb51832e6b7fbb91639a50a7f088545f8ffac9372194e7a9ce6023df452f65a1 not found: ID does not exist" Mar 07 15:05:31 crc kubenswrapper[4943]: I0307 15:05:31.235918 4943 scope.go:117] "RemoveContainer" containerID="8b4e8434294e0cea6704dc61a671a566faad06ccb1e051768608b920b1442f70" Mar 07 15:05:31 crc kubenswrapper[4943]: E0307 15:05:31.236110 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b4e8434294e0cea6704dc61a671a566faad06ccb1e051768608b920b1442f70\": container with ID starting with 8b4e8434294e0cea6704dc61a671a566faad06ccb1e051768608b920b1442f70 not found: ID does not exist" containerID="8b4e8434294e0cea6704dc61a671a566faad06ccb1e051768608b920b1442f70" Mar 07 15:05:31 crc kubenswrapper[4943]: I0307 15:05:31.236132 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b4e8434294e0cea6704dc61a671a566faad06ccb1e051768608b920b1442f70"} err="failed to get container status \"8b4e8434294e0cea6704dc61a671a566faad06ccb1e051768608b920b1442f70\": rpc error: code = NotFound desc = could not find container \"8b4e8434294e0cea6704dc61a671a566faad06ccb1e051768608b920b1442f70\": container with ID starting with 8b4e8434294e0cea6704dc61a671a566faad06ccb1e051768608b920b1442f70 not found: ID does not exist" Mar 07 15:05:31 crc kubenswrapper[4943]: I0307 15:05:31.236146 4943 scope.go:117] "RemoveContainer" containerID="df67ca685d907dbb5eb1e9a16add75f6bf16a76fb6303f0829069fd8057266d6" Mar 07 15:05:31 crc kubenswrapper[4943]: E0307 15:05:31.236376 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df67ca685d907dbb5eb1e9a16add75f6bf16a76fb6303f0829069fd8057266d6\": container with ID starting with df67ca685d907dbb5eb1e9a16add75f6bf16a76fb6303f0829069fd8057266d6 not found: ID does not exist" containerID="df67ca685d907dbb5eb1e9a16add75f6bf16a76fb6303f0829069fd8057266d6" Mar 07 15:05:31 crc kubenswrapper[4943]: I0307 15:05:31.236405 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df67ca685d907dbb5eb1e9a16add75f6bf16a76fb6303f0829069fd8057266d6"} err="failed to get container status \"df67ca685d907dbb5eb1e9a16add75f6bf16a76fb6303f0829069fd8057266d6\": rpc error: code = NotFound desc = could not find container \"df67ca685d907dbb5eb1e9a16add75f6bf16a76fb6303f0829069fd8057266d6\": container with ID starting with df67ca685d907dbb5eb1e9a16add75f6bf16a76fb6303f0829069fd8057266d6 not found: ID does not exist" Mar 07 15:05:32 crc kubenswrapper[4943]: I0307 15:05:32.171365 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9rcxq" event={"ID":"28c32ced-1924-491a-add9-496101b1c366","Type":"ContainerStarted","Data":"c9874a40cc2995cba1d7f38c57fdffcac8876054c02bb9c34eb4b93bb8f291cb"} Mar 07 15:05:32 crc kubenswrapper[4943]: I0307 15:05:32.769829 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f2c31eb-4758-44de-8ebc-f76e877cc899" path="/var/lib/kubelet/pods/3f2c31eb-4758-44de-8ebc-f76e877cc899/volumes" Mar 07 15:05:33 crc kubenswrapper[4943]: I0307 15:05:33.181776 4943 generic.go:334] "Generic (PLEG): container finished" podID="28c32ced-1924-491a-add9-496101b1c366" containerID="c9874a40cc2995cba1d7f38c57fdffcac8876054c02bb9c34eb4b93bb8f291cb" exitCode=0 Mar 07 15:05:33 crc kubenswrapper[4943]: I0307 15:05:33.182216 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9rcxq" event={"ID":"28c32ced-1924-491a-add9-496101b1c366","Type":"ContainerDied","Data":"c9874a40cc2995cba1d7f38c57fdffcac8876054c02bb9c34eb4b93bb8f291cb"} Mar 07 15:05:33 crc kubenswrapper[4943]: I0307 15:05:33.754840 4943 scope.go:117] "RemoveContainer" containerID="26f35e1f941cca0c34e3650d7aa047834e769dd76d2a61c3a39cad5bb2d090ae" Mar 07 15:05:33 crc kubenswrapper[4943]: E0307 15:05:33.755142 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cpdmm_openshift-machine-config-operator(4c1f18ea-e557-41fc-95dd-ce4e1722a61e)\"" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" Mar 07 15:05:34 crc kubenswrapper[4943]: I0307 15:05:34.194605 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9rcxq" event={"ID":"28c32ced-1924-491a-add9-496101b1c366","Type":"ContainerStarted","Data":"9d4814d26d6d8ae5aec85b3e458238ebe360e22b4576fdccc766fbb3bce351e2"} Mar 07 15:05:34 crc kubenswrapper[4943]: I0307 15:05:34.224120 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9rcxq" podStartSLOduration=1.74193218 podStartE2EDuration="4.224103532s" podCreationTimestamp="2026-03-07 15:05:30 +0000 UTC" firstStartedPulling="2026-03-07 15:05:31.152354287 +0000 UTC m=+1573.104490785" lastFinishedPulling="2026-03-07 15:05:33.634525629 +0000 UTC m=+1575.586662137" observedRunningTime="2026-03-07 15:05:34.222488832 +0000 UTC m=+1576.174625360" watchObservedRunningTime="2026-03-07 15:05:34.224103532 +0000 UTC m=+1576.176240040" Mar 07 15:05:40 crc kubenswrapper[4943]: I0307 15:05:40.507452 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9rcxq" Mar 07 15:05:40 crc kubenswrapper[4943]: I0307 15:05:40.508192 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9rcxq" Mar 07 15:05:40 crc kubenswrapper[4943]: I0307 15:05:40.576241 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9rcxq" Mar 07 15:05:41 crc kubenswrapper[4943]: I0307 15:05:41.339183 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9rcxq" Mar 07 15:05:41 crc kubenswrapper[4943]: I0307 15:05:41.396633 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9rcxq"] Mar 07 15:05:42 crc kubenswrapper[4943]: I0307 15:05:42.041253 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/keystone-db-create-r2wkc"] Mar 07 15:05:42 crc kubenswrapper[4943]: I0307 15:05:42.051877 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/keystone-db-create-r2wkc"] Mar 07 15:05:42 crc kubenswrapper[4943]: I0307 15:05:42.769848 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d94e16f-d333-4d99-b7e8-441cfb8d7252" path="/var/lib/kubelet/pods/2d94e16f-d333-4d99-b7e8-441cfb8d7252/volumes" Mar 07 15:05:43 crc kubenswrapper[4943]: I0307 15:05:43.034426 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/keystone-c24b-account-create-update-jj69g"] Mar 07 15:05:43 crc kubenswrapper[4943]: I0307 15:05:43.045188 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/keystone-c24b-account-create-update-jj69g"] Mar 07 15:05:43 crc kubenswrapper[4943]: I0307 15:05:43.285726 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9rcxq" podUID="28c32ced-1924-491a-add9-496101b1c366" containerName="registry-server" containerID="cri-o://9d4814d26d6d8ae5aec85b3e458238ebe360e22b4576fdccc766fbb3bce351e2" gracePeriod=2 Mar 07 15:05:43 crc kubenswrapper[4943]: I0307 15:05:43.723141 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9rcxq" Mar 07 15:05:43 crc kubenswrapper[4943]: I0307 15:05:43.869668 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28c32ced-1924-491a-add9-496101b1c366-utilities\") pod \"28c32ced-1924-491a-add9-496101b1c366\" (UID: \"28c32ced-1924-491a-add9-496101b1c366\") " Mar 07 15:05:43 crc kubenswrapper[4943]: I0307 15:05:43.870316 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zwgnv\" (UniqueName: \"kubernetes.io/projected/28c32ced-1924-491a-add9-496101b1c366-kube-api-access-zwgnv\") pod \"28c32ced-1924-491a-add9-496101b1c366\" (UID: \"28c32ced-1924-491a-add9-496101b1c366\") " Mar 07 15:05:43 crc kubenswrapper[4943]: I0307 15:05:43.870389 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28c32ced-1924-491a-add9-496101b1c366-catalog-content\") pod \"28c32ced-1924-491a-add9-496101b1c366\" (UID: \"28c32ced-1924-491a-add9-496101b1c366\") " Mar 07 15:05:43 crc kubenswrapper[4943]: I0307 15:05:43.870767 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/28c32ced-1924-491a-add9-496101b1c366-utilities" (OuterVolumeSpecName: "utilities") pod "28c32ced-1924-491a-add9-496101b1c366" (UID: "28c32ced-1924-491a-add9-496101b1c366"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:05:43 crc kubenswrapper[4943]: I0307 15:05:43.870979 4943 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28c32ced-1924-491a-add9-496101b1c366-utilities\") on node \"crc\" DevicePath \"\"" Mar 07 15:05:43 crc kubenswrapper[4943]: I0307 15:05:43.880501 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28c32ced-1924-491a-add9-496101b1c366-kube-api-access-zwgnv" (OuterVolumeSpecName: "kube-api-access-zwgnv") pod "28c32ced-1924-491a-add9-496101b1c366" (UID: "28c32ced-1924-491a-add9-496101b1c366"). InnerVolumeSpecName "kube-api-access-zwgnv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:05:43 crc kubenswrapper[4943]: I0307 15:05:43.959343 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/28c32ced-1924-491a-add9-496101b1c366-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "28c32ced-1924-491a-add9-496101b1c366" (UID: "28c32ced-1924-491a-add9-496101b1c366"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:05:43 crc kubenswrapper[4943]: I0307 15:05:43.972618 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zwgnv\" (UniqueName: \"kubernetes.io/projected/28c32ced-1924-491a-add9-496101b1c366-kube-api-access-zwgnv\") on node \"crc\" DevicePath \"\"" Mar 07 15:05:43 crc kubenswrapper[4943]: I0307 15:05:43.972677 4943 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28c32ced-1924-491a-add9-496101b1c366-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 07 15:05:44 crc kubenswrapper[4943]: I0307 15:05:44.295954 4943 generic.go:334] "Generic (PLEG): container finished" podID="28c32ced-1924-491a-add9-496101b1c366" containerID="9d4814d26d6d8ae5aec85b3e458238ebe360e22b4576fdccc766fbb3bce351e2" exitCode=0 Mar 07 15:05:44 crc kubenswrapper[4943]: I0307 15:05:44.295998 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9rcxq" event={"ID":"28c32ced-1924-491a-add9-496101b1c366","Type":"ContainerDied","Data":"9d4814d26d6d8ae5aec85b3e458238ebe360e22b4576fdccc766fbb3bce351e2"} Mar 07 15:05:44 crc kubenswrapper[4943]: I0307 15:05:44.296063 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9rcxq" event={"ID":"28c32ced-1924-491a-add9-496101b1c366","Type":"ContainerDied","Data":"65b94629bf32afbddbe9af4e7eda555bfcf3c9b8a9642438a0cac07f0c004c97"} Mar 07 15:05:44 crc kubenswrapper[4943]: I0307 15:05:44.296065 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9rcxq" Mar 07 15:05:44 crc kubenswrapper[4943]: I0307 15:05:44.296092 4943 scope.go:117] "RemoveContainer" containerID="9d4814d26d6d8ae5aec85b3e458238ebe360e22b4576fdccc766fbb3bce351e2" Mar 07 15:05:44 crc kubenswrapper[4943]: I0307 15:05:44.332447 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9rcxq"] Mar 07 15:05:44 crc kubenswrapper[4943]: I0307 15:05:44.335221 4943 scope.go:117] "RemoveContainer" containerID="c9874a40cc2995cba1d7f38c57fdffcac8876054c02bb9c34eb4b93bb8f291cb" Mar 07 15:05:44 crc kubenswrapper[4943]: I0307 15:05:44.344230 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9rcxq"] Mar 07 15:05:44 crc kubenswrapper[4943]: I0307 15:05:44.369745 4943 scope.go:117] "RemoveContainer" containerID="6e7492e44c9a99e0bfa5adab9201d8384619d6eff67e925effa806f9c17bfec7" Mar 07 15:05:44 crc kubenswrapper[4943]: I0307 15:05:44.404830 4943 scope.go:117] "RemoveContainer" containerID="9d4814d26d6d8ae5aec85b3e458238ebe360e22b4576fdccc766fbb3bce351e2" Mar 07 15:05:44 crc kubenswrapper[4943]: E0307 15:05:44.405341 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d4814d26d6d8ae5aec85b3e458238ebe360e22b4576fdccc766fbb3bce351e2\": container with ID starting with 9d4814d26d6d8ae5aec85b3e458238ebe360e22b4576fdccc766fbb3bce351e2 not found: ID does not exist" containerID="9d4814d26d6d8ae5aec85b3e458238ebe360e22b4576fdccc766fbb3bce351e2" Mar 07 15:05:44 crc kubenswrapper[4943]: I0307 15:05:44.405394 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d4814d26d6d8ae5aec85b3e458238ebe360e22b4576fdccc766fbb3bce351e2"} err="failed to get container status \"9d4814d26d6d8ae5aec85b3e458238ebe360e22b4576fdccc766fbb3bce351e2\": rpc error: code = NotFound desc = could not find container \"9d4814d26d6d8ae5aec85b3e458238ebe360e22b4576fdccc766fbb3bce351e2\": container with ID starting with 9d4814d26d6d8ae5aec85b3e458238ebe360e22b4576fdccc766fbb3bce351e2 not found: ID does not exist" Mar 07 15:05:44 crc kubenswrapper[4943]: I0307 15:05:44.405423 4943 scope.go:117] "RemoveContainer" containerID="c9874a40cc2995cba1d7f38c57fdffcac8876054c02bb9c34eb4b93bb8f291cb" Mar 07 15:05:44 crc kubenswrapper[4943]: E0307 15:05:44.405723 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9874a40cc2995cba1d7f38c57fdffcac8876054c02bb9c34eb4b93bb8f291cb\": container with ID starting with c9874a40cc2995cba1d7f38c57fdffcac8876054c02bb9c34eb4b93bb8f291cb not found: ID does not exist" containerID="c9874a40cc2995cba1d7f38c57fdffcac8876054c02bb9c34eb4b93bb8f291cb" Mar 07 15:05:44 crc kubenswrapper[4943]: I0307 15:05:44.405841 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9874a40cc2995cba1d7f38c57fdffcac8876054c02bb9c34eb4b93bb8f291cb"} err="failed to get container status \"c9874a40cc2995cba1d7f38c57fdffcac8876054c02bb9c34eb4b93bb8f291cb\": rpc error: code = NotFound desc = could not find container \"c9874a40cc2995cba1d7f38c57fdffcac8876054c02bb9c34eb4b93bb8f291cb\": container with ID starting with c9874a40cc2995cba1d7f38c57fdffcac8876054c02bb9c34eb4b93bb8f291cb not found: ID does not exist" Mar 07 15:05:44 crc kubenswrapper[4943]: I0307 15:05:44.405943 4943 scope.go:117] "RemoveContainer" containerID="6e7492e44c9a99e0bfa5adab9201d8384619d6eff67e925effa806f9c17bfec7" Mar 07 15:05:44 crc kubenswrapper[4943]: E0307 15:05:44.406477 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e7492e44c9a99e0bfa5adab9201d8384619d6eff67e925effa806f9c17bfec7\": container with ID starting with 6e7492e44c9a99e0bfa5adab9201d8384619d6eff67e925effa806f9c17bfec7 not found: ID does not exist" containerID="6e7492e44c9a99e0bfa5adab9201d8384619d6eff67e925effa806f9c17bfec7" Mar 07 15:05:44 crc kubenswrapper[4943]: I0307 15:05:44.406498 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e7492e44c9a99e0bfa5adab9201d8384619d6eff67e925effa806f9c17bfec7"} err="failed to get container status \"6e7492e44c9a99e0bfa5adab9201d8384619d6eff67e925effa806f9c17bfec7\": rpc error: code = NotFound desc = could not find container \"6e7492e44c9a99e0bfa5adab9201d8384619d6eff67e925effa806f9c17bfec7\": container with ID starting with 6e7492e44c9a99e0bfa5adab9201d8384619d6eff67e925effa806f9c17bfec7 not found: ID does not exist" Mar 07 15:05:44 crc kubenswrapper[4943]: I0307 15:05:44.755587 4943 scope.go:117] "RemoveContainer" containerID="26f35e1f941cca0c34e3650d7aa047834e769dd76d2a61c3a39cad5bb2d090ae" Mar 07 15:05:44 crc kubenswrapper[4943]: E0307 15:05:44.756147 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cpdmm_openshift-machine-config-operator(4c1f18ea-e557-41fc-95dd-ce4e1722a61e)\"" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" Mar 07 15:05:44 crc kubenswrapper[4943]: I0307 15:05:44.772678 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28c32ced-1924-491a-add9-496101b1c366" path="/var/lib/kubelet/pods/28c32ced-1924-491a-add9-496101b1c366/volumes" Mar 07 15:05:44 crc kubenswrapper[4943]: I0307 15:05:44.774026 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e5f2081-60f4-4af6-ab3f-50ba7ba11c6c" path="/var/lib/kubelet/pods/8e5f2081-60f4-4af6-ab3f-50ba7ba11c6c/volumes" Mar 07 15:05:58 crc kubenswrapper[4943]: I0307 15:05:58.045665 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/keystone-db-sync-nq8hl"] Mar 07 15:05:58 crc kubenswrapper[4943]: I0307 15:05:58.055521 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/keystone-db-sync-nq8hl"] Mar 07 15:05:58 crc kubenswrapper[4943]: I0307 15:05:58.762420 4943 scope.go:117] "RemoveContainer" containerID="26f35e1f941cca0c34e3650d7aa047834e769dd76d2a61c3a39cad5bb2d090ae" Mar 07 15:05:58 crc kubenswrapper[4943]: E0307 15:05:58.762801 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cpdmm_openshift-machine-config-operator(4c1f18ea-e557-41fc-95dd-ce4e1722a61e)\"" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" Mar 07 15:05:58 crc kubenswrapper[4943]: I0307 15:05:58.769407 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea23fd1e-8c76-4483-a624-be13a1715513" path="/var/lib/kubelet/pods/ea23fd1e-8c76-4483-a624-be13a1715513/volumes" Mar 07 15:06:00 crc kubenswrapper[4943]: I0307 15:06:00.160249 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29548266-tw972"] Mar 07 15:06:00 crc kubenswrapper[4943]: E0307 15:06:00.160757 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28c32ced-1924-491a-add9-496101b1c366" containerName="extract-content" Mar 07 15:06:00 crc kubenswrapper[4943]: I0307 15:06:00.160781 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="28c32ced-1924-491a-add9-496101b1c366" containerName="extract-content" Mar 07 15:06:00 crc kubenswrapper[4943]: E0307 15:06:00.160805 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f2c31eb-4758-44de-8ebc-f76e877cc899" containerName="extract-content" Mar 07 15:06:00 crc kubenswrapper[4943]: I0307 15:06:00.160819 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f2c31eb-4758-44de-8ebc-f76e877cc899" containerName="extract-content" Mar 07 15:06:00 crc kubenswrapper[4943]: E0307 15:06:00.160860 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28c32ced-1924-491a-add9-496101b1c366" containerName="extract-utilities" Mar 07 15:06:00 crc kubenswrapper[4943]: I0307 15:06:00.160873 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="28c32ced-1924-491a-add9-496101b1c366" containerName="extract-utilities" Mar 07 15:06:00 crc kubenswrapper[4943]: E0307 15:06:00.160896 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f2c31eb-4758-44de-8ebc-f76e877cc899" containerName="registry-server" Mar 07 15:06:00 crc kubenswrapper[4943]: I0307 15:06:00.160909 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f2c31eb-4758-44de-8ebc-f76e877cc899" containerName="registry-server" Mar 07 15:06:00 crc kubenswrapper[4943]: E0307 15:06:00.160948 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28c32ced-1924-491a-add9-496101b1c366" containerName="registry-server" Mar 07 15:06:00 crc kubenswrapper[4943]: I0307 15:06:00.160962 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="28c32ced-1924-491a-add9-496101b1c366" containerName="registry-server" Mar 07 15:06:00 crc kubenswrapper[4943]: E0307 15:06:00.160999 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f2c31eb-4758-44de-8ebc-f76e877cc899" containerName="extract-utilities" Mar 07 15:06:00 crc kubenswrapper[4943]: I0307 15:06:00.161013 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f2c31eb-4758-44de-8ebc-f76e877cc899" containerName="extract-utilities" Mar 07 15:06:00 crc kubenswrapper[4943]: I0307 15:06:00.161305 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f2c31eb-4758-44de-8ebc-f76e877cc899" containerName="registry-server" Mar 07 15:06:00 crc kubenswrapper[4943]: I0307 15:06:00.161339 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="28c32ced-1924-491a-add9-496101b1c366" containerName="registry-server" Mar 07 15:06:00 crc kubenswrapper[4943]: I0307 15:06:00.162164 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548266-tw972" Mar 07 15:06:00 crc kubenswrapper[4943]: I0307 15:06:00.164309 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 07 15:06:00 crc kubenswrapper[4943]: I0307 15:06:00.165253 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 07 15:06:00 crc kubenswrapper[4943]: I0307 15:06:00.168463 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29548266-tw972"] Mar 07 15:06:00 crc kubenswrapper[4943]: I0307 15:06:00.169868 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-qqlk5" Mar 07 15:06:00 crc kubenswrapper[4943]: I0307 15:06:00.364025 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97m2j\" (UniqueName: \"kubernetes.io/projected/ed54b4ee-fddf-44f5-928a-326405fac246-kube-api-access-97m2j\") pod \"auto-csr-approver-29548266-tw972\" (UID: \"ed54b4ee-fddf-44f5-928a-326405fac246\") " pod="openshift-infra/auto-csr-approver-29548266-tw972" Mar 07 15:06:00 crc kubenswrapper[4943]: I0307 15:06:00.465991 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97m2j\" (UniqueName: \"kubernetes.io/projected/ed54b4ee-fddf-44f5-928a-326405fac246-kube-api-access-97m2j\") pod \"auto-csr-approver-29548266-tw972\" (UID: \"ed54b4ee-fddf-44f5-928a-326405fac246\") " pod="openshift-infra/auto-csr-approver-29548266-tw972" Mar 07 15:06:00 crc kubenswrapper[4943]: I0307 15:06:00.503906 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97m2j\" (UniqueName: \"kubernetes.io/projected/ed54b4ee-fddf-44f5-928a-326405fac246-kube-api-access-97m2j\") pod \"auto-csr-approver-29548266-tw972\" (UID: \"ed54b4ee-fddf-44f5-928a-326405fac246\") " pod="openshift-infra/auto-csr-approver-29548266-tw972" Mar 07 15:06:00 crc kubenswrapper[4943]: I0307 15:06:00.803741 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548266-tw972" Mar 07 15:06:01 crc kubenswrapper[4943]: I0307 15:06:01.081839 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29548266-tw972"] Mar 07 15:06:01 crc kubenswrapper[4943]: I0307 15:06:01.469480 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548266-tw972" event={"ID":"ed54b4ee-fddf-44f5-928a-326405fac246","Type":"ContainerStarted","Data":"342d1bf3b978dacdacea85fd8771eb95038e8f7f51f8aaa763b72be419ef54da"} Mar 07 15:06:02 crc kubenswrapper[4943]: I0307 15:06:02.495806 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548266-tw972" event={"ID":"ed54b4ee-fddf-44f5-928a-326405fac246","Type":"ContainerStarted","Data":"05fe59cccad959e643fc20072c6b5a061c26fac692faa569a5c6d8519505b13d"} Mar 07 15:06:02 crc kubenswrapper[4943]: I0307 15:06:02.527013 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29548266-tw972" podStartSLOduration=1.6972935630000001 podStartE2EDuration="2.52699022s" podCreationTimestamp="2026-03-07 15:06:00 +0000 UTC" firstStartedPulling="2026-03-07 15:06:01.08981692 +0000 UTC m=+1603.041953418" lastFinishedPulling="2026-03-07 15:06:01.919513537 +0000 UTC m=+1603.871650075" observedRunningTime="2026-03-07 15:06:02.521138396 +0000 UTC m=+1604.473274934" watchObservedRunningTime="2026-03-07 15:06:02.52699022 +0000 UTC m=+1604.479126758" Mar 07 15:06:03 crc kubenswrapper[4943]: I0307 15:06:03.508660 4943 generic.go:334] "Generic (PLEG): container finished" podID="ed54b4ee-fddf-44f5-928a-326405fac246" containerID="05fe59cccad959e643fc20072c6b5a061c26fac692faa569a5c6d8519505b13d" exitCode=0 Mar 07 15:06:03 crc kubenswrapper[4943]: I0307 15:06:03.508729 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548266-tw972" event={"ID":"ed54b4ee-fddf-44f5-928a-326405fac246","Type":"ContainerDied","Data":"05fe59cccad959e643fc20072c6b5a061c26fac692faa569a5c6d8519505b13d"} Mar 07 15:06:04 crc kubenswrapper[4943]: I0307 15:06:04.902651 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548266-tw972" Mar 07 15:06:04 crc kubenswrapper[4943]: I0307 15:06:04.961190 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97m2j\" (UniqueName: \"kubernetes.io/projected/ed54b4ee-fddf-44f5-928a-326405fac246-kube-api-access-97m2j\") pod \"ed54b4ee-fddf-44f5-928a-326405fac246\" (UID: \"ed54b4ee-fddf-44f5-928a-326405fac246\") " Mar 07 15:06:04 crc kubenswrapper[4943]: I0307 15:06:04.973487 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed54b4ee-fddf-44f5-928a-326405fac246-kube-api-access-97m2j" (OuterVolumeSpecName: "kube-api-access-97m2j") pod "ed54b4ee-fddf-44f5-928a-326405fac246" (UID: "ed54b4ee-fddf-44f5-928a-326405fac246"). InnerVolumeSpecName "kube-api-access-97m2j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:06:05 crc kubenswrapper[4943]: I0307 15:06:05.045202 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/keystone-bootstrap-fz942"] Mar 07 15:06:05 crc kubenswrapper[4943]: I0307 15:06:05.058207 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/keystone-bootstrap-fz942"] Mar 07 15:06:05 crc kubenswrapper[4943]: I0307 15:06:05.063057 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97m2j\" (UniqueName: \"kubernetes.io/projected/ed54b4ee-fddf-44f5-928a-326405fac246-kube-api-access-97m2j\") on node \"crc\" DevicePath \"\"" Mar 07 15:06:05 crc kubenswrapper[4943]: I0307 15:06:05.527700 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548266-tw972" event={"ID":"ed54b4ee-fddf-44f5-928a-326405fac246","Type":"ContainerDied","Data":"342d1bf3b978dacdacea85fd8771eb95038e8f7f51f8aaa763b72be419ef54da"} Mar 07 15:06:05 crc kubenswrapper[4943]: I0307 15:06:05.528183 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="342d1bf3b978dacdacea85fd8771eb95038e8f7f51f8aaa763b72be419ef54da" Mar 07 15:06:05 crc kubenswrapper[4943]: I0307 15:06:05.527783 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548266-tw972" Mar 07 15:06:05 crc kubenswrapper[4943]: I0307 15:06:05.606650 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29548260-ljrss"] Mar 07 15:06:05 crc kubenswrapper[4943]: I0307 15:06:05.615834 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29548260-ljrss"] Mar 07 15:06:06 crc kubenswrapper[4943]: I0307 15:06:06.770827 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2bbf9dc7-77bd-470f-8f33-3b62d014c239" path="/var/lib/kubelet/pods/2bbf9dc7-77bd-470f-8f33-3b62d014c239/volumes" Mar 07 15:06:06 crc kubenswrapper[4943]: I0307 15:06:06.771538 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="601960fc-b3bc-4a62-9843-fd22ba6a1850" path="/var/lib/kubelet/pods/601960fc-b3bc-4a62-9843-fd22ba6a1850/volumes" Mar 07 15:06:12 crc kubenswrapper[4943]: I0307 15:06:12.335493 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-48nrq"] Mar 07 15:06:12 crc kubenswrapper[4943]: E0307 15:06:12.336550 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed54b4ee-fddf-44f5-928a-326405fac246" containerName="oc" Mar 07 15:06:12 crc kubenswrapper[4943]: I0307 15:06:12.336573 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed54b4ee-fddf-44f5-928a-326405fac246" containerName="oc" Mar 07 15:06:12 crc kubenswrapper[4943]: I0307 15:06:12.336821 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed54b4ee-fddf-44f5-928a-326405fac246" containerName="oc" Mar 07 15:06:12 crc kubenswrapper[4943]: I0307 15:06:12.338767 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-48nrq" Mar 07 15:06:12 crc kubenswrapper[4943]: I0307 15:06:12.356163 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-48nrq"] Mar 07 15:06:12 crc kubenswrapper[4943]: I0307 15:06:12.424512 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f215184-f0b3-420c-9819-675e4d3a555b-utilities\") pod \"redhat-operators-48nrq\" (UID: \"0f215184-f0b3-420c-9819-675e4d3a555b\") " pod="openshift-marketplace/redhat-operators-48nrq" Mar 07 15:06:12 crc kubenswrapper[4943]: I0307 15:06:12.424590 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f215184-f0b3-420c-9819-675e4d3a555b-catalog-content\") pod \"redhat-operators-48nrq\" (UID: \"0f215184-f0b3-420c-9819-675e4d3a555b\") " pod="openshift-marketplace/redhat-operators-48nrq" Mar 07 15:06:12 crc kubenswrapper[4943]: I0307 15:06:12.424645 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tt7qk\" (UniqueName: \"kubernetes.io/projected/0f215184-f0b3-420c-9819-675e4d3a555b-kube-api-access-tt7qk\") pod \"redhat-operators-48nrq\" (UID: \"0f215184-f0b3-420c-9819-675e4d3a555b\") " pod="openshift-marketplace/redhat-operators-48nrq" Mar 07 15:06:12 crc kubenswrapper[4943]: I0307 15:06:12.526398 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f215184-f0b3-420c-9819-675e4d3a555b-utilities\") pod \"redhat-operators-48nrq\" (UID: \"0f215184-f0b3-420c-9819-675e4d3a555b\") " pod="openshift-marketplace/redhat-operators-48nrq" Mar 07 15:06:12 crc kubenswrapper[4943]: I0307 15:06:12.526814 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f215184-f0b3-420c-9819-675e4d3a555b-catalog-content\") pod \"redhat-operators-48nrq\" (UID: \"0f215184-f0b3-420c-9819-675e4d3a555b\") " pod="openshift-marketplace/redhat-operators-48nrq" Mar 07 15:06:12 crc kubenswrapper[4943]: I0307 15:06:12.526866 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tt7qk\" (UniqueName: \"kubernetes.io/projected/0f215184-f0b3-420c-9819-675e4d3a555b-kube-api-access-tt7qk\") pod \"redhat-operators-48nrq\" (UID: \"0f215184-f0b3-420c-9819-675e4d3a555b\") " pod="openshift-marketplace/redhat-operators-48nrq" Mar 07 15:06:12 crc kubenswrapper[4943]: I0307 15:06:12.527116 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f215184-f0b3-420c-9819-675e4d3a555b-utilities\") pod \"redhat-operators-48nrq\" (UID: \"0f215184-f0b3-420c-9819-675e4d3a555b\") " pod="openshift-marketplace/redhat-operators-48nrq" Mar 07 15:06:12 crc kubenswrapper[4943]: I0307 15:06:12.527336 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f215184-f0b3-420c-9819-675e4d3a555b-catalog-content\") pod \"redhat-operators-48nrq\" (UID: \"0f215184-f0b3-420c-9819-675e4d3a555b\") " pod="openshift-marketplace/redhat-operators-48nrq" Mar 07 15:06:12 crc kubenswrapper[4943]: I0307 15:06:12.554548 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tt7qk\" (UniqueName: \"kubernetes.io/projected/0f215184-f0b3-420c-9819-675e4d3a555b-kube-api-access-tt7qk\") pod \"redhat-operators-48nrq\" (UID: \"0f215184-f0b3-420c-9819-675e4d3a555b\") " pod="openshift-marketplace/redhat-operators-48nrq" Mar 07 15:06:12 crc kubenswrapper[4943]: I0307 15:06:12.660297 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-48nrq" Mar 07 15:06:12 crc kubenswrapper[4943]: I0307 15:06:12.755843 4943 scope.go:117] "RemoveContainer" containerID="26f35e1f941cca0c34e3650d7aa047834e769dd76d2a61c3a39cad5bb2d090ae" Mar 07 15:06:12 crc kubenswrapper[4943]: E0307 15:06:12.756155 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cpdmm_openshift-machine-config-operator(4c1f18ea-e557-41fc-95dd-ce4e1722a61e)\"" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" Mar 07 15:06:13 crc kubenswrapper[4943]: I0307 15:06:13.130031 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-48nrq"] Mar 07 15:06:13 crc kubenswrapper[4943]: I0307 15:06:13.603774 4943 generic.go:334] "Generic (PLEG): container finished" podID="0f215184-f0b3-420c-9819-675e4d3a555b" containerID="a4347c3a48b1b9983989ff0a8b560c4af2e4b53570d9f37d823332425732ebce" exitCode=0 Mar 07 15:06:13 crc kubenswrapper[4943]: I0307 15:06:13.603814 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-48nrq" event={"ID":"0f215184-f0b3-420c-9819-675e4d3a555b","Type":"ContainerDied","Data":"a4347c3a48b1b9983989ff0a8b560c4af2e4b53570d9f37d823332425732ebce"} Mar 07 15:06:13 crc kubenswrapper[4943]: I0307 15:06:13.603872 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-48nrq" event={"ID":"0f215184-f0b3-420c-9819-675e4d3a555b","Type":"ContainerStarted","Data":"060b1774416c6a2946434d9d7ce74aacfa92102b1291499db8c8aff6ab42a796"} Mar 07 15:06:15 crc kubenswrapper[4943]: I0307 15:06:15.628231 4943 generic.go:334] "Generic (PLEG): container finished" podID="0f215184-f0b3-420c-9819-675e4d3a555b" containerID="fc4ec60e37643467cc68f92f442ba7a5dd0bb191e30bab70fcd023ee6b670ab3" exitCode=0 Mar 07 15:06:15 crc kubenswrapper[4943]: I0307 15:06:15.628294 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-48nrq" event={"ID":"0f215184-f0b3-420c-9819-675e4d3a555b","Type":"ContainerDied","Data":"fc4ec60e37643467cc68f92f442ba7a5dd0bb191e30bab70fcd023ee6b670ab3"} Mar 07 15:06:16 crc kubenswrapper[4943]: I0307 15:06:16.639635 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-48nrq" event={"ID":"0f215184-f0b3-420c-9819-675e4d3a555b","Type":"ContainerStarted","Data":"1c8813df60c38ec7ffff365cc223c935a33ab9ee4c707e514123328a6948c305"} Mar 07 15:06:16 crc kubenswrapper[4943]: I0307 15:06:16.661085 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-48nrq" podStartSLOduration=2.229790734 podStartE2EDuration="4.661062227s" podCreationTimestamp="2026-03-07 15:06:12 +0000 UTC" firstStartedPulling="2026-03-07 15:06:13.605776935 +0000 UTC m=+1615.557913443" lastFinishedPulling="2026-03-07 15:06:16.037048398 +0000 UTC m=+1617.989184936" observedRunningTime="2026-03-07 15:06:16.659892288 +0000 UTC m=+1618.612028826" watchObservedRunningTime="2026-03-07 15:06:16.661062227 +0000 UTC m=+1618.613198755" Mar 07 15:06:20 crc kubenswrapper[4943]: I0307 15:06:20.530496 4943 scope.go:117] "RemoveContainer" containerID="2d4a24d7450c798d91a834f2b32ef0ddd47a854c5e4a00328222985152149372" Mar 07 15:06:20 crc kubenswrapper[4943]: I0307 15:06:20.586262 4943 scope.go:117] "RemoveContainer" containerID="e10b1432f5f9b11449cf70b1dcf772f9002fc6d36fe8ecbcfc87b27610e44016" Mar 07 15:06:20 crc kubenswrapper[4943]: I0307 15:06:20.633158 4943 scope.go:117] "RemoveContainer" containerID="7868657e9c49897c842757eadb44bc3eac4bcdd4e427e2a8d8b43b9a3362e40c" Mar 07 15:06:20 crc kubenswrapper[4943]: I0307 15:06:20.649766 4943 scope.go:117] "RemoveContainer" containerID="3249b974d903bb8bc3583a79f36e4779b8523fc5199d38645e56bcc5bcbec1d1" Mar 07 15:06:20 crc kubenswrapper[4943]: I0307 15:06:20.685128 4943 scope.go:117] "RemoveContainer" containerID="0f58432047df35f921d354c24be9618c04c8812420d1943928b1bb6ef760be7a" Mar 07 15:06:20 crc kubenswrapper[4943]: I0307 15:06:20.724402 4943 scope.go:117] "RemoveContainer" containerID="cb632251025997ee50958ee0c15255ead5450b4159b77c62918a7990f1ad1002" Mar 07 15:06:22 crc kubenswrapper[4943]: I0307 15:06:22.660766 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-48nrq" Mar 07 15:06:22 crc kubenswrapper[4943]: I0307 15:06:22.661293 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-48nrq" Mar 07 15:06:23 crc kubenswrapper[4943]: I0307 15:06:23.719155 4943 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-48nrq" podUID="0f215184-f0b3-420c-9819-675e4d3a555b" containerName="registry-server" probeResult="failure" output=< Mar 07 15:06:23 crc kubenswrapper[4943]: timeout: failed to connect service ":50051" within 1s Mar 07 15:06:23 crc kubenswrapper[4943]: > Mar 07 15:06:24 crc kubenswrapper[4943]: I0307 15:06:24.043048 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/barbican-db-create-52f2b"] Mar 07 15:06:24 crc kubenswrapper[4943]: I0307 15:06:24.056409 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/barbican-f3ae-account-create-update-5pbbk"] Mar 07 15:06:24 crc kubenswrapper[4943]: I0307 15:06:24.069041 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/barbican-db-create-52f2b"] Mar 07 15:06:24 crc kubenswrapper[4943]: I0307 15:06:24.081702 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/barbican-f3ae-account-create-update-5pbbk"] Mar 07 15:06:24 crc kubenswrapper[4943]: I0307 15:06:24.770248 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc429dfa-6f9e-4715-9eb7-daccbbad62b6" path="/var/lib/kubelet/pods/dc429dfa-6f9e-4715-9eb7-daccbbad62b6/volumes" Mar 07 15:06:24 crc kubenswrapper[4943]: I0307 15:06:24.771396 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2bd919a-7322-446b-82e3-dab28d834f96" path="/var/lib/kubelet/pods/f2bd919a-7322-446b-82e3-dab28d834f96/volumes" Mar 07 15:06:26 crc kubenswrapper[4943]: I0307 15:06:26.756836 4943 scope.go:117] "RemoveContainer" containerID="26f35e1f941cca0c34e3650d7aa047834e769dd76d2a61c3a39cad5bb2d090ae" Mar 07 15:06:26 crc kubenswrapper[4943]: E0307 15:06:26.757596 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cpdmm_openshift-machine-config-operator(4c1f18ea-e557-41fc-95dd-ce4e1722a61e)\"" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" Mar 07 15:06:27 crc kubenswrapper[4943]: E0307 15:06:27.268336 4943 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.75:57858->38.102.83.75:36531: write tcp 38.102.83.75:57858->38.102.83.75:36531: write: broken pipe Mar 07 15:06:32 crc kubenswrapper[4943]: I0307 15:06:32.735697 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-48nrq" Mar 07 15:06:32 crc kubenswrapper[4943]: I0307 15:06:32.813790 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-48nrq" Mar 07 15:06:32 crc kubenswrapper[4943]: I0307 15:06:32.974840 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-48nrq"] Mar 07 15:06:33 crc kubenswrapper[4943]: I0307 15:06:33.835836 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-48nrq" podUID="0f215184-f0b3-420c-9819-675e4d3a555b" containerName="registry-server" containerID="cri-o://1c8813df60c38ec7ffff365cc223c935a33ab9ee4c707e514123328a6948c305" gracePeriod=2 Mar 07 15:06:34 crc kubenswrapper[4943]: I0307 15:06:34.286003 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-48nrq" Mar 07 15:06:34 crc kubenswrapper[4943]: I0307 15:06:34.409730 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f215184-f0b3-420c-9819-675e4d3a555b-utilities\") pod \"0f215184-f0b3-420c-9819-675e4d3a555b\" (UID: \"0f215184-f0b3-420c-9819-675e4d3a555b\") " Mar 07 15:06:34 crc kubenswrapper[4943]: I0307 15:06:34.410018 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tt7qk\" (UniqueName: \"kubernetes.io/projected/0f215184-f0b3-420c-9819-675e4d3a555b-kube-api-access-tt7qk\") pod \"0f215184-f0b3-420c-9819-675e4d3a555b\" (UID: \"0f215184-f0b3-420c-9819-675e4d3a555b\") " Mar 07 15:06:34 crc kubenswrapper[4943]: I0307 15:06:34.410100 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f215184-f0b3-420c-9819-675e4d3a555b-catalog-content\") pod \"0f215184-f0b3-420c-9819-675e4d3a555b\" (UID: \"0f215184-f0b3-420c-9819-675e4d3a555b\") " Mar 07 15:06:34 crc kubenswrapper[4943]: I0307 15:06:34.411094 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f215184-f0b3-420c-9819-675e4d3a555b-utilities" (OuterVolumeSpecName: "utilities") pod "0f215184-f0b3-420c-9819-675e4d3a555b" (UID: "0f215184-f0b3-420c-9819-675e4d3a555b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:06:34 crc kubenswrapper[4943]: I0307 15:06:34.415906 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f215184-f0b3-420c-9819-675e4d3a555b-kube-api-access-tt7qk" (OuterVolumeSpecName: "kube-api-access-tt7qk") pod "0f215184-f0b3-420c-9819-675e4d3a555b" (UID: "0f215184-f0b3-420c-9819-675e4d3a555b"). InnerVolumeSpecName "kube-api-access-tt7qk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:06:34 crc kubenswrapper[4943]: I0307 15:06:34.512400 4943 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f215184-f0b3-420c-9819-675e4d3a555b-utilities\") on node \"crc\" DevicePath \"\"" Mar 07 15:06:34 crc kubenswrapper[4943]: I0307 15:06:34.512459 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tt7qk\" (UniqueName: \"kubernetes.io/projected/0f215184-f0b3-420c-9819-675e4d3a555b-kube-api-access-tt7qk\") on node \"crc\" DevicePath \"\"" Mar 07 15:06:34 crc kubenswrapper[4943]: I0307 15:06:34.585553 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f215184-f0b3-420c-9819-675e4d3a555b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0f215184-f0b3-420c-9819-675e4d3a555b" (UID: "0f215184-f0b3-420c-9819-675e4d3a555b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:06:34 crc kubenswrapper[4943]: I0307 15:06:34.615847 4943 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f215184-f0b3-420c-9819-675e4d3a555b-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 07 15:06:34 crc kubenswrapper[4943]: I0307 15:06:34.864456 4943 generic.go:334] "Generic (PLEG): container finished" podID="0f215184-f0b3-420c-9819-675e4d3a555b" containerID="1c8813df60c38ec7ffff365cc223c935a33ab9ee4c707e514123328a6948c305" exitCode=0 Mar 07 15:06:34 crc kubenswrapper[4943]: I0307 15:06:34.864685 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-48nrq" event={"ID":"0f215184-f0b3-420c-9819-675e4d3a555b","Type":"ContainerDied","Data":"1c8813df60c38ec7ffff365cc223c935a33ab9ee4c707e514123328a6948c305"} Mar 07 15:06:34 crc kubenswrapper[4943]: I0307 15:06:34.864729 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-48nrq" event={"ID":"0f215184-f0b3-420c-9819-675e4d3a555b","Type":"ContainerDied","Data":"060b1774416c6a2946434d9d7ce74aacfa92102b1291499db8c8aff6ab42a796"} Mar 07 15:06:34 crc kubenswrapper[4943]: I0307 15:06:34.864765 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-48nrq" Mar 07 15:06:34 crc kubenswrapper[4943]: I0307 15:06:34.865221 4943 scope.go:117] "RemoveContainer" containerID="1c8813df60c38ec7ffff365cc223c935a33ab9ee4c707e514123328a6948c305" Mar 07 15:06:34 crc kubenswrapper[4943]: I0307 15:06:34.910921 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-48nrq"] Mar 07 15:06:34 crc kubenswrapper[4943]: I0307 15:06:34.918983 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-48nrq"] Mar 07 15:06:34 crc kubenswrapper[4943]: I0307 15:06:34.921681 4943 scope.go:117] "RemoveContainer" containerID="fc4ec60e37643467cc68f92f442ba7a5dd0bb191e30bab70fcd023ee6b670ab3" Mar 07 15:06:34 crc kubenswrapper[4943]: I0307 15:06:34.941672 4943 scope.go:117] "RemoveContainer" containerID="a4347c3a48b1b9983989ff0a8b560c4af2e4b53570d9f37d823332425732ebce" Mar 07 15:06:34 crc kubenswrapper[4943]: I0307 15:06:34.969496 4943 scope.go:117] "RemoveContainer" containerID="1c8813df60c38ec7ffff365cc223c935a33ab9ee4c707e514123328a6948c305" Mar 07 15:06:34 crc kubenswrapper[4943]: E0307 15:06:34.969839 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c8813df60c38ec7ffff365cc223c935a33ab9ee4c707e514123328a6948c305\": container with ID starting with 1c8813df60c38ec7ffff365cc223c935a33ab9ee4c707e514123328a6948c305 not found: ID does not exist" containerID="1c8813df60c38ec7ffff365cc223c935a33ab9ee4c707e514123328a6948c305" Mar 07 15:06:34 crc kubenswrapper[4943]: I0307 15:06:34.969872 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c8813df60c38ec7ffff365cc223c935a33ab9ee4c707e514123328a6948c305"} err="failed to get container status \"1c8813df60c38ec7ffff365cc223c935a33ab9ee4c707e514123328a6948c305\": rpc error: code = NotFound desc = could not find container \"1c8813df60c38ec7ffff365cc223c935a33ab9ee4c707e514123328a6948c305\": container with ID starting with 1c8813df60c38ec7ffff365cc223c935a33ab9ee4c707e514123328a6948c305 not found: ID does not exist" Mar 07 15:06:34 crc kubenswrapper[4943]: I0307 15:06:34.969908 4943 scope.go:117] "RemoveContainer" containerID="fc4ec60e37643467cc68f92f442ba7a5dd0bb191e30bab70fcd023ee6b670ab3" Mar 07 15:06:34 crc kubenswrapper[4943]: E0307 15:06:34.970252 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc4ec60e37643467cc68f92f442ba7a5dd0bb191e30bab70fcd023ee6b670ab3\": container with ID starting with fc4ec60e37643467cc68f92f442ba7a5dd0bb191e30bab70fcd023ee6b670ab3 not found: ID does not exist" containerID="fc4ec60e37643467cc68f92f442ba7a5dd0bb191e30bab70fcd023ee6b670ab3" Mar 07 15:06:34 crc kubenswrapper[4943]: I0307 15:06:34.970283 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc4ec60e37643467cc68f92f442ba7a5dd0bb191e30bab70fcd023ee6b670ab3"} err="failed to get container status \"fc4ec60e37643467cc68f92f442ba7a5dd0bb191e30bab70fcd023ee6b670ab3\": rpc error: code = NotFound desc = could not find container \"fc4ec60e37643467cc68f92f442ba7a5dd0bb191e30bab70fcd023ee6b670ab3\": container with ID starting with fc4ec60e37643467cc68f92f442ba7a5dd0bb191e30bab70fcd023ee6b670ab3 not found: ID does not exist" Mar 07 15:06:34 crc kubenswrapper[4943]: I0307 15:06:34.970302 4943 scope.go:117] "RemoveContainer" containerID="a4347c3a48b1b9983989ff0a8b560c4af2e4b53570d9f37d823332425732ebce" Mar 07 15:06:34 crc kubenswrapper[4943]: E0307 15:06:34.970605 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4347c3a48b1b9983989ff0a8b560c4af2e4b53570d9f37d823332425732ebce\": container with ID starting with a4347c3a48b1b9983989ff0a8b560c4af2e4b53570d9f37d823332425732ebce not found: ID does not exist" containerID="a4347c3a48b1b9983989ff0a8b560c4af2e4b53570d9f37d823332425732ebce" Mar 07 15:06:34 crc kubenswrapper[4943]: I0307 15:06:34.970629 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4347c3a48b1b9983989ff0a8b560c4af2e4b53570d9f37d823332425732ebce"} err="failed to get container status \"a4347c3a48b1b9983989ff0a8b560c4af2e4b53570d9f37d823332425732ebce\": rpc error: code = NotFound desc = could not find container \"a4347c3a48b1b9983989ff0a8b560c4af2e4b53570d9f37d823332425732ebce\": container with ID starting with a4347c3a48b1b9983989ff0a8b560c4af2e4b53570d9f37d823332425732ebce not found: ID does not exist" Mar 07 15:06:36 crc kubenswrapper[4943]: I0307 15:06:36.773854 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f215184-f0b3-420c-9819-675e4d3a555b" path="/var/lib/kubelet/pods/0f215184-f0b3-420c-9819-675e4d3a555b/volumes" Mar 07 15:06:38 crc kubenswrapper[4943]: I0307 15:06:38.763603 4943 scope.go:117] "RemoveContainer" containerID="26f35e1f941cca0c34e3650d7aa047834e769dd76d2a61c3a39cad5bb2d090ae" Mar 07 15:06:38 crc kubenswrapper[4943]: E0307 15:06:38.764478 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cpdmm_openshift-machine-config-operator(4c1f18ea-e557-41fc-95dd-ce4e1722a61e)\"" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" Mar 07 15:06:51 crc kubenswrapper[4943]: I0307 15:06:51.756142 4943 scope.go:117] "RemoveContainer" containerID="26f35e1f941cca0c34e3650d7aa047834e769dd76d2a61c3a39cad5bb2d090ae" Mar 07 15:06:51 crc kubenswrapper[4943]: E0307 15:06:51.757475 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cpdmm_openshift-machine-config-operator(4c1f18ea-e557-41fc-95dd-ce4e1722a61e)\"" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" Mar 07 15:07:05 crc kubenswrapper[4943]: I0307 15:07:05.754824 4943 scope.go:117] "RemoveContainer" containerID="26f35e1f941cca0c34e3650d7aa047834e769dd76d2a61c3a39cad5bb2d090ae" Mar 07 15:07:05 crc kubenswrapper[4943]: E0307 15:07:05.755581 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cpdmm_openshift-machine-config-operator(4c1f18ea-e557-41fc-95dd-ce4e1722a61e)\"" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" Mar 07 15:07:17 crc kubenswrapper[4943]: I0307 15:07:17.755518 4943 scope.go:117] "RemoveContainer" containerID="26f35e1f941cca0c34e3650d7aa047834e769dd76d2a61c3a39cad5bb2d090ae" Mar 07 15:07:17 crc kubenswrapper[4943]: E0307 15:07:17.756596 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cpdmm_openshift-machine-config-operator(4c1f18ea-e557-41fc-95dd-ce4e1722a61e)\"" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" Mar 07 15:07:20 crc kubenswrapper[4943]: I0307 15:07:20.853172 4943 scope.go:117] "RemoveContainer" containerID="4e0a94c932a04db92eef6513ced511a8e2e1584cab874c3afd4d3cb291b46377" Mar 07 15:07:20 crc kubenswrapper[4943]: I0307 15:07:20.901337 4943 scope.go:117] "RemoveContainer" containerID="a4e6945468fd8ddc3d4d40a6d039ba2e245da2a8ca630f738f186b6895066a42" Mar 07 15:07:30 crc kubenswrapper[4943]: I0307 15:07:30.755401 4943 scope.go:117] "RemoveContainer" containerID="26f35e1f941cca0c34e3650d7aa047834e769dd76d2a61c3a39cad5bb2d090ae" Mar 07 15:07:30 crc kubenswrapper[4943]: E0307 15:07:30.756415 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cpdmm_openshift-machine-config-operator(4c1f18ea-e557-41fc-95dd-ce4e1722a61e)\"" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" Mar 07 15:07:42 crc kubenswrapper[4943]: I0307 15:07:42.758275 4943 scope.go:117] "RemoveContainer" containerID="26f35e1f941cca0c34e3650d7aa047834e769dd76d2a61c3a39cad5bb2d090ae" Mar 07 15:07:42 crc kubenswrapper[4943]: E0307 15:07:42.759361 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cpdmm_openshift-machine-config-operator(4c1f18ea-e557-41fc-95dd-ce4e1722a61e)\"" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" Mar 07 15:07:57 crc kubenswrapper[4943]: I0307 15:07:57.755722 4943 scope.go:117] "RemoveContainer" containerID="26f35e1f941cca0c34e3650d7aa047834e769dd76d2a61c3a39cad5bb2d090ae" Mar 07 15:07:57 crc kubenswrapper[4943]: E0307 15:07:57.756487 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cpdmm_openshift-machine-config-operator(4c1f18ea-e557-41fc-95dd-ce4e1722a61e)\"" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" Mar 07 15:08:00 crc kubenswrapper[4943]: I0307 15:08:00.163919 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29548268-nlpcf"] Mar 07 15:08:00 crc kubenswrapper[4943]: E0307 15:08:00.164574 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f215184-f0b3-420c-9819-675e4d3a555b" containerName="extract-content" Mar 07 15:08:00 crc kubenswrapper[4943]: I0307 15:08:00.164588 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f215184-f0b3-420c-9819-675e4d3a555b" containerName="extract-content" Mar 07 15:08:00 crc kubenswrapper[4943]: E0307 15:08:00.164603 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f215184-f0b3-420c-9819-675e4d3a555b" containerName="registry-server" Mar 07 15:08:00 crc kubenswrapper[4943]: I0307 15:08:00.164612 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f215184-f0b3-420c-9819-675e4d3a555b" containerName="registry-server" Mar 07 15:08:00 crc kubenswrapper[4943]: E0307 15:08:00.164627 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f215184-f0b3-420c-9819-675e4d3a555b" containerName="extract-utilities" Mar 07 15:08:00 crc kubenswrapper[4943]: I0307 15:08:00.164636 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f215184-f0b3-420c-9819-675e4d3a555b" containerName="extract-utilities" Mar 07 15:08:00 crc kubenswrapper[4943]: I0307 15:08:00.164818 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f215184-f0b3-420c-9819-675e4d3a555b" containerName="registry-server" Mar 07 15:08:00 crc kubenswrapper[4943]: I0307 15:08:00.165485 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548268-nlpcf" Mar 07 15:08:00 crc kubenswrapper[4943]: I0307 15:08:00.169868 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 07 15:08:00 crc kubenswrapper[4943]: I0307 15:08:00.170040 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 07 15:08:00 crc kubenswrapper[4943]: I0307 15:08:00.171465 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-qqlk5" Mar 07 15:08:00 crc kubenswrapper[4943]: I0307 15:08:00.177606 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29548268-nlpcf"] Mar 07 15:08:00 crc kubenswrapper[4943]: I0307 15:08:00.254696 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gx7mb\" (UniqueName: \"kubernetes.io/projected/eb3b8ef8-e951-4f8c-8579-c9515a42b092-kube-api-access-gx7mb\") pod \"auto-csr-approver-29548268-nlpcf\" (UID: \"eb3b8ef8-e951-4f8c-8579-c9515a42b092\") " pod="openshift-infra/auto-csr-approver-29548268-nlpcf" Mar 07 15:08:00 crc kubenswrapper[4943]: I0307 15:08:00.355965 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gx7mb\" (UniqueName: \"kubernetes.io/projected/eb3b8ef8-e951-4f8c-8579-c9515a42b092-kube-api-access-gx7mb\") pod \"auto-csr-approver-29548268-nlpcf\" (UID: \"eb3b8ef8-e951-4f8c-8579-c9515a42b092\") " pod="openshift-infra/auto-csr-approver-29548268-nlpcf" Mar 07 15:08:00 crc kubenswrapper[4943]: I0307 15:08:00.388180 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gx7mb\" (UniqueName: \"kubernetes.io/projected/eb3b8ef8-e951-4f8c-8579-c9515a42b092-kube-api-access-gx7mb\") pod \"auto-csr-approver-29548268-nlpcf\" (UID: \"eb3b8ef8-e951-4f8c-8579-c9515a42b092\") " pod="openshift-infra/auto-csr-approver-29548268-nlpcf" Mar 07 15:08:00 crc kubenswrapper[4943]: I0307 15:08:00.494569 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548268-nlpcf" Mar 07 15:08:01 crc kubenswrapper[4943]: I0307 15:08:01.007585 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29548268-nlpcf"] Mar 07 15:08:01 crc kubenswrapper[4943]: I0307 15:08:01.673960 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548268-nlpcf" event={"ID":"eb3b8ef8-e951-4f8c-8579-c9515a42b092","Type":"ContainerStarted","Data":"22dda2a5ac5e9d01fd0c8da5d40fe1114a80142849cbc000e25150a2fd181465"} Mar 07 15:08:02 crc kubenswrapper[4943]: E0307 15:08:02.656713 4943 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb3b8ef8_e951_4f8c_8579_c9515a42b092.slice/crio-conmon-084497ed13c5cd353f7e40a35741d561595a66526e4545b3576e21cb941b9128.scope\": RecentStats: unable to find data in memory cache]" Mar 07 15:08:02 crc kubenswrapper[4943]: I0307 15:08:02.691374 4943 generic.go:334] "Generic (PLEG): container finished" podID="eb3b8ef8-e951-4f8c-8579-c9515a42b092" containerID="084497ed13c5cd353f7e40a35741d561595a66526e4545b3576e21cb941b9128" exitCode=0 Mar 07 15:08:02 crc kubenswrapper[4943]: I0307 15:08:02.691428 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548268-nlpcf" event={"ID":"eb3b8ef8-e951-4f8c-8579-c9515a42b092","Type":"ContainerDied","Data":"084497ed13c5cd353f7e40a35741d561595a66526e4545b3576e21cb941b9128"} Mar 07 15:08:04 crc kubenswrapper[4943]: I0307 15:08:04.090264 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548268-nlpcf" Mar 07 15:08:04 crc kubenswrapper[4943]: I0307 15:08:04.230437 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gx7mb\" (UniqueName: \"kubernetes.io/projected/eb3b8ef8-e951-4f8c-8579-c9515a42b092-kube-api-access-gx7mb\") pod \"eb3b8ef8-e951-4f8c-8579-c9515a42b092\" (UID: \"eb3b8ef8-e951-4f8c-8579-c9515a42b092\") " Mar 07 15:08:04 crc kubenswrapper[4943]: I0307 15:08:04.240139 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb3b8ef8-e951-4f8c-8579-c9515a42b092-kube-api-access-gx7mb" (OuterVolumeSpecName: "kube-api-access-gx7mb") pod "eb3b8ef8-e951-4f8c-8579-c9515a42b092" (UID: "eb3b8ef8-e951-4f8c-8579-c9515a42b092"). InnerVolumeSpecName "kube-api-access-gx7mb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:08:04 crc kubenswrapper[4943]: I0307 15:08:04.332350 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gx7mb\" (UniqueName: \"kubernetes.io/projected/eb3b8ef8-e951-4f8c-8579-c9515a42b092-kube-api-access-gx7mb\") on node \"crc\" DevicePath \"\"" Mar 07 15:08:04 crc kubenswrapper[4943]: I0307 15:08:04.713225 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548268-nlpcf" event={"ID":"eb3b8ef8-e951-4f8c-8579-c9515a42b092","Type":"ContainerDied","Data":"22dda2a5ac5e9d01fd0c8da5d40fe1114a80142849cbc000e25150a2fd181465"} Mar 07 15:08:04 crc kubenswrapper[4943]: I0307 15:08:04.713279 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="22dda2a5ac5e9d01fd0c8da5d40fe1114a80142849cbc000e25150a2fd181465" Mar 07 15:08:04 crc kubenswrapper[4943]: I0307 15:08:04.713290 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548268-nlpcf" Mar 07 15:08:05 crc kubenswrapper[4943]: I0307 15:08:05.153343 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29548262-k2c8b"] Mar 07 15:08:05 crc kubenswrapper[4943]: I0307 15:08:05.160778 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29548262-k2c8b"] Mar 07 15:08:06 crc kubenswrapper[4943]: I0307 15:08:06.764587 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="feed891c-4b59-472b-9cce-89d1c90652a5" path="/var/lib/kubelet/pods/feed891c-4b59-472b-9cce-89d1c90652a5/volumes" Mar 07 15:08:12 crc kubenswrapper[4943]: I0307 15:08:12.757260 4943 scope.go:117] "RemoveContainer" containerID="26f35e1f941cca0c34e3650d7aa047834e769dd76d2a61c3a39cad5bb2d090ae" Mar 07 15:08:12 crc kubenswrapper[4943]: E0307 15:08:12.758689 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cpdmm_openshift-machine-config-operator(4c1f18ea-e557-41fc-95dd-ce4e1722a61e)\"" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" Mar 07 15:08:21 crc kubenswrapper[4943]: I0307 15:08:21.004523 4943 scope.go:117] "RemoveContainer" containerID="d445bd4612bc3ec557b286d279d0eea9cd49a0656c2b26d29b28f1675b62d5bb" Mar 07 15:08:23 crc kubenswrapper[4943]: I0307 15:08:23.754992 4943 scope.go:117] "RemoveContainer" containerID="26f35e1f941cca0c34e3650d7aa047834e769dd76d2a61c3a39cad5bb2d090ae" Mar 07 15:08:23 crc kubenswrapper[4943]: E0307 15:08:23.756236 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cpdmm_openshift-machine-config-operator(4c1f18ea-e557-41fc-95dd-ce4e1722a61e)\"" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" Mar 07 15:08:38 crc kubenswrapper[4943]: I0307 15:08:38.762491 4943 scope.go:117] "RemoveContainer" containerID="26f35e1f941cca0c34e3650d7aa047834e769dd76d2a61c3a39cad5bb2d090ae" Mar 07 15:08:38 crc kubenswrapper[4943]: E0307 15:08:38.763687 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cpdmm_openshift-machine-config-operator(4c1f18ea-e557-41fc-95dd-ce4e1722a61e)\"" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" Mar 07 15:08:53 crc kubenswrapper[4943]: I0307 15:08:53.755265 4943 scope.go:117] "RemoveContainer" containerID="26f35e1f941cca0c34e3650d7aa047834e769dd76d2a61c3a39cad5bb2d090ae" Mar 07 15:08:53 crc kubenswrapper[4943]: E0307 15:08:53.756920 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cpdmm_openshift-machine-config-operator(4c1f18ea-e557-41fc-95dd-ce4e1722a61e)\"" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" Mar 07 15:09:07 crc kubenswrapper[4943]: I0307 15:09:07.756087 4943 scope.go:117] "RemoveContainer" containerID="26f35e1f941cca0c34e3650d7aa047834e769dd76d2a61c3a39cad5bb2d090ae" Mar 07 15:09:07 crc kubenswrapper[4943]: E0307 15:09:07.757134 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cpdmm_openshift-machine-config-operator(4c1f18ea-e557-41fc-95dd-ce4e1722a61e)\"" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" Mar 07 15:09:18 crc kubenswrapper[4943]: I0307 15:09:18.759028 4943 scope.go:117] "RemoveContainer" containerID="26f35e1f941cca0c34e3650d7aa047834e769dd76d2a61c3a39cad5bb2d090ae" Mar 07 15:09:18 crc kubenswrapper[4943]: E0307 15:09:18.759822 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cpdmm_openshift-machine-config-operator(4c1f18ea-e557-41fc-95dd-ce4e1722a61e)\"" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" Mar 07 15:09:30 crc kubenswrapper[4943]: I0307 15:09:30.676373 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-khdj6"] Mar 07 15:09:30 crc kubenswrapper[4943]: E0307 15:09:30.677341 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb3b8ef8-e951-4f8c-8579-c9515a42b092" containerName="oc" Mar 07 15:09:30 crc kubenswrapper[4943]: I0307 15:09:30.677361 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb3b8ef8-e951-4f8c-8579-c9515a42b092" containerName="oc" Mar 07 15:09:30 crc kubenswrapper[4943]: I0307 15:09:30.677613 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb3b8ef8-e951-4f8c-8579-c9515a42b092" containerName="oc" Mar 07 15:09:30 crc kubenswrapper[4943]: I0307 15:09:30.678457 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-khdj6" Mar 07 15:09:30 crc kubenswrapper[4943]: I0307 15:09:30.682213 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:09:30 crc kubenswrapper[4943]: I0307 15:09:30.682601 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:09:30 crc kubenswrapper[4943]: I0307 15:09:30.688507 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-khdj6"] Mar 07 15:09:30 crc kubenswrapper[4943]: I0307 15:09:30.741045 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/beeb25d9-4a5d-4e7c-a148-55c164a99704-scripts\") pod \"swift-ring-rebalance-debug-khdj6\" (UID: \"beeb25d9-4a5d-4e7c-a148-55c164a99704\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-khdj6" Mar 07 15:09:30 crc kubenswrapper[4943]: I0307 15:09:30.741125 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfb4c\" (UniqueName: \"kubernetes.io/projected/beeb25d9-4a5d-4e7c-a148-55c164a99704-kube-api-access-lfb4c\") pod \"swift-ring-rebalance-debug-khdj6\" (UID: \"beeb25d9-4a5d-4e7c-a148-55c164a99704\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-khdj6" Mar 07 15:09:30 crc kubenswrapper[4943]: I0307 15:09:30.741201 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/beeb25d9-4a5d-4e7c-a148-55c164a99704-dispersionconf\") pod \"swift-ring-rebalance-debug-khdj6\" (UID: \"beeb25d9-4a5d-4e7c-a148-55c164a99704\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-khdj6" Mar 07 15:09:30 crc kubenswrapper[4943]: I0307 15:09:30.741285 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/beeb25d9-4a5d-4e7c-a148-55c164a99704-etc-swift\") pod \"swift-ring-rebalance-debug-khdj6\" (UID: \"beeb25d9-4a5d-4e7c-a148-55c164a99704\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-khdj6" Mar 07 15:09:30 crc kubenswrapper[4943]: I0307 15:09:30.741355 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/beeb25d9-4a5d-4e7c-a148-55c164a99704-ring-data-devices\") pod \"swift-ring-rebalance-debug-khdj6\" (UID: \"beeb25d9-4a5d-4e7c-a148-55c164a99704\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-khdj6" Mar 07 15:09:30 crc kubenswrapper[4943]: I0307 15:09:30.741422 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/beeb25d9-4a5d-4e7c-a148-55c164a99704-swiftconf\") pod \"swift-ring-rebalance-debug-khdj6\" (UID: \"beeb25d9-4a5d-4e7c-a148-55c164a99704\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-khdj6" Mar 07 15:09:30 crc kubenswrapper[4943]: I0307 15:09:30.842616 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfb4c\" (UniqueName: \"kubernetes.io/projected/beeb25d9-4a5d-4e7c-a148-55c164a99704-kube-api-access-lfb4c\") pod \"swift-ring-rebalance-debug-khdj6\" (UID: \"beeb25d9-4a5d-4e7c-a148-55c164a99704\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-khdj6" Mar 07 15:09:30 crc kubenswrapper[4943]: I0307 15:09:30.842726 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/beeb25d9-4a5d-4e7c-a148-55c164a99704-dispersionconf\") pod \"swift-ring-rebalance-debug-khdj6\" (UID: \"beeb25d9-4a5d-4e7c-a148-55c164a99704\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-khdj6" Mar 07 15:09:30 crc kubenswrapper[4943]: I0307 15:09:30.842807 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/beeb25d9-4a5d-4e7c-a148-55c164a99704-etc-swift\") pod \"swift-ring-rebalance-debug-khdj6\" (UID: \"beeb25d9-4a5d-4e7c-a148-55c164a99704\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-khdj6" Mar 07 15:09:30 crc kubenswrapper[4943]: I0307 15:09:30.842851 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/beeb25d9-4a5d-4e7c-a148-55c164a99704-ring-data-devices\") pod \"swift-ring-rebalance-debug-khdj6\" (UID: \"beeb25d9-4a5d-4e7c-a148-55c164a99704\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-khdj6" Mar 07 15:09:30 crc kubenswrapper[4943]: I0307 15:09:30.842915 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/beeb25d9-4a5d-4e7c-a148-55c164a99704-swiftconf\") pod \"swift-ring-rebalance-debug-khdj6\" (UID: \"beeb25d9-4a5d-4e7c-a148-55c164a99704\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-khdj6" Mar 07 15:09:30 crc kubenswrapper[4943]: I0307 15:09:30.843017 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/beeb25d9-4a5d-4e7c-a148-55c164a99704-scripts\") pod \"swift-ring-rebalance-debug-khdj6\" (UID: \"beeb25d9-4a5d-4e7c-a148-55c164a99704\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-khdj6" Mar 07 15:09:30 crc kubenswrapper[4943]: I0307 15:09:30.843857 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/beeb25d9-4a5d-4e7c-a148-55c164a99704-scripts\") pod \"swift-ring-rebalance-debug-khdj6\" (UID: \"beeb25d9-4a5d-4e7c-a148-55c164a99704\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-khdj6" Mar 07 15:09:30 crc kubenswrapper[4943]: I0307 15:09:30.844177 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/beeb25d9-4a5d-4e7c-a148-55c164a99704-etc-swift\") pod \"swift-ring-rebalance-debug-khdj6\" (UID: \"beeb25d9-4a5d-4e7c-a148-55c164a99704\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-khdj6" Mar 07 15:09:30 crc kubenswrapper[4943]: I0307 15:09:30.844751 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/beeb25d9-4a5d-4e7c-a148-55c164a99704-ring-data-devices\") pod \"swift-ring-rebalance-debug-khdj6\" (UID: \"beeb25d9-4a5d-4e7c-a148-55c164a99704\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-khdj6" Mar 07 15:09:30 crc kubenswrapper[4943]: I0307 15:09:30.848560 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/beeb25d9-4a5d-4e7c-a148-55c164a99704-dispersionconf\") pod \"swift-ring-rebalance-debug-khdj6\" (UID: \"beeb25d9-4a5d-4e7c-a148-55c164a99704\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-khdj6" Mar 07 15:09:30 crc kubenswrapper[4943]: I0307 15:09:30.851132 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/beeb25d9-4a5d-4e7c-a148-55c164a99704-swiftconf\") pod \"swift-ring-rebalance-debug-khdj6\" (UID: \"beeb25d9-4a5d-4e7c-a148-55c164a99704\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-khdj6" Mar 07 15:09:30 crc kubenswrapper[4943]: I0307 15:09:30.873832 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfb4c\" (UniqueName: \"kubernetes.io/projected/beeb25d9-4a5d-4e7c-a148-55c164a99704-kube-api-access-lfb4c\") pod \"swift-ring-rebalance-debug-khdj6\" (UID: \"beeb25d9-4a5d-4e7c-a148-55c164a99704\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-khdj6" Mar 07 15:09:31 crc kubenswrapper[4943]: I0307 15:09:31.005700 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-khdj6" Mar 07 15:09:31 crc kubenswrapper[4943]: I0307 15:09:31.496874 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-khdj6"] Mar 07 15:09:31 crc kubenswrapper[4943]: I0307 15:09:31.531200 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-khdj6" event={"ID":"beeb25d9-4a5d-4e7c-a148-55c164a99704","Type":"ContainerStarted","Data":"3a2620b80e90a4294781c40c4a246ae281d4654be0bb7fd3ead613ca59f1c518"} Mar 07 15:09:31 crc kubenswrapper[4943]: I0307 15:09:31.792961 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-storage-1"] Mar 07 15:09:31 crc kubenswrapper[4943]: I0307 15:09:31.820629 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-storage-1" Mar 07 15:09:31 crc kubenswrapper[4943]: I0307 15:09:31.822292 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-storage-2"] Mar 07 15:09:31 crc kubenswrapper[4943]: I0307 15:09:31.852046 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-storage-1"] Mar 07 15:09:31 crc kubenswrapper[4943]: I0307 15:09:31.852161 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-storage-2" Mar 07 15:09:31 crc kubenswrapper[4943]: I0307 15:09:31.860202 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/40da0e9b-e661-487c-8898-a54286149f85-etc-swift\") pod \"swift-storage-1\" (UID: \"40da0e9b-e661-487c-8898-a54286149f85\") " pod="swift-kuttl-tests/swift-storage-1" Mar 07 15:09:31 crc kubenswrapper[4943]: I0307 15:09:31.860326 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"swift-storage-1\" (UID: \"40da0e9b-e661-487c-8898-a54286149f85\") " pod="swift-kuttl-tests/swift-storage-1" Mar 07 15:09:31 crc kubenswrapper[4943]: I0307 15:09:31.860359 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/40da0e9b-e661-487c-8898-a54286149f85-cache\") pod \"swift-storage-1\" (UID: \"40da0e9b-e661-487c-8898-a54286149f85\") " pod="swift-kuttl-tests/swift-storage-1" Mar 07 15:09:31 crc kubenswrapper[4943]: I0307 15:09:31.860486 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nq5nm\" (UniqueName: \"kubernetes.io/projected/40da0e9b-e661-487c-8898-a54286149f85-kube-api-access-nq5nm\") pod \"swift-storage-1\" (UID: \"40da0e9b-e661-487c-8898-a54286149f85\") " pod="swift-kuttl-tests/swift-storage-1" Mar 07 15:09:31 crc kubenswrapper[4943]: I0307 15:09:31.860574 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/40da0e9b-e661-487c-8898-a54286149f85-lock\") pod \"swift-storage-1\" (UID: \"40da0e9b-e661-487c-8898-a54286149f85\") " pod="swift-kuttl-tests/swift-storage-1" Mar 07 15:09:31 crc kubenswrapper[4943]: I0307 15:09:31.861730 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-storage-2"] Mar 07 15:09:31 crc kubenswrapper[4943]: I0307 15:09:31.964672 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"swift-storage-1\" (UID: \"40da0e9b-e661-487c-8898-a54286149f85\") " pod="swift-kuttl-tests/swift-storage-1" Mar 07 15:09:31 crc kubenswrapper[4943]: I0307 15:09:31.964718 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/40da0e9b-e661-487c-8898-a54286149f85-cache\") pod \"swift-storage-1\" (UID: \"40da0e9b-e661-487c-8898-a54286149f85\") " pod="swift-kuttl-tests/swift-storage-1" Mar 07 15:09:31 crc kubenswrapper[4943]: I0307 15:09:31.964740 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/c77eb68a-e396-4958-8657-8e62013b774e-lock\") pod \"swift-storage-2\" (UID: \"c77eb68a-e396-4958-8657-8e62013b774e\") " pod="swift-kuttl-tests/swift-storage-2" Mar 07 15:09:31 crc kubenswrapper[4943]: I0307 15:09:31.964768 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-2\" (UID: \"c77eb68a-e396-4958-8657-8e62013b774e\") " pod="swift-kuttl-tests/swift-storage-2" Mar 07 15:09:31 crc kubenswrapper[4943]: I0307 15:09:31.964805 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nq5nm\" (UniqueName: \"kubernetes.io/projected/40da0e9b-e661-487c-8898-a54286149f85-kube-api-access-nq5nm\") pod \"swift-storage-1\" (UID: \"40da0e9b-e661-487c-8898-a54286149f85\") " pod="swift-kuttl-tests/swift-storage-1" Mar 07 15:09:31 crc kubenswrapper[4943]: I0307 15:09:31.964833 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c77eb68a-e396-4958-8657-8e62013b774e-etc-swift\") pod \"swift-storage-2\" (UID: \"c77eb68a-e396-4958-8657-8e62013b774e\") " pod="swift-kuttl-tests/swift-storage-2" Mar 07 15:09:31 crc kubenswrapper[4943]: I0307 15:09:31.964856 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/40da0e9b-e661-487c-8898-a54286149f85-lock\") pod \"swift-storage-1\" (UID: \"40da0e9b-e661-487c-8898-a54286149f85\") " pod="swift-kuttl-tests/swift-storage-1" Mar 07 15:09:31 crc kubenswrapper[4943]: I0307 15:09:31.964872 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qnlx\" (UniqueName: \"kubernetes.io/projected/c77eb68a-e396-4958-8657-8e62013b774e-kube-api-access-4qnlx\") pod \"swift-storage-2\" (UID: \"c77eb68a-e396-4958-8657-8e62013b774e\") " pod="swift-kuttl-tests/swift-storage-2" Mar 07 15:09:31 crc kubenswrapper[4943]: I0307 15:09:31.964897 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/40da0e9b-e661-487c-8898-a54286149f85-etc-swift\") pod \"swift-storage-1\" (UID: \"40da0e9b-e661-487c-8898-a54286149f85\") " pod="swift-kuttl-tests/swift-storage-1" Mar 07 15:09:31 crc kubenswrapper[4943]: I0307 15:09:31.964918 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/c77eb68a-e396-4958-8657-8e62013b774e-cache\") pod \"swift-storage-2\" (UID: \"c77eb68a-e396-4958-8657-8e62013b774e\") " pod="swift-kuttl-tests/swift-storage-2" Mar 07 15:09:31 crc kubenswrapper[4943]: I0307 15:09:31.965518 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/40da0e9b-e661-487c-8898-a54286149f85-cache\") pod \"swift-storage-1\" (UID: \"40da0e9b-e661-487c-8898-a54286149f85\") " pod="swift-kuttl-tests/swift-storage-1" Mar 07 15:09:31 crc kubenswrapper[4943]: I0307 15:09:31.965633 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/40da0e9b-e661-487c-8898-a54286149f85-lock\") pod \"swift-storage-1\" (UID: \"40da0e9b-e661-487c-8898-a54286149f85\") " pod="swift-kuttl-tests/swift-storage-1" Mar 07 15:09:31 crc kubenswrapper[4943]: I0307 15:09:31.965795 4943 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"swift-storage-1\" (UID: \"40da0e9b-e661-487c-8898-a54286149f85\") device mount path \"/mnt/openstack/pv01\"" pod="swift-kuttl-tests/swift-storage-1" Mar 07 15:09:31 crc kubenswrapper[4943]: I0307 15:09:31.971571 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/40da0e9b-e661-487c-8898-a54286149f85-etc-swift\") pod \"swift-storage-1\" (UID: \"40da0e9b-e661-487c-8898-a54286149f85\") " pod="swift-kuttl-tests/swift-storage-1" Mar 07 15:09:31 crc kubenswrapper[4943]: I0307 15:09:31.989216 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nq5nm\" (UniqueName: \"kubernetes.io/projected/40da0e9b-e661-487c-8898-a54286149f85-kube-api-access-nq5nm\") pod \"swift-storage-1\" (UID: \"40da0e9b-e661-487c-8898-a54286149f85\") " pod="swift-kuttl-tests/swift-storage-1" Mar 07 15:09:31 crc kubenswrapper[4943]: I0307 15:09:31.993436 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"swift-storage-1\" (UID: \"40da0e9b-e661-487c-8898-a54286149f85\") " pod="swift-kuttl-tests/swift-storage-1" Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.056885 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-proxy-76c998454c-kngkw"] Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.058060 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-proxy-76c998454c-kngkw" Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.065780 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/c77eb68a-e396-4958-8657-8e62013b774e-cache\") pod \"swift-storage-2\" (UID: \"c77eb68a-e396-4958-8657-8e62013b774e\") " pod="swift-kuttl-tests/swift-storage-2" Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.065858 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/c77eb68a-e396-4958-8657-8e62013b774e-lock\") pod \"swift-storage-2\" (UID: \"c77eb68a-e396-4958-8657-8e62013b774e\") " pod="swift-kuttl-tests/swift-storage-2" Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.065892 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-2\" (UID: \"c77eb68a-e396-4958-8657-8e62013b774e\") " pod="swift-kuttl-tests/swift-storage-2" Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.065977 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c77eb68a-e396-4958-8657-8e62013b774e-etc-swift\") pod \"swift-storage-2\" (UID: \"c77eb68a-e396-4958-8657-8e62013b774e\") " pod="swift-kuttl-tests/swift-storage-2" Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.066012 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qnlx\" (UniqueName: \"kubernetes.io/projected/c77eb68a-e396-4958-8657-8e62013b774e-kube-api-access-4qnlx\") pod \"swift-storage-2\" (UID: \"c77eb68a-e396-4958-8657-8e62013b774e\") " pod="swift-kuttl-tests/swift-storage-2" Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.066141 4943 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-2\" (UID: \"c77eb68a-e396-4958-8657-8e62013b774e\") device mount path \"/mnt/openstack/pv06\"" pod="swift-kuttl-tests/swift-storage-2" Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.066725 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/c77eb68a-e396-4958-8657-8e62013b774e-lock\") pod \"swift-storage-2\" (UID: \"c77eb68a-e396-4958-8657-8e62013b774e\") " pod="swift-kuttl-tests/swift-storage-2" Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.069476 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/c77eb68a-e396-4958-8657-8e62013b774e-cache\") pod \"swift-storage-2\" (UID: \"c77eb68a-e396-4958-8657-8e62013b774e\") " pod="swift-kuttl-tests/swift-storage-2" Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.073210 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-proxy-76c998454c-kngkw"] Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.073861 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c77eb68a-e396-4958-8657-8e62013b774e-etc-swift\") pod \"swift-storage-2\" (UID: \"c77eb68a-e396-4958-8657-8e62013b774e\") " pod="swift-kuttl-tests/swift-storage-2" Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.085084 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-2\" (UID: \"c77eb68a-e396-4958-8657-8e62013b774e\") " pod="swift-kuttl-tests/swift-storage-2" Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.089802 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qnlx\" (UniqueName: \"kubernetes.io/projected/c77eb68a-e396-4958-8657-8e62013b774e-kube-api-access-4qnlx\") pod \"swift-storage-2\" (UID: \"c77eb68a-e396-4958-8657-8e62013b774e\") " pod="swift-kuttl-tests/swift-storage-2" Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.134716 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-mvdbn"] Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.142512 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-mvdbn"] Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.166989 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/32f9bc3c-6fe2-42ab-8485-d7fda4d10c78-log-httpd\") pod \"swift-proxy-76c998454c-kngkw\" (UID: \"32f9bc3c-6fe2-42ab-8485-d7fda4d10c78\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-kngkw" Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.167043 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32f9bc3c-6fe2-42ab-8485-d7fda4d10c78-config-data\") pod \"swift-proxy-76c998454c-kngkw\" (UID: \"32f9bc3c-6fe2-42ab-8485-d7fda4d10c78\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-kngkw" Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.167077 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/32f9bc3c-6fe2-42ab-8485-d7fda4d10c78-etc-swift\") pod \"swift-proxy-76c998454c-kngkw\" (UID: \"32f9bc3c-6fe2-42ab-8485-d7fda4d10c78\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-kngkw" Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.167114 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/32f9bc3c-6fe2-42ab-8485-d7fda4d10c78-run-httpd\") pod \"swift-proxy-76c998454c-kngkw\" (UID: \"32f9bc3c-6fe2-42ab-8485-d7fda4d10c78\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-kngkw" Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.167384 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfflq\" (UniqueName: \"kubernetes.io/projected/32f9bc3c-6fe2-42ab-8485-d7fda4d10c78-kube-api-access-qfflq\") pod \"swift-proxy-76c998454c-kngkw\" (UID: \"32f9bc3c-6fe2-42ab-8485-d7fda4d10c78\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-kngkw" Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.170400 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-s5kqz"] Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.171446 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-s5kqz" Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.181782 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-s5kqz"] Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.191815 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-storage-1" Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.198704 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-storage-2" Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.268890 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/32f9bc3c-6fe2-42ab-8485-d7fda4d10c78-etc-swift\") pod \"swift-proxy-76c998454c-kngkw\" (UID: \"32f9bc3c-6fe2-42ab-8485-d7fda4d10c78\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-kngkw" Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.268972 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r52nz\" (UniqueName: \"kubernetes.io/projected/85e1e3e8-b9f8-4c3d-940e-4060262201ee-kube-api-access-r52nz\") pod \"swift-ring-rebalance-s5kqz\" (UID: \"85e1e3e8-b9f8-4c3d-940e-4060262201ee\") " pod="swift-kuttl-tests/swift-ring-rebalance-s5kqz" Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.269001 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/32f9bc3c-6fe2-42ab-8485-d7fda4d10c78-run-httpd\") pod \"swift-proxy-76c998454c-kngkw\" (UID: \"32f9bc3c-6fe2-42ab-8485-d7fda4d10c78\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-kngkw" Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.269031 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/85e1e3e8-b9f8-4c3d-940e-4060262201ee-etc-swift\") pod \"swift-ring-rebalance-s5kqz\" (UID: \"85e1e3e8-b9f8-4c3d-940e-4060262201ee\") " pod="swift-kuttl-tests/swift-ring-rebalance-s5kqz" Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.269054 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/85e1e3e8-b9f8-4c3d-940e-4060262201ee-dispersionconf\") pod \"swift-ring-rebalance-s5kqz\" (UID: \"85e1e3e8-b9f8-4c3d-940e-4060262201ee\") " pod="swift-kuttl-tests/swift-ring-rebalance-s5kqz" Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.269101 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/85e1e3e8-b9f8-4c3d-940e-4060262201ee-swiftconf\") pod \"swift-ring-rebalance-s5kqz\" (UID: \"85e1e3e8-b9f8-4c3d-940e-4060262201ee\") " pod="swift-kuttl-tests/swift-ring-rebalance-s5kqz" Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.269131 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/85e1e3e8-b9f8-4c3d-940e-4060262201ee-scripts\") pod \"swift-ring-rebalance-s5kqz\" (UID: \"85e1e3e8-b9f8-4c3d-940e-4060262201ee\") " pod="swift-kuttl-tests/swift-ring-rebalance-s5kqz" Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.269184 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfflq\" (UniqueName: \"kubernetes.io/projected/32f9bc3c-6fe2-42ab-8485-d7fda4d10c78-kube-api-access-qfflq\") pod \"swift-proxy-76c998454c-kngkw\" (UID: \"32f9bc3c-6fe2-42ab-8485-d7fda4d10c78\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-kngkw" Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.269220 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/32f9bc3c-6fe2-42ab-8485-d7fda4d10c78-log-httpd\") pod \"swift-proxy-76c998454c-kngkw\" (UID: \"32f9bc3c-6fe2-42ab-8485-d7fda4d10c78\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-kngkw" Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.269249 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/85e1e3e8-b9f8-4c3d-940e-4060262201ee-ring-data-devices\") pod \"swift-ring-rebalance-s5kqz\" (UID: \"85e1e3e8-b9f8-4c3d-940e-4060262201ee\") " pod="swift-kuttl-tests/swift-ring-rebalance-s5kqz" Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.269273 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32f9bc3c-6fe2-42ab-8485-d7fda4d10c78-config-data\") pod \"swift-proxy-76c998454c-kngkw\" (UID: \"32f9bc3c-6fe2-42ab-8485-d7fda4d10c78\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-kngkw" Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.270164 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/32f9bc3c-6fe2-42ab-8485-d7fda4d10c78-run-httpd\") pod \"swift-proxy-76c998454c-kngkw\" (UID: \"32f9bc3c-6fe2-42ab-8485-d7fda4d10c78\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-kngkw" Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.270977 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/32f9bc3c-6fe2-42ab-8485-d7fda4d10c78-log-httpd\") pod \"swift-proxy-76c998454c-kngkw\" (UID: \"32f9bc3c-6fe2-42ab-8485-d7fda4d10c78\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-kngkw" Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.273338 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32f9bc3c-6fe2-42ab-8485-d7fda4d10c78-config-data\") pod \"swift-proxy-76c998454c-kngkw\" (UID: \"32f9bc3c-6fe2-42ab-8485-d7fda4d10c78\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-kngkw" Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.273398 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/32f9bc3c-6fe2-42ab-8485-d7fda4d10c78-etc-swift\") pod \"swift-proxy-76c998454c-kngkw\" (UID: \"32f9bc3c-6fe2-42ab-8485-d7fda4d10c78\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-kngkw" Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.285511 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfflq\" (UniqueName: \"kubernetes.io/projected/32f9bc3c-6fe2-42ab-8485-d7fda4d10c78-kube-api-access-qfflq\") pod \"swift-proxy-76c998454c-kngkw\" (UID: \"32f9bc3c-6fe2-42ab-8485-d7fda4d10c78\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-kngkw" Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.370940 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/85e1e3e8-b9f8-4c3d-940e-4060262201ee-ring-data-devices\") pod \"swift-ring-rebalance-s5kqz\" (UID: \"85e1e3e8-b9f8-4c3d-940e-4060262201ee\") " pod="swift-kuttl-tests/swift-ring-rebalance-s5kqz" Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.371021 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r52nz\" (UniqueName: \"kubernetes.io/projected/85e1e3e8-b9f8-4c3d-940e-4060262201ee-kube-api-access-r52nz\") pod \"swift-ring-rebalance-s5kqz\" (UID: \"85e1e3e8-b9f8-4c3d-940e-4060262201ee\") " pod="swift-kuttl-tests/swift-ring-rebalance-s5kqz" Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.371053 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/85e1e3e8-b9f8-4c3d-940e-4060262201ee-etc-swift\") pod \"swift-ring-rebalance-s5kqz\" (UID: \"85e1e3e8-b9f8-4c3d-940e-4060262201ee\") " pod="swift-kuttl-tests/swift-ring-rebalance-s5kqz" Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.371074 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/85e1e3e8-b9f8-4c3d-940e-4060262201ee-dispersionconf\") pod \"swift-ring-rebalance-s5kqz\" (UID: \"85e1e3e8-b9f8-4c3d-940e-4060262201ee\") " pod="swift-kuttl-tests/swift-ring-rebalance-s5kqz" Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.371123 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/85e1e3e8-b9f8-4c3d-940e-4060262201ee-swiftconf\") pod \"swift-ring-rebalance-s5kqz\" (UID: \"85e1e3e8-b9f8-4c3d-940e-4060262201ee\") " pod="swift-kuttl-tests/swift-ring-rebalance-s5kqz" Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.371160 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/85e1e3e8-b9f8-4c3d-940e-4060262201ee-scripts\") pod \"swift-ring-rebalance-s5kqz\" (UID: \"85e1e3e8-b9f8-4c3d-940e-4060262201ee\") " pod="swift-kuttl-tests/swift-ring-rebalance-s5kqz" Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.371726 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/85e1e3e8-b9f8-4c3d-940e-4060262201ee-etc-swift\") pod \"swift-ring-rebalance-s5kqz\" (UID: \"85e1e3e8-b9f8-4c3d-940e-4060262201ee\") " pod="swift-kuttl-tests/swift-ring-rebalance-s5kqz" Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.372459 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/85e1e3e8-b9f8-4c3d-940e-4060262201ee-ring-data-devices\") pod \"swift-ring-rebalance-s5kqz\" (UID: \"85e1e3e8-b9f8-4c3d-940e-4060262201ee\") " pod="swift-kuttl-tests/swift-ring-rebalance-s5kqz" Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.372627 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/85e1e3e8-b9f8-4c3d-940e-4060262201ee-scripts\") pod \"swift-ring-rebalance-s5kqz\" (UID: \"85e1e3e8-b9f8-4c3d-940e-4060262201ee\") " pod="swift-kuttl-tests/swift-ring-rebalance-s5kqz" Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.374759 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/85e1e3e8-b9f8-4c3d-940e-4060262201ee-swiftconf\") pod \"swift-ring-rebalance-s5kqz\" (UID: \"85e1e3e8-b9f8-4c3d-940e-4060262201ee\") " pod="swift-kuttl-tests/swift-ring-rebalance-s5kqz" Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.374794 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/85e1e3e8-b9f8-4c3d-940e-4060262201ee-dispersionconf\") pod \"swift-ring-rebalance-s5kqz\" (UID: \"85e1e3e8-b9f8-4c3d-940e-4060262201ee\") " pod="swift-kuttl-tests/swift-ring-rebalance-s5kqz" Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.375180 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-proxy-76c998454c-kngkw" Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.392014 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r52nz\" (UniqueName: \"kubernetes.io/projected/85e1e3e8-b9f8-4c3d-940e-4060262201ee-kube-api-access-r52nz\") pod \"swift-ring-rebalance-s5kqz\" (UID: \"85e1e3e8-b9f8-4c3d-940e-4060262201ee\") " pod="swift-kuttl-tests/swift-ring-rebalance-s5kqz" Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.484228 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-s5kqz" Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.539960 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-khdj6" event={"ID":"beeb25d9-4a5d-4e7c-a148-55c164a99704","Type":"ContainerStarted","Data":"0b3d2f580d1445cec8ffea6df297252f539ecf099f6b63cb72727022bdf8093d"} Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.568987 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-khdj6" podStartSLOduration=2.568970448 podStartE2EDuration="2.568970448s" podCreationTimestamp="2026-03-07 15:09:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:09:32.564011966 +0000 UTC m=+1814.516148464" watchObservedRunningTime="2026-03-07 15:09:32.568970448 +0000 UTC m=+1814.521106946" Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.664971 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-storage-2"] Mar 07 15:09:32 crc kubenswrapper[4943]: W0307 15:09:32.670544 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc77eb68a_e396_4958_8657_8e62013b774e.slice/crio-a7a86750512559f4717b0029229b22cdacf5a98f34f3ef2f75c11a48f23bb305 WatchSource:0}: Error finding container a7a86750512559f4717b0029229b22cdacf5a98f34f3ef2f75c11a48f23bb305: Status 404 returned error can't find the container with id a7a86750512559f4717b0029229b22cdacf5a98f34f3ef2f75c11a48f23bb305 Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.753273 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-storage-1"] Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.759947 4943 scope.go:117] "RemoveContainer" containerID="26f35e1f941cca0c34e3650d7aa047834e769dd76d2a61c3a39cad5bb2d090ae" Mar 07 15:09:32 crc kubenswrapper[4943]: E0307 15:09:32.760171 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cpdmm_openshift-machine-config-operator(4c1f18ea-e557-41fc-95dd-ce4e1722a61e)\"" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.770912 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24902a59-0296-47f0-b64e-2291d9fe0ee2" path="/var/lib/kubelet/pods/24902a59-0296-47f0-b64e-2291d9fe0ee2/volumes" Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.789024 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-proxy-76c998454c-kngkw"] Mar 07 15:09:32 crc kubenswrapper[4943]: I0307 15:09:32.908100 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-s5kqz"] Mar 07 15:09:32 crc kubenswrapper[4943]: W0307 15:09:32.916458 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod85e1e3e8_b9f8_4c3d_940e_4060262201ee.slice/crio-b17c0d4741b2bdff0e0844e11447e44b3a997fc20dd4268105cec6eebed81d38 WatchSource:0}: Error finding container b17c0d4741b2bdff0e0844e11447e44b3a997fc20dd4268105cec6eebed81d38: Status 404 returned error can't find the container with id b17c0d4741b2bdff0e0844e11447e44b3a997fc20dd4268105cec6eebed81d38 Mar 07 15:09:33 crc kubenswrapper[4943]: I0307 15:09:33.559084 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"c77eb68a-e396-4958-8657-8e62013b774e","Type":"ContainerStarted","Data":"404d1a326f89fcdaf9562093d293b5fc43b73d2006024ea34454388e1a75f6a2"} Mar 07 15:09:33 crc kubenswrapper[4943]: I0307 15:09:33.559134 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"c77eb68a-e396-4958-8657-8e62013b774e","Type":"ContainerStarted","Data":"5f5d6de06f1893da372a87dce2e877a2f7965841f9cc68b776009c8c0d9cc9ee"} Mar 07 15:09:33 crc kubenswrapper[4943]: I0307 15:09:33.559144 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"c77eb68a-e396-4958-8657-8e62013b774e","Type":"ContainerStarted","Data":"0de049b6dadfef3eb9c27e1de578b00820dbe9354a1d4854921375d0d4830c95"} Mar 07 15:09:33 crc kubenswrapper[4943]: I0307 15:09:33.559152 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"c77eb68a-e396-4958-8657-8e62013b774e","Type":"ContainerStarted","Data":"a7a86750512559f4717b0029229b22cdacf5a98f34f3ef2f75c11a48f23bb305"} Mar 07 15:09:33 crc kubenswrapper[4943]: I0307 15:09:33.563154 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"40da0e9b-e661-487c-8898-a54286149f85","Type":"ContainerStarted","Data":"9dae1d8779c4c5b7dd0c95d98c2b6cfeea018076e3402bfb7f68d8507db9d04f"} Mar 07 15:09:33 crc kubenswrapper[4943]: I0307 15:09:33.563182 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"40da0e9b-e661-487c-8898-a54286149f85","Type":"ContainerStarted","Data":"4a1bd8cbfe1dfcdefd48d86de621eaaafdf1c311b615fb1c51d735453beafa4c"} Mar 07 15:09:33 crc kubenswrapper[4943]: I0307 15:09:33.563191 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"40da0e9b-e661-487c-8898-a54286149f85","Type":"ContainerStarted","Data":"f7ae884859303cef904d56e34794eabcf379816495837646329dbbc51b65b8ef"} Mar 07 15:09:33 crc kubenswrapper[4943]: I0307 15:09:33.563200 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"40da0e9b-e661-487c-8898-a54286149f85","Type":"ContainerStarted","Data":"a5e16536e8ffb8b85b761fec1c2f8f37f92839db9c87741558a620f6c3c8cc73"} Mar 07 15:09:33 crc kubenswrapper[4943]: I0307 15:09:33.563210 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"40da0e9b-e661-487c-8898-a54286149f85","Type":"ContainerStarted","Data":"9ddd7252c72b6d6fc3b29dde4c6feb6f86c36a90c13a118c03973c25328b8c17"} Mar 07 15:09:33 crc kubenswrapper[4943]: I0307 15:09:33.566453 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-s5kqz" event={"ID":"85e1e3e8-b9f8-4c3d-940e-4060262201ee","Type":"ContainerStarted","Data":"554f64008e702a987e22726c874147b00d5ca6af3bd7acae9d07bdab80612bb1"} Mar 07 15:09:33 crc kubenswrapper[4943]: I0307 15:09:33.566498 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-s5kqz" event={"ID":"85e1e3e8-b9f8-4c3d-940e-4060262201ee","Type":"ContainerStarted","Data":"b17c0d4741b2bdff0e0844e11447e44b3a997fc20dd4268105cec6eebed81d38"} Mar 07 15:09:33 crc kubenswrapper[4943]: I0307 15:09:33.574183 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-proxy-76c998454c-kngkw" event={"ID":"32f9bc3c-6fe2-42ab-8485-d7fda4d10c78","Type":"ContainerStarted","Data":"3704ab966cd8d1dbd2a46a6ca8bf107a7d9d1873182df3da70053ff4af30b35b"} Mar 07 15:09:33 crc kubenswrapper[4943]: I0307 15:09:33.574220 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-proxy-76c998454c-kngkw" event={"ID":"32f9bc3c-6fe2-42ab-8485-d7fda4d10c78","Type":"ContainerStarted","Data":"0a22553c0bf9faa524461f70d3879e9d4d2719e98a35a2a4d6db523e16077d74"} Mar 07 15:09:33 crc kubenswrapper[4943]: I0307 15:09:33.574247 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="swift-kuttl-tests/swift-proxy-76c998454c-kngkw" Mar 07 15:09:33 crc kubenswrapper[4943]: I0307 15:09:33.574259 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-proxy-76c998454c-kngkw" event={"ID":"32f9bc3c-6fe2-42ab-8485-d7fda4d10c78","Type":"ContainerStarted","Data":"fabc0e67e1d588e331e86144329c32d61064b9dfe56c607d4dda3bb340e89de7"} Mar 07 15:09:33 crc kubenswrapper[4943]: I0307 15:09:33.574280 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="swift-kuttl-tests/swift-proxy-76c998454c-kngkw" Mar 07 15:09:33 crc kubenswrapper[4943]: I0307 15:09:33.610879 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-s5kqz" podStartSLOduration=1.6108530810000001 podStartE2EDuration="1.610853081s" podCreationTimestamp="2026-03-07 15:09:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:09:33.593046674 +0000 UTC m=+1815.545183172" watchObservedRunningTime="2026-03-07 15:09:33.610853081 +0000 UTC m=+1815.562989579" Mar 07 15:09:34 crc kubenswrapper[4943]: I0307 15:09:34.596038 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"c77eb68a-e396-4958-8657-8e62013b774e","Type":"ContainerStarted","Data":"d99ec228ad8cdd7f23ea2fafbf1a4e2d872f58d96d6db84550fdbd3b636b0612"} Mar 07 15:09:34 crc kubenswrapper[4943]: I0307 15:09:34.596457 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"c77eb68a-e396-4958-8657-8e62013b774e","Type":"ContainerStarted","Data":"df5b6d61815364484801d77b7a4566b44530d7d7d1c8458423b9e608afb516b0"} Mar 07 15:09:34 crc kubenswrapper[4943]: I0307 15:09:34.596468 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"c77eb68a-e396-4958-8657-8e62013b774e","Type":"ContainerStarted","Data":"d55a2ed66e4f92fd6eca61937a726f238712973499b4a4cefc7679cf37e35d98"} Mar 07 15:09:34 crc kubenswrapper[4943]: I0307 15:09:34.596476 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"c77eb68a-e396-4958-8657-8e62013b774e","Type":"ContainerStarted","Data":"64e04ffc668989d0fc9aa0a7fbd67837640f118c00251f7a7c9f16a103a7c53f"} Mar 07 15:09:34 crc kubenswrapper[4943]: I0307 15:09:34.596484 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"c77eb68a-e396-4958-8657-8e62013b774e","Type":"ContainerStarted","Data":"c5081e6f839f056f7669b020a16f8c2063f2cbe2d70457b95e06168421eda92c"} Mar 07 15:09:34 crc kubenswrapper[4943]: I0307 15:09:34.596493 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"c77eb68a-e396-4958-8657-8e62013b774e","Type":"ContainerStarted","Data":"ee5238926b53decee931446e8bd5b441e5a679b98dc02d0ff3433bb25ee65702"} Mar 07 15:09:34 crc kubenswrapper[4943]: I0307 15:09:34.608154 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"40da0e9b-e661-487c-8898-a54286149f85","Type":"ContainerStarted","Data":"939f27df9ba049ae9a95206b91f9421bda0aabd20166312edb5b571c9ee76db0"} Mar 07 15:09:34 crc kubenswrapper[4943]: I0307 15:09:34.608201 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"40da0e9b-e661-487c-8898-a54286149f85","Type":"ContainerStarted","Data":"12acab80c53f2bbaca75658a6344fd322ddcf4a9221236b26247da4e6a724967"} Mar 07 15:09:34 crc kubenswrapper[4943]: I0307 15:09:34.608212 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"40da0e9b-e661-487c-8898-a54286149f85","Type":"ContainerStarted","Data":"a440900ccc716cb786a3ef4ab08eb8ef73b335f3bbcf4a5fa9b824fe095871a5"} Mar 07 15:09:34 crc kubenswrapper[4943]: I0307 15:09:34.608221 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"40da0e9b-e661-487c-8898-a54286149f85","Type":"ContainerStarted","Data":"7f1bcff16c16fbd8f7875320f8f3515b39e46822770e7a4a67f077ec7c828314"} Mar 07 15:09:34 crc kubenswrapper[4943]: I0307 15:09:34.608232 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"40da0e9b-e661-487c-8898-a54286149f85","Type":"ContainerStarted","Data":"f08918de76869f935ad3c9da3960249fd5aa2248511f6c23bf11168f108c424e"} Mar 07 15:09:34 crc kubenswrapper[4943]: I0307 15:09:34.610264 4943 generic.go:334] "Generic (PLEG): container finished" podID="beeb25d9-4a5d-4e7c-a148-55c164a99704" containerID="0b3d2f580d1445cec8ffea6df297252f539ecf099f6b63cb72727022bdf8093d" exitCode=0 Mar 07 15:09:34 crc kubenswrapper[4943]: I0307 15:09:34.610464 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-khdj6" event={"ID":"beeb25d9-4a5d-4e7c-a148-55c164a99704","Type":"ContainerDied","Data":"0b3d2f580d1445cec8ffea6df297252f539ecf099f6b63cb72727022bdf8093d"} Mar 07 15:09:34 crc kubenswrapper[4943]: I0307 15:09:34.633502 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-proxy-76c998454c-kngkw" podStartSLOduration=2.633480792 podStartE2EDuration="2.633480792s" podCreationTimestamp="2026-03-07 15:09:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:09:33.615172787 +0000 UTC m=+1815.567309285" watchObservedRunningTime="2026-03-07 15:09:34.633480792 +0000 UTC m=+1816.585617300" Mar 07 15:09:35 crc kubenswrapper[4943]: I0307 15:09:35.624636 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"c77eb68a-e396-4958-8657-8e62013b774e","Type":"ContainerStarted","Data":"95434302f8337cff157b53fb64b5ecb9f80062e3117f982f8defc4616dc778d4"} Mar 07 15:09:35 crc kubenswrapper[4943]: I0307 15:09:35.624694 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"c77eb68a-e396-4958-8657-8e62013b774e","Type":"ContainerStarted","Data":"fe5802e7c45509ab5b15e45e06fc126fd5c0fabc243b88237cb21ba70984c51e"} Mar 07 15:09:35 crc kubenswrapper[4943]: I0307 15:09:35.624704 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"c77eb68a-e396-4958-8657-8e62013b774e","Type":"ContainerStarted","Data":"480701a754d395c09cb92218f86706ff9deca84f1dd205605262a19bc0496b76"} Mar 07 15:09:35 crc kubenswrapper[4943]: I0307 15:09:35.624711 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"c77eb68a-e396-4958-8657-8e62013b774e","Type":"ContainerStarted","Data":"3840cdd6bfbeb63e204cef19908a3554010dab72daf44004d87d1a999b960c88"} Mar 07 15:09:35 crc kubenswrapper[4943]: I0307 15:09:35.624722 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"c77eb68a-e396-4958-8657-8e62013b774e","Type":"ContainerStarted","Data":"83da4699a9cb42e4f645907d9a026e2b30ed9cf2b9ba8f866c7a670582f31f78"} Mar 07 15:09:35 crc kubenswrapper[4943]: I0307 15:09:35.624733 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"c77eb68a-e396-4958-8657-8e62013b774e","Type":"ContainerStarted","Data":"b835f9e32153fe05a051ac1cbab09cd5912c710a7f110a3af56fdb702a8cb3f6"} Mar 07 15:09:35 crc kubenswrapper[4943]: I0307 15:09:35.631062 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"40da0e9b-e661-487c-8898-a54286149f85","Type":"ContainerStarted","Data":"62f184290010649808823e131e5a7eea0732420c8f7ffaaf9319ba64a925c4fc"} Mar 07 15:09:35 crc kubenswrapper[4943]: I0307 15:09:35.631148 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"40da0e9b-e661-487c-8898-a54286149f85","Type":"ContainerStarted","Data":"e0bde4b893898878f18d8b8435e5ea916b4bf1299b71ed580fdcc0e3dc6aed5b"} Mar 07 15:09:35 crc kubenswrapper[4943]: I0307 15:09:35.631169 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"40da0e9b-e661-487c-8898-a54286149f85","Type":"ContainerStarted","Data":"33d07436d0df915a99faf9a80787a3dc8ecd921694fa2daaf7a14ed1d5251892"} Mar 07 15:09:35 crc kubenswrapper[4943]: I0307 15:09:35.631249 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"40da0e9b-e661-487c-8898-a54286149f85","Type":"ContainerStarted","Data":"cfdce2ae515747c28989a9d4ffa137f36c5249231dd6ec975d4d81cf1ea7dee1"} Mar 07 15:09:35 crc kubenswrapper[4943]: I0307 15:09:35.631269 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"40da0e9b-e661-487c-8898-a54286149f85","Type":"ContainerStarted","Data":"72b78e68f059ec897322ad001ece625cfb357a0c8e0a8bbdbc3c0a3753d4b2d6"} Mar 07 15:09:35 crc kubenswrapper[4943]: I0307 15:09:35.631286 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"40da0e9b-e661-487c-8898-a54286149f85","Type":"ContainerStarted","Data":"8bf2ed07927525786e508b81b49690874823a5b76192111e5720239d82b2254c"} Mar 07 15:09:35 crc kubenswrapper[4943]: I0307 15:09:35.677785 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-storage-2" podStartSLOduration=5.677766465 podStartE2EDuration="5.677766465s" podCreationTimestamp="2026-03-07 15:09:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:09:35.676396831 +0000 UTC m=+1817.628533349" watchObservedRunningTime="2026-03-07 15:09:35.677766465 +0000 UTC m=+1817.629902963" Mar 07 15:09:35 crc kubenswrapper[4943]: I0307 15:09:35.719805 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-storage-1" podStartSLOduration=5.7197768159999995 podStartE2EDuration="5.719776816s" podCreationTimestamp="2026-03-07 15:09:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:09:35.715993333 +0000 UTC m=+1817.668129851" watchObservedRunningTime="2026-03-07 15:09:35.719776816 +0000 UTC m=+1817.671913324" Mar 07 15:09:35 crc kubenswrapper[4943]: I0307 15:09:35.742001 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-storage-0"] Mar 07 15:09:35 crc kubenswrapper[4943]: I0307 15:09:35.742829 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="account-server" containerID="cri-o://9b6765692b58c5ebd2755fcbcc0bf6fee3ddbe781cfa1b626ad6078346d3e402" gracePeriod=30 Mar 07 15:09:35 crc kubenswrapper[4943]: I0307 15:09:35.743105 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="account-auditor" containerID="cri-o://5a68e14eb4c97f4f249f077f9c0d016c858544f0da60e1d6c84ee0b53bc3a259" gracePeriod=30 Mar 07 15:09:35 crc kubenswrapper[4943]: I0307 15:09:35.742980 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="rsync" containerID="cri-o://b96feec40ee02a1bd0a7a1c8d67abf8300f9b59dbae2f6e54e6ae30df4a3971b" gracePeriod=30 Mar 07 15:09:35 crc kubenswrapper[4943]: I0307 15:09:35.743002 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="container-replicator" containerID="cri-o://a269396243ea039c5a1ff714244738a9af9f115de9be8ca31a6f67a9a2bb2e94" gracePeriod=30 Mar 07 15:09:35 crc kubenswrapper[4943]: I0307 15:09:35.743044 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="object-server" containerID="cri-o://4009d66a16af4cf49d7b909280c182c6a2839bf7dc72660a38b3c02320550d9f" gracePeriod=30 Mar 07 15:09:35 crc kubenswrapper[4943]: I0307 15:09:35.743019 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="object-replicator" containerID="cri-o://a75dfedf2adeb9783f42686241f798122e938a4edd4336b201d2e124ee87e1da" gracePeriod=30 Mar 07 15:09:35 crc kubenswrapper[4943]: I0307 15:09:35.743010 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="object-updater" containerID="cri-o://a3b3ff008ced9b38eb096a8fe7d155531763f2bdf2e5399ff2af469f95e76448" gracePeriod=30 Mar 07 15:09:35 crc kubenswrapper[4943]: I0307 15:09:35.743063 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="container-updater" containerID="cri-o://40df134ae24e3ed4a8fd377fb725953a304574200ac3148e3f9b3b3806e44a47" gracePeriod=30 Mar 07 15:09:35 crc kubenswrapper[4943]: I0307 15:09:35.743074 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="object-expirer" containerID="cri-o://f290059bb335b82ed354d4d229fb8530d57e7f29a151602dd18122d411960328" gracePeriod=30 Mar 07 15:09:35 crc kubenswrapper[4943]: I0307 15:09:35.743080 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="account-replicator" containerID="cri-o://c16bdb03bef6cc9ab7fb55a6953d241e9190e80787f4e3de43c893cd1ed349a2" gracePeriod=30 Mar 07 15:09:35 crc kubenswrapper[4943]: I0307 15:09:35.743071 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="container-auditor" containerID="cri-o://cc374a57378de01b25998555023a72a91a81bfe6f48e0c2d74901d09dc497799" gracePeriod=30 Mar 07 15:09:35 crc kubenswrapper[4943]: I0307 15:09:35.743091 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="container-server" containerID="cri-o://059e70bef0f14c03d2e83f29ffb3a0153b9683edc84e88eed7fa99bc0a0d8579" gracePeriod=30 Mar 07 15:09:35 crc kubenswrapper[4943]: I0307 15:09:35.743087 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="account-reaper" containerID="cri-o://f4a58ec1a6b1a5a66faa2ca708935002ee19db5bef2c8c99a431c0bb5ec9eea3" gracePeriod=30 Mar 07 15:09:35 crc kubenswrapper[4943]: I0307 15:09:35.743096 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="object-auditor" containerID="cri-o://3d606e2c04c5358f4c4404f1af78027ac23f0b1d763a512e4323fdd0240efd19" gracePeriod=30 Mar 07 15:09:35 crc kubenswrapper[4943]: I0307 15:09:35.743032 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="container-sharder" containerID="cri-o://a1e3e8dc6afbf3013626e71f36dcde17e61c0ed8650ec0a48d8c2ddccb2aa585" gracePeriod=30 Mar 07 15:09:35 crc kubenswrapper[4943]: I0307 15:09:35.743046 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="swift-recon-cron" containerID="cri-o://06b728b91732e6aa41de478ccef9fb0046a86c5a1dc8baedd3243f9a3c0ed18c" gracePeriod=30 Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.000894 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-khdj6" Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.041963 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-khdj6"] Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.047578 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-khdj6"] Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.130246 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/beeb25d9-4a5d-4e7c-a148-55c164a99704-etc-swift\") pod \"beeb25d9-4a5d-4e7c-a148-55c164a99704\" (UID: \"beeb25d9-4a5d-4e7c-a148-55c164a99704\") " Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.130589 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/beeb25d9-4a5d-4e7c-a148-55c164a99704-scripts\") pod \"beeb25d9-4a5d-4e7c-a148-55c164a99704\" (UID: \"beeb25d9-4a5d-4e7c-a148-55c164a99704\") " Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.130843 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lfb4c\" (UniqueName: \"kubernetes.io/projected/beeb25d9-4a5d-4e7c-a148-55c164a99704-kube-api-access-lfb4c\") pod \"beeb25d9-4a5d-4e7c-a148-55c164a99704\" (UID: \"beeb25d9-4a5d-4e7c-a148-55c164a99704\") " Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.131072 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/beeb25d9-4a5d-4e7c-a148-55c164a99704-dispersionconf\") pod \"beeb25d9-4a5d-4e7c-a148-55c164a99704\" (UID: \"beeb25d9-4a5d-4e7c-a148-55c164a99704\") " Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.131283 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/beeb25d9-4a5d-4e7c-a148-55c164a99704-swiftconf\") pod \"beeb25d9-4a5d-4e7c-a148-55c164a99704\" (UID: \"beeb25d9-4a5d-4e7c-a148-55c164a99704\") " Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.131653 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/beeb25d9-4a5d-4e7c-a148-55c164a99704-ring-data-devices\") pod \"beeb25d9-4a5d-4e7c-a148-55c164a99704\" (UID: \"beeb25d9-4a5d-4e7c-a148-55c164a99704\") " Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.131665 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/beeb25d9-4a5d-4e7c-a148-55c164a99704-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "beeb25d9-4a5d-4e7c-a148-55c164a99704" (UID: "beeb25d9-4a5d-4e7c-a148-55c164a99704"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.132264 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/beeb25d9-4a5d-4e7c-a148-55c164a99704-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "beeb25d9-4a5d-4e7c-a148-55c164a99704" (UID: "beeb25d9-4a5d-4e7c-a148-55c164a99704"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.132858 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/beeb25d9-4a5d-4e7c-a148-55c164a99704-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.133055 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/beeb25d9-4a5d-4e7c-a148-55c164a99704-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.137390 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/beeb25d9-4a5d-4e7c-a148-55c164a99704-kube-api-access-lfb4c" (OuterVolumeSpecName: "kube-api-access-lfb4c") pod "beeb25d9-4a5d-4e7c-a148-55c164a99704" (UID: "beeb25d9-4a5d-4e7c-a148-55c164a99704"). InnerVolumeSpecName "kube-api-access-lfb4c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.153872 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/beeb25d9-4a5d-4e7c-a148-55c164a99704-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "beeb25d9-4a5d-4e7c-a148-55c164a99704" (UID: "beeb25d9-4a5d-4e7c-a148-55c164a99704"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.154976 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/beeb25d9-4a5d-4e7c-a148-55c164a99704-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "beeb25d9-4a5d-4e7c-a148-55c164a99704" (UID: "beeb25d9-4a5d-4e7c-a148-55c164a99704"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.155270 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/beeb25d9-4a5d-4e7c-a148-55c164a99704-scripts" (OuterVolumeSpecName: "scripts") pod "beeb25d9-4a5d-4e7c-a148-55c164a99704" (UID: "beeb25d9-4a5d-4e7c-a148-55c164a99704"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.210073 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-5lvgv"] Mar 07 15:09:36 crc kubenswrapper[4943]: E0307 15:09:36.210367 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="beeb25d9-4a5d-4e7c-a148-55c164a99704" containerName="swift-ring-rebalance" Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.210384 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="beeb25d9-4a5d-4e7c-a148-55c164a99704" containerName="swift-ring-rebalance" Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.210519 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="beeb25d9-4a5d-4e7c-a148-55c164a99704" containerName="swift-ring-rebalance" Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.211020 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5lvgv" Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.225858 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-5lvgv"] Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.235307 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lfb4c\" (UniqueName: \"kubernetes.io/projected/beeb25d9-4a5d-4e7c-a148-55c164a99704-kube-api-access-lfb4c\") on node \"crc\" DevicePath \"\"" Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.235353 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/beeb25d9-4a5d-4e7c-a148-55c164a99704-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.235371 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/beeb25d9-4a5d-4e7c-a148-55c164a99704-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.235388 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/beeb25d9-4a5d-4e7c-a148-55c164a99704-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.337071 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/de11450f-e45d-48d4-99a7-89907ae9da71-etc-swift\") pod \"swift-ring-rebalance-debug-5lvgv\" (UID: \"de11450f-e45d-48d4-99a7-89907ae9da71\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5lvgv" Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.337400 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htmds\" (UniqueName: \"kubernetes.io/projected/de11450f-e45d-48d4-99a7-89907ae9da71-kube-api-access-htmds\") pod \"swift-ring-rebalance-debug-5lvgv\" (UID: \"de11450f-e45d-48d4-99a7-89907ae9da71\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5lvgv" Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.337576 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/de11450f-e45d-48d4-99a7-89907ae9da71-scripts\") pod \"swift-ring-rebalance-debug-5lvgv\" (UID: \"de11450f-e45d-48d4-99a7-89907ae9da71\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5lvgv" Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.337741 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/de11450f-e45d-48d4-99a7-89907ae9da71-swiftconf\") pod \"swift-ring-rebalance-debug-5lvgv\" (UID: \"de11450f-e45d-48d4-99a7-89907ae9da71\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5lvgv" Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.338001 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/de11450f-e45d-48d4-99a7-89907ae9da71-dispersionconf\") pod \"swift-ring-rebalance-debug-5lvgv\" (UID: \"de11450f-e45d-48d4-99a7-89907ae9da71\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5lvgv" Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.338170 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/de11450f-e45d-48d4-99a7-89907ae9da71-ring-data-devices\") pod \"swift-ring-rebalance-debug-5lvgv\" (UID: \"de11450f-e45d-48d4-99a7-89907ae9da71\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5lvgv" Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.439542 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/de11450f-e45d-48d4-99a7-89907ae9da71-etc-swift\") pod \"swift-ring-rebalance-debug-5lvgv\" (UID: \"de11450f-e45d-48d4-99a7-89907ae9da71\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5lvgv" Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.439694 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htmds\" (UniqueName: \"kubernetes.io/projected/de11450f-e45d-48d4-99a7-89907ae9da71-kube-api-access-htmds\") pod \"swift-ring-rebalance-debug-5lvgv\" (UID: \"de11450f-e45d-48d4-99a7-89907ae9da71\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5lvgv" Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.439768 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/de11450f-e45d-48d4-99a7-89907ae9da71-scripts\") pod \"swift-ring-rebalance-debug-5lvgv\" (UID: \"de11450f-e45d-48d4-99a7-89907ae9da71\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5lvgv" Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.439815 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/de11450f-e45d-48d4-99a7-89907ae9da71-swiftconf\") pod \"swift-ring-rebalance-debug-5lvgv\" (UID: \"de11450f-e45d-48d4-99a7-89907ae9da71\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5lvgv" Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.439878 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/de11450f-e45d-48d4-99a7-89907ae9da71-dispersionconf\") pod \"swift-ring-rebalance-debug-5lvgv\" (UID: \"de11450f-e45d-48d4-99a7-89907ae9da71\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5lvgv" Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.439908 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/de11450f-e45d-48d4-99a7-89907ae9da71-ring-data-devices\") pod \"swift-ring-rebalance-debug-5lvgv\" (UID: \"de11450f-e45d-48d4-99a7-89907ae9da71\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5lvgv" Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.440410 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/de11450f-e45d-48d4-99a7-89907ae9da71-etc-swift\") pod \"swift-ring-rebalance-debug-5lvgv\" (UID: \"de11450f-e45d-48d4-99a7-89907ae9da71\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5lvgv" Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.441122 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/de11450f-e45d-48d4-99a7-89907ae9da71-ring-data-devices\") pod \"swift-ring-rebalance-debug-5lvgv\" (UID: \"de11450f-e45d-48d4-99a7-89907ae9da71\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5lvgv" Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.441140 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/de11450f-e45d-48d4-99a7-89907ae9da71-scripts\") pod \"swift-ring-rebalance-debug-5lvgv\" (UID: \"de11450f-e45d-48d4-99a7-89907ae9da71\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5lvgv" Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.447488 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/de11450f-e45d-48d4-99a7-89907ae9da71-swiftconf\") pod \"swift-ring-rebalance-debug-5lvgv\" (UID: \"de11450f-e45d-48d4-99a7-89907ae9da71\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5lvgv" Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.447836 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/de11450f-e45d-48d4-99a7-89907ae9da71-dispersionconf\") pod \"swift-ring-rebalance-debug-5lvgv\" (UID: \"de11450f-e45d-48d4-99a7-89907ae9da71\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5lvgv" Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.465608 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htmds\" (UniqueName: \"kubernetes.io/projected/de11450f-e45d-48d4-99a7-89907ae9da71-kube-api-access-htmds\") pod \"swift-ring-rebalance-debug-5lvgv\" (UID: \"de11450f-e45d-48d4-99a7-89907ae9da71\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5lvgv" Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.579462 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5lvgv" Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.649244 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a2620b80e90a4294781c40c4a246ae281d4654be0bb7fd3ead613ca59f1c518" Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.650738 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-khdj6" Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.665623 4943 generic.go:334] "Generic (PLEG): container finished" podID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerID="a1e3e8dc6afbf3013626e71f36dcde17e61c0ed8650ec0a48d8c2ddccb2aa585" exitCode=0 Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.665667 4943 generic.go:334] "Generic (PLEG): container finished" podID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerID="b96feec40ee02a1bd0a7a1c8d67abf8300f9b59dbae2f6e54e6ae30df4a3971b" exitCode=0 Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.665689 4943 generic.go:334] "Generic (PLEG): container finished" podID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerID="f290059bb335b82ed354d4d229fb8530d57e7f29a151602dd18122d411960328" exitCode=0 Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.665711 4943 generic.go:334] "Generic (PLEG): container finished" podID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerID="a3b3ff008ced9b38eb096a8fe7d155531763f2bdf2e5399ff2af469f95e76448" exitCode=0 Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.665724 4943 generic.go:334] "Generic (PLEG): container finished" podID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerID="3d606e2c04c5358f4c4404f1af78027ac23f0b1d763a512e4323fdd0240efd19" exitCode=0 Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.665737 4943 generic.go:334] "Generic (PLEG): container finished" podID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerID="a75dfedf2adeb9783f42686241f798122e938a4edd4336b201d2e124ee87e1da" exitCode=0 Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.665749 4943 generic.go:334] "Generic (PLEG): container finished" podID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerID="4009d66a16af4cf49d7b909280c182c6a2839bf7dc72660a38b3c02320550d9f" exitCode=0 Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.665764 4943 generic.go:334] "Generic (PLEG): container finished" podID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerID="40df134ae24e3ed4a8fd377fb725953a304574200ac3148e3f9b3b3806e44a47" exitCode=0 Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.665735 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8","Type":"ContainerDied","Data":"a1e3e8dc6afbf3013626e71f36dcde17e61c0ed8650ec0a48d8c2ddccb2aa585"} Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.665837 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8","Type":"ContainerDied","Data":"b96feec40ee02a1bd0a7a1c8d67abf8300f9b59dbae2f6e54e6ae30df4a3971b"} Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.665864 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8","Type":"ContainerDied","Data":"f290059bb335b82ed354d4d229fb8530d57e7f29a151602dd18122d411960328"} Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.665884 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8","Type":"ContainerDied","Data":"a3b3ff008ced9b38eb096a8fe7d155531763f2bdf2e5399ff2af469f95e76448"} Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.665903 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8","Type":"ContainerDied","Data":"3d606e2c04c5358f4c4404f1af78027ac23f0b1d763a512e4323fdd0240efd19"} Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.665781 4943 generic.go:334] "Generic (PLEG): container finished" podID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerID="cc374a57378de01b25998555023a72a91a81bfe6f48e0c2d74901d09dc497799" exitCode=0 Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.665921 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8","Type":"ContainerDied","Data":"a75dfedf2adeb9783f42686241f798122e938a4edd4336b201d2e124ee87e1da"} Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.665971 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8","Type":"ContainerDied","Data":"4009d66a16af4cf49d7b909280c182c6a2839bf7dc72660a38b3c02320550d9f"} Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.666003 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8","Type":"ContainerDied","Data":"40df134ae24e3ed4a8fd377fb725953a304574200ac3148e3f9b3b3806e44a47"} Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.666029 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8","Type":"ContainerDied","Data":"cc374a57378de01b25998555023a72a91a81bfe6f48e0c2d74901d09dc497799"} Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.666055 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8","Type":"ContainerDied","Data":"a269396243ea039c5a1ff714244738a9af9f115de9be8ca31a6f67a9a2bb2e94"} Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.665977 4943 generic.go:334] "Generic (PLEG): container finished" podID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerID="a269396243ea039c5a1ff714244738a9af9f115de9be8ca31a6f67a9a2bb2e94" exitCode=0 Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.666101 4943 generic.go:334] "Generic (PLEG): container finished" podID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerID="059e70bef0f14c03d2e83f29ffb3a0153b9683edc84e88eed7fa99bc0a0d8579" exitCode=0 Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.666121 4943 generic.go:334] "Generic (PLEG): container finished" podID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerID="f4a58ec1a6b1a5a66faa2ca708935002ee19db5bef2c8c99a431c0bb5ec9eea3" exitCode=0 Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.666135 4943 generic.go:334] "Generic (PLEG): container finished" podID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerID="5a68e14eb4c97f4f249f077f9c0d016c858544f0da60e1d6c84ee0b53bc3a259" exitCode=0 Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.666152 4943 generic.go:334] "Generic (PLEG): container finished" podID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerID="c16bdb03bef6cc9ab7fb55a6953d241e9190e80787f4e3de43c893cd1ed349a2" exitCode=0 Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.666167 4943 generic.go:334] "Generic (PLEG): container finished" podID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerID="9b6765692b58c5ebd2755fcbcc0bf6fee3ddbe781cfa1b626ad6078346d3e402" exitCode=0 Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.666164 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8","Type":"ContainerDied","Data":"059e70bef0f14c03d2e83f29ffb3a0153b9683edc84e88eed7fa99bc0a0d8579"} Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.666256 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8","Type":"ContainerDied","Data":"f4a58ec1a6b1a5a66faa2ca708935002ee19db5bef2c8c99a431c0bb5ec9eea3"} Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.666302 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8","Type":"ContainerDied","Data":"5a68e14eb4c97f4f249f077f9c0d016c858544f0da60e1d6c84ee0b53bc3a259"} Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.666323 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8","Type":"ContainerDied","Data":"c16bdb03bef6cc9ab7fb55a6953d241e9190e80787f4e3de43c893cd1ed349a2"} Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.666341 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8","Type":"ContainerDied","Data":"9b6765692b58c5ebd2755fcbcc0bf6fee3ddbe781cfa1b626ad6078346d3e402"} Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.765468 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="beeb25d9-4a5d-4e7c-a148-55c164a99704" path="/var/lib/kubelet/pods/beeb25d9-4a5d-4e7c-a148-55c164a99704/volumes" Mar 07 15:09:36 crc kubenswrapper[4943]: I0307 15:09:36.882189 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-5lvgv"] Mar 07 15:09:36 crc kubenswrapper[4943]: W0307 15:09:36.882323 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podde11450f_e45d_48d4_99a7_89907ae9da71.slice/crio-4093b73774d310ae50f7bfe7a6ff77a8443c0b62241e17dc23fccd6755d59c20 WatchSource:0}: Error finding container 4093b73774d310ae50f7bfe7a6ff77a8443c0b62241e17dc23fccd6755d59c20: Status 404 returned error can't find the container with id 4093b73774d310ae50f7bfe7a6ff77a8443c0b62241e17dc23fccd6755d59c20 Mar 07 15:09:37 crc kubenswrapper[4943]: I0307 15:09:37.257174 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-5lvgv"] Mar 07 15:09:37 crc kubenswrapper[4943]: I0307 15:09:37.675774 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5lvgv" event={"ID":"de11450f-e45d-48d4-99a7-89907ae9da71","Type":"ContainerStarted","Data":"9a27b9b16b6a885e47f4731dee29920caf08019a737681a20b88cbe0375277cc"} Mar 07 15:09:37 crc kubenswrapper[4943]: I0307 15:09:37.676073 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5lvgv" event={"ID":"de11450f-e45d-48d4-99a7-89907ae9da71","Type":"ContainerStarted","Data":"4093b73774d310ae50f7bfe7a6ff77a8443c0b62241e17dc23fccd6755d59c20"} Mar 07 15:09:37 crc kubenswrapper[4943]: I0307 15:09:37.700437 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5lvgv" podStartSLOduration=1.700418421 podStartE2EDuration="1.700418421s" podCreationTimestamp="2026-03-07 15:09:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:09:37.694810064 +0000 UTC m=+1819.646946552" watchObservedRunningTime="2026-03-07 15:09:37.700418421 +0000 UTC m=+1819.652554919" Mar 07 15:09:38 crc kubenswrapper[4943]: I0307 15:09:38.685992 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5lvgv" podUID="de11450f-e45d-48d4-99a7-89907ae9da71" containerName="swift-ring-rebalance" containerID="cri-o://9a27b9b16b6a885e47f4731dee29920caf08019a737681a20b88cbe0375277cc" gracePeriod=30 Mar 07 15:09:41 crc kubenswrapper[4943]: I0307 15:09:41.721648 4943 generic.go:334] "Generic (PLEG): container finished" podID="85e1e3e8-b9f8-4c3d-940e-4060262201ee" containerID="554f64008e702a987e22726c874147b00d5ca6af3bd7acae9d07bdab80612bb1" exitCode=0 Mar 07 15:09:41 crc kubenswrapper[4943]: I0307 15:09:41.721733 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-s5kqz" event={"ID":"85e1e3e8-b9f8-4c3d-940e-4060262201ee","Type":"ContainerDied","Data":"554f64008e702a987e22726c874147b00d5ca6af3bd7acae9d07bdab80612bb1"} Mar 07 15:09:42 crc kubenswrapper[4943]: I0307 15:09:42.379487 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="swift-kuttl-tests/swift-proxy-76c998454c-kngkw" Mar 07 15:09:42 crc kubenswrapper[4943]: I0307 15:09:42.380668 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="swift-kuttl-tests/swift-proxy-76c998454c-kngkw" Mar 07 15:09:42 crc kubenswrapper[4943]: I0307 15:09:42.498642 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-proxy-646594d5c9-q4vdw"] Mar 07 15:09:42 crc kubenswrapper[4943]: I0307 15:09:42.498967 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-proxy-646594d5c9-q4vdw" podUID="6d17441e-f6b6-4fab-8239-0409b7bb3910" containerName="proxy-httpd" containerID="cri-o://ea89ba800a6d23b6b4bc4c690cd8df8bb7d32b6ee9b1f7bebbebd0d0592f47f9" gracePeriod=30 Mar 07 15:09:42 crc kubenswrapper[4943]: I0307 15:09:42.499399 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-proxy-646594d5c9-q4vdw" podUID="6d17441e-f6b6-4fab-8239-0409b7bb3910" containerName="proxy-server" containerID="cri-o://be38a7d18eaf729bc3bfa860152f8612e983a3ab4366c7882bc771e0466ac733" gracePeriod=30 Mar 07 15:09:42 crc kubenswrapper[4943]: I0307 15:09:42.732699 4943 generic.go:334] "Generic (PLEG): container finished" podID="6d17441e-f6b6-4fab-8239-0409b7bb3910" containerID="ea89ba800a6d23b6b4bc4c690cd8df8bb7d32b6ee9b1f7bebbebd0d0592f47f9" exitCode=0 Mar 07 15:09:42 crc kubenswrapper[4943]: I0307 15:09:42.732781 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-proxy-646594d5c9-q4vdw" event={"ID":"6d17441e-f6b6-4fab-8239-0409b7bb3910","Type":"ContainerDied","Data":"ea89ba800a6d23b6b4bc4c690cd8df8bb7d32b6ee9b1f7bebbebd0d0592f47f9"} Mar 07 15:09:43 crc kubenswrapper[4943]: I0307 15:09:43.059253 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-s5kqz" Mar 07 15:09:43 crc kubenswrapper[4943]: I0307 15:09:43.153332 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/85e1e3e8-b9f8-4c3d-940e-4060262201ee-ring-data-devices\") pod \"85e1e3e8-b9f8-4c3d-940e-4060262201ee\" (UID: \"85e1e3e8-b9f8-4c3d-940e-4060262201ee\") " Mar 07 15:09:43 crc kubenswrapper[4943]: I0307 15:09:43.153392 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/85e1e3e8-b9f8-4c3d-940e-4060262201ee-scripts\") pod \"85e1e3e8-b9f8-4c3d-940e-4060262201ee\" (UID: \"85e1e3e8-b9f8-4c3d-940e-4060262201ee\") " Mar 07 15:09:43 crc kubenswrapper[4943]: I0307 15:09:43.153431 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/85e1e3e8-b9f8-4c3d-940e-4060262201ee-etc-swift\") pod \"85e1e3e8-b9f8-4c3d-940e-4060262201ee\" (UID: \"85e1e3e8-b9f8-4c3d-940e-4060262201ee\") " Mar 07 15:09:43 crc kubenswrapper[4943]: I0307 15:09:43.153502 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/85e1e3e8-b9f8-4c3d-940e-4060262201ee-swiftconf\") pod \"85e1e3e8-b9f8-4c3d-940e-4060262201ee\" (UID: \"85e1e3e8-b9f8-4c3d-940e-4060262201ee\") " Mar 07 15:09:43 crc kubenswrapper[4943]: I0307 15:09:43.153527 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/85e1e3e8-b9f8-4c3d-940e-4060262201ee-dispersionconf\") pod \"85e1e3e8-b9f8-4c3d-940e-4060262201ee\" (UID: \"85e1e3e8-b9f8-4c3d-940e-4060262201ee\") " Mar 07 15:09:43 crc kubenswrapper[4943]: I0307 15:09:43.153543 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r52nz\" (UniqueName: \"kubernetes.io/projected/85e1e3e8-b9f8-4c3d-940e-4060262201ee-kube-api-access-r52nz\") pod \"85e1e3e8-b9f8-4c3d-940e-4060262201ee\" (UID: \"85e1e3e8-b9f8-4c3d-940e-4060262201ee\") " Mar 07 15:09:43 crc kubenswrapper[4943]: I0307 15:09:43.155129 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85e1e3e8-b9f8-4c3d-940e-4060262201ee-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "85e1e3e8-b9f8-4c3d-940e-4060262201ee" (UID: "85e1e3e8-b9f8-4c3d-940e-4060262201ee"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:09:43 crc kubenswrapper[4943]: I0307 15:09:43.156314 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85e1e3e8-b9f8-4c3d-940e-4060262201ee-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "85e1e3e8-b9f8-4c3d-940e-4060262201ee" (UID: "85e1e3e8-b9f8-4c3d-940e-4060262201ee"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:09:43 crc kubenswrapper[4943]: I0307 15:09:43.159822 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85e1e3e8-b9f8-4c3d-940e-4060262201ee-kube-api-access-r52nz" (OuterVolumeSpecName: "kube-api-access-r52nz") pod "85e1e3e8-b9f8-4c3d-940e-4060262201ee" (UID: "85e1e3e8-b9f8-4c3d-940e-4060262201ee"). InnerVolumeSpecName "kube-api-access-r52nz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:09:43 crc kubenswrapper[4943]: I0307 15:09:43.173484 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85e1e3e8-b9f8-4c3d-940e-4060262201ee-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "85e1e3e8-b9f8-4c3d-940e-4060262201ee" (UID: "85e1e3e8-b9f8-4c3d-940e-4060262201ee"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:09:43 crc kubenswrapper[4943]: I0307 15:09:43.177923 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85e1e3e8-b9f8-4c3d-940e-4060262201ee-scripts" (OuterVolumeSpecName: "scripts") pod "85e1e3e8-b9f8-4c3d-940e-4060262201ee" (UID: "85e1e3e8-b9f8-4c3d-940e-4060262201ee"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:09:43 crc kubenswrapper[4943]: I0307 15:09:43.196461 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85e1e3e8-b9f8-4c3d-940e-4060262201ee-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "85e1e3e8-b9f8-4c3d-940e-4060262201ee" (UID: "85e1e3e8-b9f8-4c3d-940e-4060262201ee"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:09:43 crc kubenswrapper[4943]: I0307 15:09:43.254596 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/85e1e3e8-b9f8-4c3d-940e-4060262201ee-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:09:43 crc kubenswrapper[4943]: I0307 15:09:43.254796 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/85e1e3e8-b9f8-4c3d-940e-4060262201ee-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:09:43 crc kubenswrapper[4943]: I0307 15:09:43.254808 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r52nz\" (UniqueName: \"kubernetes.io/projected/85e1e3e8-b9f8-4c3d-940e-4060262201ee-kube-api-access-r52nz\") on node \"crc\" DevicePath \"\"" Mar 07 15:09:43 crc kubenswrapper[4943]: I0307 15:09:43.254816 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/85e1e3e8-b9f8-4c3d-940e-4060262201ee-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:09:43 crc kubenswrapper[4943]: I0307 15:09:43.254824 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/85e1e3e8-b9f8-4c3d-940e-4060262201ee-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:09:43 crc kubenswrapper[4943]: I0307 15:09:43.254833 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/85e1e3e8-b9f8-4c3d-940e-4060262201ee-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:09:43 crc kubenswrapper[4943]: I0307 15:09:43.485495 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-proxy-646594d5c9-q4vdw" Mar 07 15:09:43 crc kubenswrapper[4943]: I0307 15:09:43.559400 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6d17441e-f6b6-4fab-8239-0409b7bb3910-run-httpd\") pod \"6d17441e-f6b6-4fab-8239-0409b7bb3910\" (UID: \"6d17441e-f6b6-4fab-8239-0409b7bb3910\") " Mar 07 15:09:43 crc kubenswrapper[4943]: I0307 15:09:43.559515 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6d17441e-f6b6-4fab-8239-0409b7bb3910-etc-swift\") pod \"6d17441e-f6b6-4fab-8239-0409b7bb3910\" (UID: \"6d17441e-f6b6-4fab-8239-0409b7bb3910\") " Mar 07 15:09:43 crc kubenswrapper[4943]: I0307 15:09:43.559579 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9wdb4\" (UniqueName: \"kubernetes.io/projected/6d17441e-f6b6-4fab-8239-0409b7bb3910-kube-api-access-9wdb4\") pod \"6d17441e-f6b6-4fab-8239-0409b7bb3910\" (UID: \"6d17441e-f6b6-4fab-8239-0409b7bb3910\") " Mar 07 15:09:43 crc kubenswrapper[4943]: I0307 15:09:43.559675 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d17441e-f6b6-4fab-8239-0409b7bb3910-config-data\") pod \"6d17441e-f6b6-4fab-8239-0409b7bb3910\" (UID: \"6d17441e-f6b6-4fab-8239-0409b7bb3910\") " Mar 07 15:09:43 crc kubenswrapper[4943]: I0307 15:09:43.559716 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6d17441e-f6b6-4fab-8239-0409b7bb3910-log-httpd\") pod \"6d17441e-f6b6-4fab-8239-0409b7bb3910\" (UID: \"6d17441e-f6b6-4fab-8239-0409b7bb3910\") " Mar 07 15:09:43 crc kubenswrapper[4943]: I0307 15:09:43.559964 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d17441e-f6b6-4fab-8239-0409b7bb3910-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "6d17441e-f6b6-4fab-8239-0409b7bb3910" (UID: "6d17441e-f6b6-4fab-8239-0409b7bb3910"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:09:43 crc kubenswrapper[4943]: I0307 15:09:43.560300 4943 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6d17441e-f6b6-4fab-8239-0409b7bb3910-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 07 15:09:43 crc kubenswrapper[4943]: I0307 15:09:43.560380 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d17441e-f6b6-4fab-8239-0409b7bb3910-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "6d17441e-f6b6-4fab-8239-0409b7bb3910" (UID: "6d17441e-f6b6-4fab-8239-0409b7bb3910"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:09:43 crc kubenswrapper[4943]: I0307 15:09:43.563051 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d17441e-f6b6-4fab-8239-0409b7bb3910-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "6d17441e-f6b6-4fab-8239-0409b7bb3910" (UID: "6d17441e-f6b6-4fab-8239-0409b7bb3910"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:09:43 crc kubenswrapper[4943]: I0307 15:09:43.563401 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d17441e-f6b6-4fab-8239-0409b7bb3910-kube-api-access-9wdb4" (OuterVolumeSpecName: "kube-api-access-9wdb4") pod "6d17441e-f6b6-4fab-8239-0409b7bb3910" (UID: "6d17441e-f6b6-4fab-8239-0409b7bb3910"). InnerVolumeSpecName "kube-api-access-9wdb4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:09:43 crc kubenswrapper[4943]: I0307 15:09:43.626207 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d17441e-f6b6-4fab-8239-0409b7bb3910-config-data" (OuterVolumeSpecName: "config-data") pod "6d17441e-f6b6-4fab-8239-0409b7bb3910" (UID: "6d17441e-f6b6-4fab-8239-0409b7bb3910"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:09:43 crc kubenswrapper[4943]: I0307 15:09:43.661916 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6d17441e-f6b6-4fab-8239-0409b7bb3910-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:09:43 crc kubenswrapper[4943]: I0307 15:09:43.661977 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9wdb4\" (UniqueName: \"kubernetes.io/projected/6d17441e-f6b6-4fab-8239-0409b7bb3910-kube-api-access-9wdb4\") on node \"crc\" DevicePath \"\"" Mar 07 15:09:43 crc kubenswrapper[4943]: I0307 15:09:43.661992 4943 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d17441e-f6b6-4fab-8239-0409b7bb3910-config-data\") on node \"crc\" DevicePath \"\"" Mar 07 15:09:43 crc kubenswrapper[4943]: I0307 15:09:43.662020 4943 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6d17441e-f6b6-4fab-8239-0409b7bb3910-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 07 15:09:43 crc kubenswrapper[4943]: I0307 15:09:43.745715 4943 generic.go:334] "Generic (PLEG): container finished" podID="6d17441e-f6b6-4fab-8239-0409b7bb3910" containerID="be38a7d18eaf729bc3bfa860152f8612e983a3ab4366c7882bc771e0466ac733" exitCode=0 Mar 07 15:09:43 crc kubenswrapper[4943]: I0307 15:09:43.745770 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-proxy-646594d5c9-q4vdw" event={"ID":"6d17441e-f6b6-4fab-8239-0409b7bb3910","Type":"ContainerDied","Data":"be38a7d18eaf729bc3bfa860152f8612e983a3ab4366c7882bc771e0466ac733"} Mar 07 15:09:43 crc kubenswrapper[4943]: I0307 15:09:43.745822 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-proxy-646594d5c9-q4vdw" event={"ID":"6d17441e-f6b6-4fab-8239-0409b7bb3910","Type":"ContainerDied","Data":"5ba7b7da95bb2163a0f0a086f63f0cb2ccc1e9447e39660d8df0ba17604ac53d"} Mar 07 15:09:43 crc kubenswrapper[4943]: I0307 15:09:43.745828 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-proxy-646594d5c9-q4vdw" Mar 07 15:09:43 crc kubenswrapper[4943]: I0307 15:09:43.745839 4943 scope.go:117] "RemoveContainer" containerID="be38a7d18eaf729bc3bfa860152f8612e983a3ab4366c7882bc771e0466ac733" Mar 07 15:09:43 crc kubenswrapper[4943]: I0307 15:09:43.748662 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-s5kqz" event={"ID":"85e1e3e8-b9f8-4c3d-940e-4060262201ee","Type":"ContainerDied","Data":"b17c0d4741b2bdff0e0844e11447e44b3a997fc20dd4268105cec6eebed81d38"} Mar 07 15:09:43 crc kubenswrapper[4943]: I0307 15:09:43.748785 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b17c0d4741b2bdff0e0844e11447e44b3a997fc20dd4268105cec6eebed81d38" Mar 07 15:09:43 crc kubenswrapper[4943]: I0307 15:09:43.748722 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-s5kqz" Mar 07 15:09:43 crc kubenswrapper[4943]: I0307 15:09:43.765557 4943 scope.go:117] "RemoveContainer" containerID="ea89ba800a6d23b6b4bc4c690cd8df8bb7d32b6ee9b1f7bebbebd0d0592f47f9" Mar 07 15:09:43 crc kubenswrapper[4943]: I0307 15:09:43.796990 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-proxy-646594d5c9-q4vdw"] Mar 07 15:09:43 crc kubenswrapper[4943]: I0307 15:09:43.806672 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-proxy-646594d5c9-q4vdw"] Mar 07 15:09:43 crc kubenswrapper[4943]: I0307 15:09:43.810467 4943 scope.go:117] "RemoveContainer" containerID="be38a7d18eaf729bc3bfa860152f8612e983a3ab4366c7882bc771e0466ac733" Mar 07 15:09:43 crc kubenswrapper[4943]: E0307 15:09:43.810976 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be38a7d18eaf729bc3bfa860152f8612e983a3ab4366c7882bc771e0466ac733\": container with ID starting with be38a7d18eaf729bc3bfa860152f8612e983a3ab4366c7882bc771e0466ac733 not found: ID does not exist" containerID="be38a7d18eaf729bc3bfa860152f8612e983a3ab4366c7882bc771e0466ac733" Mar 07 15:09:43 crc kubenswrapper[4943]: I0307 15:09:43.811029 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be38a7d18eaf729bc3bfa860152f8612e983a3ab4366c7882bc771e0466ac733"} err="failed to get container status \"be38a7d18eaf729bc3bfa860152f8612e983a3ab4366c7882bc771e0466ac733\": rpc error: code = NotFound desc = could not find container \"be38a7d18eaf729bc3bfa860152f8612e983a3ab4366c7882bc771e0466ac733\": container with ID starting with be38a7d18eaf729bc3bfa860152f8612e983a3ab4366c7882bc771e0466ac733 not found: ID does not exist" Mar 07 15:09:43 crc kubenswrapper[4943]: I0307 15:09:43.811067 4943 scope.go:117] "RemoveContainer" containerID="ea89ba800a6d23b6b4bc4c690cd8df8bb7d32b6ee9b1f7bebbebd0d0592f47f9" Mar 07 15:09:43 crc kubenswrapper[4943]: E0307 15:09:43.811429 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea89ba800a6d23b6b4bc4c690cd8df8bb7d32b6ee9b1f7bebbebd0d0592f47f9\": container with ID starting with ea89ba800a6d23b6b4bc4c690cd8df8bb7d32b6ee9b1f7bebbebd0d0592f47f9 not found: ID does not exist" containerID="ea89ba800a6d23b6b4bc4c690cd8df8bb7d32b6ee9b1f7bebbebd0d0592f47f9" Mar 07 15:09:43 crc kubenswrapper[4943]: I0307 15:09:43.811452 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea89ba800a6d23b6b4bc4c690cd8df8bb7d32b6ee9b1f7bebbebd0d0592f47f9"} err="failed to get container status \"ea89ba800a6d23b6b4bc4c690cd8df8bb7d32b6ee9b1f7bebbebd0d0592f47f9\": rpc error: code = NotFound desc = could not find container \"ea89ba800a6d23b6b4bc4c690cd8df8bb7d32b6ee9b1f7bebbebd0d0592f47f9\": container with ID starting with ea89ba800a6d23b6b4bc4c690cd8df8bb7d32b6ee9b1f7bebbebd0d0592f47f9 not found: ID does not exist" Mar 07 15:09:44 crc kubenswrapper[4943]: I0307 15:09:44.755252 4943 scope.go:117] "RemoveContainer" containerID="26f35e1f941cca0c34e3650d7aa047834e769dd76d2a61c3a39cad5bb2d090ae" Mar 07 15:09:44 crc kubenswrapper[4943]: E0307 15:09:44.755636 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cpdmm_openshift-machine-config-operator(4c1f18ea-e557-41fc-95dd-ce4e1722a61e)\"" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" Mar 07 15:09:44 crc kubenswrapper[4943]: I0307 15:09:44.765373 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d17441e-f6b6-4fab-8239-0409b7bb3910" path="/var/lib/kubelet/pods/6d17441e-f6b6-4fab-8239-0409b7bb3910/volumes" Mar 07 15:09:56 crc kubenswrapper[4943]: I0307 15:09:56.474369 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5lvgv" Mar 07 15:09:56 crc kubenswrapper[4943]: I0307 15:09:56.587923 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/de11450f-e45d-48d4-99a7-89907ae9da71-ring-data-devices\") pod \"de11450f-e45d-48d4-99a7-89907ae9da71\" (UID: \"de11450f-e45d-48d4-99a7-89907ae9da71\") " Mar 07 15:09:56 crc kubenswrapper[4943]: I0307 15:09:56.588033 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htmds\" (UniqueName: \"kubernetes.io/projected/de11450f-e45d-48d4-99a7-89907ae9da71-kube-api-access-htmds\") pod \"de11450f-e45d-48d4-99a7-89907ae9da71\" (UID: \"de11450f-e45d-48d4-99a7-89907ae9da71\") " Mar 07 15:09:56 crc kubenswrapper[4943]: I0307 15:09:56.588833 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de11450f-e45d-48d4-99a7-89907ae9da71-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "de11450f-e45d-48d4-99a7-89907ae9da71" (UID: "de11450f-e45d-48d4-99a7-89907ae9da71"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:09:56 crc kubenswrapper[4943]: I0307 15:09:56.588072 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/de11450f-e45d-48d4-99a7-89907ae9da71-swiftconf\") pod \"de11450f-e45d-48d4-99a7-89907ae9da71\" (UID: \"de11450f-e45d-48d4-99a7-89907ae9da71\") " Mar 07 15:09:56 crc kubenswrapper[4943]: I0307 15:09:56.588999 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/de11450f-e45d-48d4-99a7-89907ae9da71-etc-swift\") pod \"de11450f-e45d-48d4-99a7-89907ae9da71\" (UID: \"de11450f-e45d-48d4-99a7-89907ae9da71\") " Mar 07 15:09:56 crc kubenswrapper[4943]: I0307 15:09:56.589057 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/de11450f-e45d-48d4-99a7-89907ae9da71-dispersionconf\") pod \"de11450f-e45d-48d4-99a7-89907ae9da71\" (UID: \"de11450f-e45d-48d4-99a7-89907ae9da71\") " Mar 07 15:09:56 crc kubenswrapper[4943]: I0307 15:09:56.589090 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/de11450f-e45d-48d4-99a7-89907ae9da71-scripts\") pod \"de11450f-e45d-48d4-99a7-89907ae9da71\" (UID: \"de11450f-e45d-48d4-99a7-89907ae9da71\") " Mar 07 15:09:56 crc kubenswrapper[4943]: I0307 15:09:56.589444 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/de11450f-e45d-48d4-99a7-89907ae9da71-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:09:56 crc kubenswrapper[4943]: I0307 15:09:56.589986 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de11450f-e45d-48d4-99a7-89907ae9da71-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "de11450f-e45d-48d4-99a7-89907ae9da71" (UID: "de11450f-e45d-48d4-99a7-89907ae9da71"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:09:56 crc kubenswrapper[4943]: I0307 15:09:56.596224 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de11450f-e45d-48d4-99a7-89907ae9da71-kube-api-access-htmds" (OuterVolumeSpecName: "kube-api-access-htmds") pod "de11450f-e45d-48d4-99a7-89907ae9da71" (UID: "de11450f-e45d-48d4-99a7-89907ae9da71"). InnerVolumeSpecName "kube-api-access-htmds". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:09:56 crc kubenswrapper[4943]: I0307 15:09:56.612775 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de11450f-e45d-48d4-99a7-89907ae9da71-scripts" (OuterVolumeSpecName: "scripts") pod "de11450f-e45d-48d4-99a7-89907ae9da71" (UID: "de11450f-e45d-48d4-99a7-89907ae9da71"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:09:56 crc kubenswrapper[4943]: I0307 15:09:56.616403 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de11450f-e45d-48d4-99a7-89907ae9da71-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "de11450f-e45d-48d4-99a7-89907ae9da71" (UID: "de11450f-e45d-48d4-99a7-89907ae9da71"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:09:56 crc kubenswrapper[4943]: I0307 15:09:56.623827 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de11450f-e45d-48d4-99a7-89907ae9da71-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "de11450f-e45d-48d4-99a7-89907ae9da71" (UID: "de11450f-e45d-48d4-99a7-89907ae9da71"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:09:56 crc kubenswrapper[4943]: I0307 15:09:56.691275 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/de11450f-e45d-48d4-99a7-89907ae9da71-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:09:56 crc kubenswrapper[4943]: I0307 15:09:56.691313 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/de11450f-e45d-48d4-99a7-89907ae9da71-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:09:56 crc kubenswrapper[4943]: I0307 15:09:56.691332 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/de11450f-e45d-48d4-99a7-89907ae9da71-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:09:56 crc kubenswrapper[4943]: I0307 15:09:56.691349 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/de11450f-e45d-48d4-99a7-89907ae9da71-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:09:56 crc kubenswrapper[4943]: I0307 15:09:56.691364 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htmds\" (UniqueName: \"kubernetes.io/projected/de11450f-e45d-48d4-99a7-89907ae9da71-kube-api-access-htmds\") on node \"crc\" DevicePath \"\"" Mar 07 15:09:56 crc kubenswrapper[4943]: I0307 15:09:56.897213 4943 generic.go:334] "Generic (PLEG): container finished" podID="de11450f-e45d-48d4-99a7-89907ae9da71" containerID="9a27b9b16b6a885e47f4731dee29920caf08019a737681a20b88cbe0375277cc" exitCode=1 Mar 07 15:09:56 crc kubenswrapper[4943]: I0307 15:09:56.897273 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5lvgv" event={"ID":"de11450f-e45d-48d4-99a7-89907ae9da71","Type":"ContainerDied","Data":"9a27b9b16b6a885e47f4731dee29920caf08019a737681a20b88cbe0375277cc"} Mar 07 15:09:56 crc kubenswrapper[4943]: I0307 15:09:56.897284 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5lvgv" Mar 07 15:09:56 crc kubenswrapper[4943]: I0307 15:09:56.897321 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5lvgv" event={"ID":"de11450f-e45d-48d4-99a7-89907ae9da71","Type":"ContainerDied","Data":"4093b73774d310ae50f7bfe7a6ff77a8443c0b62241e17dc23fccd6755d59c20"} Mar 07 15:09:56 crc kubenswrapper[4943]: I0307 15:09:56.897350 4943 scope.go:117] "RemoveContainer" containerID="9a27b9b16b6a885e47f4731dee29920caf08019a737681a20b88cbe0375277cc" Mar 07 15:09:56 crc kubenswrapper[4943]: I0307 15:09:56.913808 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-5lvgv"] Mar 07 15:09:56 crc kubenswrapper[4943]: I0307 15:09:56.918612 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-5lvgv"] Mar 07 15:09:56 crc kubenswrapper[4943]: I0307 15:09:56.937293 4943 scope.go:117] "RemoveContainer" containerID="9a27b9b16b6a885e47f4731dee29920caf08019a737681a20b88cbe0375277cc" Mar 07 15:09:56 crc kubenswrapper[4943]: E0307 15:09:56.937850 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a27b9b16b6a885e47f4731dee29920caf08019a737681a20b88cbe0375277cc\": container with ID starting with 9a27b9b16b6a885e47f4731dee29920caf08019a737681a20b88cbe0375277cc not found: ID does not exist" containerID="9a27b9b16b6a885e47f4731dee29920caf08019a737681a20b88cbe0375277cc" Mar 07 15:09:56 crc kubenswrapper[4943]: I0307 15:09:56.937882 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a27b9b16b6a885e47f4731dee29920caf08019a737681a20b88cbe0375277cc"} err="failed to get container status \"9a27b9b16b6a885e47f4731dee29920caf08019a737681a20b88cbe0375277cc\": rpc error: code = NotFound desc = could not find container \"9a27b9b16b6a885e47f4731dee29920caf08019a737681a20b88cbe0375277cc\": container with ID starting with 9a27b9b16b6a885e47f4731dee29920caf08019a737681a20b88cbe0375277cc not found: ID does not exist" Mar 07 15:09:57 crc kubenswrapper[4943]: I0307 15:09:57.755525 4943 scope.go:117] "RemoveContainer" containerID="26f35e1f941cca0c34e3650d7aa047834e769dd76d2a61c3a39cad5bb2d090ae" Mar 07 15:09:57 crc kubenswrapper[4943]: E0307 15:09:57.756201 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cpdmm_openshift-machine-config-operator(4c1f18ea-e557-41fc-95dd-ce4e1722a61e)\"" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" Mar 07 15:09:58 crc kubenswrapper[4943]: I0307 15:09:58.770815 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de11450f-e45d-48d4-99a7-89907ae9da71" path="/var/lib/kubelet/pods/de11450f-e45d-48d4-99a7-89907ae9da71/volumes" Mar 07 15:09:58 crc kubenswrapper[4943]: I0307 15:09:58.911545 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-tc4dt"] Mar 07 15:09:58 crc kubenswrapper[4943]: E0307 15:09:58.912224 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de11450f-e45d-48d4-99a7-89907ae9da71" containerName="swift-ring-rebalance" Mar 07 15:09:58 crc kubenswrapper[4943]: I0307 15:09:58.912263 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="de11450f-e45d-48d4-99a7-89907ae9da71" containerName="swift-ring-rebalance" Mar 07 15:09:58 crc kubenswrapper[4943]: E0307 15:09:58.912301 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85e1e3e8-b9f8-4c3d-940e-4060262201ee" containerName="swift-ring-rebalance" Mar 07 15:09:58 crc kubenswrapper[4943]: I0307 15:09:58.912320 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="85e1e3e8-b9f8-4c3d-940e-4060262201ee" containerName="swift-ring-rebalance" Mar 07 15:09:58 crc kubenswrapper[4943]: E0307 15:09:58.912357 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d17441e-f6b6-4fab-8239-0409b7bb3910" containerName="proxy-server" Mar 07 15:09:58 crc kubenswrapper[4943]: I0307 15:09:58.912375 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d17441e-f6b6-4fab-8239-0409b7bb3910" containerName="proxy-server" Mar 07 15:09:58 crc kubenswrapper[4943]: E0307 15:09:58.912418 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d17441e-f6b6-4fab-8239-0409b7bb3910" containerName="proxy-httpd" Mar 07 15:09:58 crc kubenswrapper[4943]: I0307 15:09:58.912436 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d17441e-f6b6-4fab-8239-0409b7bb3910" containerName="proxy-httpd" Mar 07 15:09:58 crc kubenswrapper[4943]: I0307 15:09:58.912763 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d17441e-f6b6-4fab-8239-0409b7bb3910" containerName="proxy-httpd" Mar 07 15:09:58 crc kubenswrapper[4943]: I0307 15:09:58.912793 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d17441e-f6b6-4fab-8239-0409b7bb3910" containerName="proxy-server" Mar 07 15:09:58 crc kubenswrapper[4943]: I0307 15:09:58.912819 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="85e1e3e8-b9f8-4c3d-940e-4060262201ee" containerName="swift-ring-rebalance" Mar 07 15:09:58 crc kubenswrapper[4943]: I0307 15:09:58.912851 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="de11450f-e45d-48d4-99a7-89907ae9da71" containerName="swift-ring-rebalance" Mar 07 15:09:58 crc kubenswrapper[4943]: I0307 15:09:58.913671 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-tc4dt" Mar 07 15:09:58 crc kubenswrapper[4943]: I0307 15:09:58.917267 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:09:58 crc kubenswrapper[4943]: I0307 15:09:58.920208 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:09:58 crc kubenswrapper[4943]: I0307 15:09:58.929020 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-tc4dt"] Mar 07 15:09:59 crc kubenswrapper[4943]: I0307 15:09:59.026151 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/072c2e48-01cc-4912-8771-81c4e2713582-etc-swift\") pod \"swift-ring-rebalance-debug-tc4dt\" (UID: \"072c2e48-01cc-4912-8771-81c4e2713582\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tc4dt" Mar 07 15:09:59 crc kubenswrapper[4943]: I0307 15:09:59.026202 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/072c2e48-01cc-4912-8771-81c4e2713582-swiftconf\") pod \"swift-ring-rebalance-debug-tc4dt\" (UID: \"072c2e48-01cc-4912-8771-81c4e2713582\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tc4dt" Mar 07 15:09:59 crc kubenswrapper[4943]: I0307 15:09:59.026240 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kp5pf\" (UniqueName: \"kubernetes.io/projected/072c2e48-01cc-4912-8771-81c4e2713582-kube-api-access-kp5pf\") pod \"swift-ring-rebalance-debug-tc4dt\" (UID: \"072c2e48-01cc-4912-8771-81c4e2713582\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tc4dt" Mar 07 15:09:59 crc kubenswrapper[4943]: I0307 15:09:59.026282 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/072c2e48-01cc-4912-8771-81c4e2713582-scripts\") pod \"swift-ring-rebalance-debug-tc4dt\" (UID: \"072c2e48-01cc-4912-8771-81c4e2713582\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tc4dt" Mar 07 15:09:59 crc kubenswrapper[4943]: I0307 15:09:59.026305 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/072c2e48-01cc-4912-8771-81c4e2713582-ring-data-devices\") pod \"swift-ring-rebalance-debug-tc4dt\" (UID: \"072c2e48-01cc-4912-8771-81c4e2713582\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tc4dt" Mar 07 15:09:59 crc kubenswrapper[4943]: I0307 15:09:59.026380 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/072c2e48-01cc-4912-8771-81c4e2713582-dispersionconf\") pod \"swift-ring-rebalance-debug-tc4dt\" (UID: \"072c2e48-01cc-4912-8771-81c4e2713582\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tc4dt" Mar 07 15:09:59 crc kubenswrapper[4943]: I0307 15:09:59.129308 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/072c2e48-01cc-4912-8771-81c4e2713582-scripts\") pod \"swift-ring-rebalance-debug-tc4dt\" (UID: \"072c2e48-01cc-4912-8771-81c4e2713582\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tc4dt" Mar 07 15:09:59 crc kubenswrapper[4943]: I0307 15:09:59.129784 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/072c2e48-01cc-4912-8771-81c4e2713582-ring-data-devices\") pod \"swift-ring-rebalance-debug-tc4dt\" (UID: \"072c2e48-01cc-4912-8771-81c4e2713582\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tc4dt" Mar 07 15:09:59 crc kubenswrapper[4943]: I0307 15:09:59.129854 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/072c2e48-01cc-4912-8771-81c4e2713582-dispersionconf\") pod \"swift-ring-rebalance-debug-tc4dt\" (UID: \"072c2e48-01cc-4912-8771-81c4e2713582\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tc4dt" Mar 07 15:09:59 crc kubenswrapper[4943]: I0307 15:09:59.130007 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/072c2e48-01cc-4912-8771-81c4e2713582-etc-swift\") pod \"swift-ring-rebalance-debug-tc4dt\" (UID: \"072c2e48-01cc-4912-8771-81c4e2713582\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tc4dt" Mar 07 15:09:59 crc kubenswrapper[4943]: I0307 15:09:59.130052 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/072c2e48-01cc-4912-8771-81c4e2713582-swiftconf\") pod \"swift-ring-rebalance-debug-tc4dt\" (UID: \"072c2e48-01cc-4912-8771-81c4e2713582\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tc4dt" Mar 07 15:09:59 crc kubenswrapper[4943]: I0307 15:09:59.130832 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kp5pf\" (UniqueName: \"kubernetes.io/projected/072c2e48-01cc-4912-8771-81c4e2713582-kube-api-access-kp5pf\") pod \"swift-ring-rebalance-debug-tc4dt\" (UID: \"072c2e48-01cc-4912-8771-81c4e2713582\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tc4dt" Mar 07 15:09:59 crc kubenswrapper[4943]: I0307 15:09:59.131027 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/072c2e48-01cc-4912-8771-81c4e2713582-scripts\") pod \"swift-ring-rebalance-debug-tc4dt\" (UID: \"072c2e48-01cc-4912-8771-81c4e2713582\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tc4dt" Mar 07 15:09:59 crc kubenswrapper[4943]: I0307 15:09:59.130725 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/072c2e48-01cc-4912-8771-81c4e2713582-etc-swift\") pod \"swift-ring-rebalance-debug-tc4dt\" (UID: \"072c2e48-01cc-4912-8771-81c4e2713582\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tc4dt" Mar 07 15:09:59 crc kubenswrapper[4943]: I0307 15:09:59.131329 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/072c2e48-01cc-4912-8771-81c4e2713582-ring-data-devices\") pod \"swift-ring-rebalance-debug-tc4dt\" (UID: \"072c2e48-01cc-4912-8771-81c4e2713582\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tc4dt" Mar 07 15:09:59 crc kubenswrapper[4943]: I0307 15:09:59.141482 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/072c2e48-01cc-4912-8771-81c4e2713582-swiftconf\") pod \"swift-ring-rebalance-debug-tc4dt\" (UID: \"072c2e48-01cc-4912-8771-81c4e2713582\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tc4dt" Mar 07 15:09:59 crc kubenswrapper[4943]: I0307 15:09:59.143735 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/072c2e48-01cc-4912-8771-81c4e2713582-dispersionconf\") pod \"swift-ring-rebalance-debug-tc4dt\" (UID: \"072c2e48-01cc-4912-8771-81c4e2713582\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tc4dt" Mar 07 15:09:59 crc kubenswrapper[4943]: I0307 15:09:59.164048 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kp5pf\" (UniqueName: \"kubernetes.io/projected/072c2e48-01cc-4912-8771-81c4e2713582-kube-api-access-kp5pf\") pod \"swift-ring-rebalance-debug-tc4dt\" (UID: \"072c2e48-01cc-4912-8771-81c4e2713582\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tc4dt" Mar 07 15:09:59 crc kubenswrapper[4943]: I0307 15:09:59.247460 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-tc4dt" Mar 07 15:09:59 crc kubenswrapper[4943]: I0307 15:09:59.531552 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-tc4dt"] Mar 07 15:09:59 crc kubenswrapper[4943]: I0307 15:09:59.931168 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-tc4dt" event={"ID":"072c2e48-01cc-4912-8771-81c4e2713582","Type":"ContainerStarted","Data":"e6c8eda6fc15ae26bdb6d81e503888d7982d8fcf99f9bc1e7390f99ec44c938b"} Mar 07 15:10:00 crc kubenswrapper[4943]: I0307 15:10:00.158350 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29548270-v9j8n"] Mar 07 15:10:00 crc kubenswrapper[4943]: I0307 15:10:00.159585 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548270-v9j8n" Mar 07 15:10:00 crc kubenswrapper[4943]: I0307 15:10:00.162345 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 07 15:10:00 crc kubenswrapper[4943]: I0307 15:10:00.163697 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 07 15:10:00 crc kubenswrapper[4943]: I0307 15:10:00.164009 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-qqlk5" Mar 07 15:10:00 crc kubenswrapper[4943]: I0307 15:10:00.175447 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29548270-v9j8n"] Mar 07 15:10:00 crc kubenswrapper[4943]: I0307 15:10:00.251317 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hsdq2\" (UniqueName: \"kubernetes.io/projected/62e0e72a-7faf-4dbb-86d7-cbb756a90d63-kube-api-access-hsdq2\") pod \"auto-csr-approver-29548270-v9j8n\" (UID: \"62e0e72a-7faf-4dbb-86d7-cbb756a90d63\") " pod="openshift-infra/auto-csr-approver-29548270-v9j8n" Mar 07 15:10:00 crc kubenswrapper[4943]: I0307 15:10:00.353816 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hsdq2\" (UniqueName: \"kubernetes.io/projected/62e0e72a-7faf-4dbb-86d7-cbb756a90d63-kube-api-access-hsdq2\") pod \"auto-csr-approver-29548270-v9j8n\" (UID: \"62e0e72a-7faf-4dbb-86d7-cbb756a90d63\") " pod="openshift-infra/auto-csr-approver-29548270-v9j8n" Mar 07 15:10:00 crc kubenswrapper[4943]: I0307 15:10:00.385794 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hsdq2\" (UniqueName: \"kubernetes.io/projected/62e0e72a-7faf-4dbb-86d7-cbb756a90d63-kube-api-access-hsdq2\") pod \"auto-csr-approver-29548270-v9j8n\" (UID: \"62e0e72a-7faf-4dbb-86d7-cbb756a90d63\") " pod="openshift-infra/auto-csr-approver-29548270-v9j8n" Mar 07 15:10:00 crc kubenswrapper[4943]: I0307 15:10:00.497957 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548270-v9j8n" Mar 07 15:10:00 crc kubenswrapper[4943]: I0307 15:10:00.958614 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-tc4dt" event={"ID":"072c2e48-01cc-4912-8771-81c4e2713582","Type":"ContainerStarted","Data":"6b2b3d26dfb9fc84ca17101f4e32b166dab99cf4d078b30e536a2eb783f0b9cb"} Mar 07 15:10:00 crc kubenswrapper[4943]: I0307 15:10:00.984634 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29548270-v9j8n"] Mar 07 15:10:00 crc kubenswrapper[4943]: I0307 15:10:00.992085 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-tc4dt" podStartSLOduration=2.99205891 podStartE2EDuration="2.99205891s" podCreationTimestamp="2026-03-07 15:09:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:10:00.9888117 +0000 UTC m=+1842.940948238" watchObservedRunningTime="2026-03-07 15:10:00.99205891 +0000 UTC m=+1842.944195448" Mar 07 15:10:00 crc kubenswrapper[4943]: W0307 15:10:00.999216 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod62e0e72a_7faf_4dbb_86d7_cbb756a90d63.slice/crio-24c6020642e4f32f29ee961af9a9e44a3e3e91612a1117a90403200690f5dbb4 WatchSource:0}: Error finding container 24c6020642e4f32f29ee961af9a9e44a3e3e91612a1117a90403200690f5dbb4: Status 404 returned error can't find the container with id 24c6020642e4f32f29ee961af9a9e44a3e3e91612a1117a90403200690f5dbb4 Mar 07 15:10:01 crc kubenswrapper[4943]: I0307 15:10:01.971835 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548270-v9j8n" event={"ID":"62e0e72a-7faf-4dbb-86d7-cbb756a90d63","Type":"ContainerStarted","Data":"24c6020642e4f32f29ee961af9a9e44a3e3e91612a1117a90403200690f5dbb4"} Mar 07 15:10:02 crc kubenswrapper[4943]: I0307 15:10:02.983822 4943 generic.go:334] "Generic (PLEG): container finished" podID="62e0e72a-7faf-4dbb-86d7-cbb756a90d63" containerID="4532bdcc9922a6a649a935ca6b7bc782b1aa3e570e68991386637b2ea30daef0" exitCode=0 Mar 07 15:10:02 crc kubenswrapper[4943]: I0307 15:10:02.983907 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548270-v9j8n" event={"ID":"62e0e72a-7faf-4dbb-86d7-cbb756a90d63","Type":"ContainerDied","Data":"4532bdcc9922a6a649a935ca6b7bc782b1aa3e570e68991386637b2ea30daef0"} Mar 07 15:10:04 crc kubenswrapper[4943]: I0307 15:10:04.371758 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548270-v9j8n" Mar 07 15:10:04 crc kubenswrapper[4943]: I0307 15:10:04.422636 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hsdq2\" (UniqueName: \"kubernetes.io/projected/62e0e72a-7faf-4dbb-86d7-cbb756a90d63-kube-api-access-hsdq2\") pod \"62e0e72a-7faf-4dbb-86d7-cbb756a90d63\" (UID: \"62e0e72a-7faf-4dbb-86d7-cbb756a90d63\") " Mar 07 15:10:04 crc kubenswrapper[4943]: I0307 15:10:04.431728 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62e0e72a-7faf-4dbb-86d7-cbb756a90d63-kube-api-access-hsdq2" (OuterVolumeSpecName: "kube-api-access-hsdq2") pod "62e0e72a-7faf-4dbb-86d7-cbb756a90d63" (UID: "62e0e72a-7faf-4dbb-86d7-cbb756a90d63"). InnerVolumeSpecName "kube-api-access-hsdq2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:10:04 crc kubenswrapper[4943]: I0307 15:10:04.525066 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hsdq2\" (UniqueName: \"kubernetes.io/projected/62e0e72a-7faf-4dbb-86d7-cbb756a90d63-kube-api-access-hsdq2\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:05 crc kubenswrapper[4943]: I0307 15:10:05.003194 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548270-v9j8n" event={"ID":"62e0e72a-7faf-4dbb-86d7-cbb756a90d63","Type":"ContainerDied","Data":"24c6020642e4f32f29ee961af9a9e44a3e3e91612a1117a90403200690f5dbb4"} Mar 07 15:10:05 crc kubenswrapper[4943]: I0307 15:10:05.003242 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="24c6020642e4f32f29ee961af9a9e44a3e3e91612a1117a90403200690f5dbb4" Mar 07 15:10:05 crc kubenswrapper[4943]: I0307 15:10:05.003309 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548270-v9j8n" Mar 07 15:10:05 crc kubenswrapper[4943]: I0307 15:10:05.451408 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29548264-qrpfx"] Mar 07 15:10:05 crc kubenswrapper[4943]: I0307 15:10:05.457259 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29548264-qrpfx"] Mar 07 15:10:06 crc kubenswrapper[4943]: I0307 15:10:06.020958 4943 generic.go:334] "Generic (PLEG): container finished" podID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerID="06b728b91732e6aa41de478ccef9fb0046a86c5a1dc8baedd3243f9a3c0ed18c" exitCode=137 Mar 07 15:10:06 crc kubenswrapper[4943]: I0307 15:10:06.021034 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8","Type":"ContainerDied","Data":"06b728b91732e6aa41de478ccef9fb0046a86c5a1dc8baedd3243f9a3c0ed18c"} Mar 07 15:10:06 crc kubenswrapper[4943]: I0307 15:10:06.176701 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-storage-0" Mar 07 15:10:06 crc kubenswrapper[4943]: I0307 15:10:06.251472 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8-etc-swift\") pod \"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8\" (UID: \"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8\") " Mar 07 15:10:06 crc kubenswrapper[4943]: I0307 15:10:06.251651 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swift\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8\" (UID: \"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8\") " Mar 07 15:10:06 crc kubenswrapper[4943]: I0307 15:10:06.253567 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-scpxj\" (UniqueName: \"kubernetes.io/projected/3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8-kube-api-access-scpxj\") pod \"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8\" (UID: \"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8\") " Mar 07 15:10:06 crc kubenswrapper[4943]: I0307 15:10:06.253627 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8-cache\") pod \"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8\" (UID: \"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8\") " Mar 07 15:10:06 crc kubenswrapper[4943]: I0307 15:10:06.253755 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8-lock\") pod \"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8\" (UID: \"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8\") " Mar 07 15:10:06 crc kubenswrapper[4943]: I0307 15:10:06.254524 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8-cache" (OuterVolumeSpecName: "cache") pod "3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" (UID: "3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8"). InnerVolumeSpecName "cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:10:06 crc kubenswrapper[4943]: I0307 15:10:06.254651 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8-lock" (OuterVolumeSpecName: "lock") pod "3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" (UID: "3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8"). InnerVolumeSpecName "lock". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:10:06 crc kubenswrapper[4943]: I0307 15:10:06.255039 4943 reconciler_common.go:293] "Volume detached for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8-cache\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:06 crc kubenswrapper[4943]: I0307 15:10:06.255067 4943 reconciler_common.go:293] "Volume detached for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8-lock\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:06 crc kubenswrapper[4943]: I0307 15:10:06.272053 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "swift") pod "3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" (UID: "3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 07 15:10:06 crc kubenswrapper[4943]: I0307 15:10:06.272351 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8-kube-api-access-scpxj" (OuterVolumeSpecName: "kube-api-access-scpxj") pod "3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" (UID: "3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8"). InnerVolumeSpecName "kube-api-access-scpxj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:10:06 crc kubenswrapper[4943]: I0307 15:10:06.273206 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" (UID: "3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:10:06 crc kubenswrapper[4943]: I0307 15:10:06.356779 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:06 crc kubenswrapper[4943]: I0307 15:10:06.356827 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-scpxj\" (UniqueName: \"kubernetes.io/projected/3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8-kube-api-access-scpxj\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:06 crc kubenswrapper[4943]: I0307 15:10:06.356880 4943 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Mar 07 15:10:06 crc kubenswrapper[4943]: I0307 15:10:06.383856 4943 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Mar 07 15:10:06 crc kubenswrapper[4943]: I0307 15:10:06.458725 4943 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:06 crc kubenswrapper[4943]: I0307 15:10:06.771810 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0838f7e8-2cbe-4b70-a399-ff1200dc1693" path="/var/lib/kubelet/pods/0838f7e8-2cbe-4b70-a399-ff1200dc1693/volumes" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.045599 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8","Type":"ContainerDied","Data":"8ec1d04cd278b86217515269c7955bd492054280abca8f7dd7e6efa559069292"} Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.045654 4943 scope.go:117] "RemoveContainer" containerID="a1e3e8dc6afbf3013626e71f36dcde17e61c0ed8650ec0a48d8c2ddccb2aa585" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.045866 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-storage-0" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.084049 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-storage-0"] Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.095417 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-storage-0"] Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.100552 4943 scope.go:117] "RemoveContainer" containerID="06b728b91732e6aa41de478ccef9fb0046a86c5a1dc8baedd3243f9a3c0ed18c" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.128213 4943 scope.go:117] "RemoveContainer" containerID="b96feec40ee02a1bd0a7a1c8d67abf8300f9b59dbae2f6e54e6ae30df4a3971b" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.147809 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-storage-0"] Mar 07 15:10:07 crc kubenswrapper[4943]: E0307 15:10:07.148190 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="account-auditor" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.148211 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="account-auditor" Mar 07 15:10:07 crc kubenswrapper[4943]: E0307 15:10:07.148225 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="container-updater" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.148233 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="container-updater" Mar 07 15:10:07 crc kubenswrapper[4943]: E0307 15:10:07.148248 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="account-replicator" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.148257 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="account-replicator" Mar 07 15:10:07 crc kubenswrapper[4943]: E0307 15:10:07.148264 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="container-server" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.148272 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="container-server" Mar 07 15:10:07 crc kubenswrapper[4943]: E0307 15:10:07.148286 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="account-server" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.148295 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="account-server" Mar 07 15:10:07 crc kubenswrapper[4943]: E0307 15:10:07.148311 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="object-auditor" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.148321 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="object-auditor" Mar 07 15:10:07 crc kubenswrapper[4943]: E0307 15:10:07.148332 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="container-replicator" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.148340 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="container-replicator" Mar 07 15:10:07 crc kubenswrapper[4943]: E0307 15:10:07.148354 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62e0e72a-7faf-4dbb-86d7-cbb756a90d63" containerName="oc" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.148362 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="62e0e72a-7faf-4dbb-86d7-cbb756a90d63" containerName="oc" Mar 07 15:10:07 crc kubenswrapper[4943]: E0307 15:10:07.148377 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="object-updater" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.148385 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="object-updater" Mar 07 15:10:07 crc kubenswrapper[4943]: E0307 15:10:07.148399 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="container-sharder" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.148407 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="container-sharder" Mar 07 15:10:07 crc kubenswrapper[4943]: E0307 15:10:07.148425 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="object-server" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.148432 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="object-server" Mar 07 15:10:07 crc kubenswrapper[4943]: E0307 15:10:07.148448 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="account-reaper" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.148456 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="account-reaper" Mar 07 15:10:07 crc kubenswrapper[4943]: E0307 15:10:07.148466 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="swift-recon-cron" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.148474 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="swift-recon-cron" Mar 07 15:10:07 crc kubenswrapper[4943]: E0307 15:10:07.148488 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="rsync" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.148496 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="rsync" Mar 07 15:10:07 crc kubenswrapper[4943]: E0307 15:10:07.148503 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="object-expirer" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.148511 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="object-expirer" Mar 07 15:10:07 crc kubenswrapper[4943]: E0307 15:10:07.148528 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="object-replicator" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.148536 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="object-replicator" Mar 07 15:10:07 crc kubenswrapper[4943]: E0307 15:10:07.148550 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="container-auditor" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.148560 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="container-auditor" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.148722 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="object-server" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.148739 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="container-replicator" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.148747 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="container-server" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.148759 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="container-auditor" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.148772 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="swift-recon-cron" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.148783 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="account-auditor" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.148792 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="container-updater" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.148803 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="object-expirer" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.148813 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="object-replicator" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.148823 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="object-updater" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.148834 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="account-replicator" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.148848 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="container-sharder" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.148857 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="object-auditor" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.148875 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="62e0e72a-7faf-4dbb-86d7-cbb756a90d63" containerName="oc" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.148884 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="account-server" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.148896 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="account-reaper" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.148905 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" containerName="rsync" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.153847 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-storage-0" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.155023 4943 scope.go:117] "RemoveContainer" containerID="f290059bb335b82ed354d4d229fb8530d57e7f29a151602dd18122d411960328" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.170139 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/93b86e6f-5b8a-45fe-ac00-62b667d92a76-etc-swift\") pod \"swift-storage-0\" (UID: \"93b86e6f-5b8a-45fe-ac00-62b667d92a76\") " pod="swift-kuttl-tests/swift-storage-0" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.170400 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/93b86e6f-5b8a-45fe-ac00-62b667d92a76-cache\") pod \"swift-storage-0\" (UID: \"93b86e6f-5b8a-45fe-ac00-62b667d92a76\") " pod="swift-kuttl-tests/swift-storage-0" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.170527 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"93b86e6f-5b8a-45fe-ac00-62b667d92a76\") " pod="swift-kuttl-tests/swift-storage-0" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.170625 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6jxw\" (UniqueName: \"kubernetes.io/projected/93b86e6f-5b8a-45fe-ac00-62b667d92a76-kube-api-access-b6jxw\") pod \"swift-storage-0\" (UID: \"93b86e6f-5b8a-45fe-ac00-62b667d92a76\") " pod="swift-kuttl-tests/swift-storage-0" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.170720 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/93b86e6f-5b8a-45fe-ac00-62b667d92a76-lock\") pod \"swift-storage-0\" (UID: \"93b86e6f-5b8a-45fe-ac00-62b667d92a76\") " pod="swift-kuttl-tests/swift-storage-0" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.184624 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-storage-0"] Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.187015 4943 scope.go:117] "RemoveContainer" containerID="a3b3ff008ced9b38eb096a8fe7d155531763f2bdf2e5399ff2af469f95e76448" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.214186 4943 scope.go:117] "RemoveContainer" containerID="3d606e2c04c5358f4c4404f1af78027ac23f0b1d763a512e4323fdd0240efd19" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.249215 4943 scope.go:117] "RemoveContainer" containerID="a75dfedf2adeb9783f42686241f798122e938a4edd4336b201d2e124ee87e1da" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.270713 4943 scope.go:117] "RemoveContainer" containerID="4009d66a16af4cf49d7b909280c182c6a2839bf7dc72660a38b3c02320550d9f" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.271912 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/93b86e6f-5b8a-45fe-ac00-62b667d92a76-etc-swift\") pod \"swift-storage-0\" (UID: \"93b86e6f-5b8a-45fe-ac00-62b667d92a76\") " pod="swift-kuttl-tests/swift-storage-0" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.271998 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/93b86e6f-5b8a-45fe-ac00-62b667d92a76-cache\") pod \"swift-storage-0\" (UID: \"93b86e6f-5b8a-45fe-ac00-62b667d92a76\") " pod="swift-kuttl-tests/swift-storage-0" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.272046 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"93b86e6f-5b8a-45fe-ac00-62b667d92a76\") " pod="swift-kuttl-tests/swift-storage-0" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.272090 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6jxw\" (UniqueName: \"kubernetes.io/projected/93b86e6f-5b8a-45fe-ac00-62b667d92a76-kube-api-access-b6jxw\") pod \"swift-storage-0\" (UID: \"93b86e6f-5b8a-45fe-ac00-62b667d92a76\") " pod="swift-kuttl-tests/swift-storage-0" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.272127 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/93b86e6f-5b8a-45fe-ac00-62b667d92a76-lock\") pod \"swift-storage-0\" (UID: \"93b86e6f-5b8a-45fe-ac00-62b667d92a76\") " pod="swift-kuttl-tests/swift-storage-0" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.272387 4943 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"93b86e6f-5b8a-45fe-ac00-62b667d92a76\") device mount path \"/mnt/openstack/pv07\"" pod="swift-kuttl-tests/swift-storage-0" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.272552 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/93b86e6f-5b8a-45fe-ac00-62b667d92a76-cache\") pod \"swift-storage-0\" (UID: \"93b86e6f-5b8a-45fe-ac00-62b667d92a76\") " pod="swift-kuttl-tests/swift-storage-0" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.272586 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/93b86e6f-5b8a-45fe-ac00-62b667d92a76-lock\") pod \"swift-storage-0\" (UID: \"93b86e6f-5b8a-45fe-ac00-62b667d92a76\") " pod="swift-kuttl-tests/swift-storage-0" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.293211 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/93b86e6f-5b8a-45fe-ac00-62b667d92a76-etc-swift\") pod \"swift-storage-0\" (UID: \"93b86e6f-5b8a-45fe-ac00-62b667d92a76\") " pod="swift-kuttl-tests/swift-storage-0" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.295638 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6jxw\" (UniqueName: \"kubernetes.io/projected/93b86e6f-5b8a-45fe-ac00-62b667d92a76-kube-api-access-b6jxw\") pod \"swift-storage-0\" (UID: \"93b86e6f-5b8a-45fe-ac00-62b667d92a76\") " pod="swift-kuttl-tests/swift-storage-0" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.296182 4943 scope.go:117] "RemoveContainer" containerID="40df134ae24e3ed4a8fd377fb725953a304574200ac3148e3f9b3b3806e44a47" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.298890 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"93b86e6f-5b8a-45fe-ac00-62b667d92a76\") " pod="swift-kuttl-tests/swift-storage-0" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.367081 4943 scope.go:117] "RemoveContainer" containerID="cc374a57378de01b25998555023a72a91a81bfe6f48e0c2d74901d09dc497799" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.409177 4943 scope.go:117] "RemoveContainer" containerID="a269396243ea039c5a1ff714244738a9af9f115de9be8ca31a6f67a9a2bb2e94" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.429274 4943 scope.go:117] "RemoveContainer" containerID="059e70bef0f14c03d2e83f29ffb3a0153b9683edc84e88eed7fa99bc0a0d8579" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.450672 4943 scope.go:117] "RemoveContainer" containerID="f4a58ec1a6b1a5a66faa2ca708935002ee19db5bef2c8c99a431c0bb5ec9eea3" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.466695 4943 scope.go:117] "RemoveContainer" containerID="5a68e14eb4c97f4f249f077f9c0d016c858544f0da60e1d6c84ee0b53bc3a259" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.479251 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-storage-0" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.483725 4943 scope.go:117] "RemoveContainer" containerID="c16bdb03bef6cc9ab7fb55a6953d241e9190e80787f4e3de43c893cd1ed349a2" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.506665 4943 scope.go:117] "RemoveContainer" containerID="9b6765692b58c5ebd2755fcbcc0bf6fee3ddbe781cfa1b626ad6078346d3e402" Mar 07 15:10:07 crc kubenswrapper[4943]: I0307 15:10:07.753617 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-storage-0"] Mar 07 15:10:08 crc kubenswrapper[4943]: I0307 15:10:08.053794 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"93b86e6f-5b8a-45fe-ac00-62b667d92a76","Type":"ContainerStarted","Data":"9ba007989fd42c622f16b5e4ace25bd26a482ad973220147c68b84f6138f3150"} Mar 07 15:10:08 crc kubenswrapper[4943]: I0307 15:10:08.054056 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"93b86e6f-5b8a-45fe-ac00-62b667d92a76","Type":"ContainerStarted","Data":"88ba5c72495f51a12cb384beff9c58f7e54914280dc558c2dad143c3b571b886"} Mar 07 15:10:08 crc kubenswrapper[4943]: I0307 15:10:08.765772 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8" path="/var/lib/kubelet/pods/3bcc7ca5-125f-46fa-9fbc-af0aedf0b4d8/volumes" Mar 07 15:10:09 crc kubenswrapper[4943]: I0307 15:10:09.086122 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"93b86e6f-5b8a-45fe-ac00-62b667d92a76","Type":"ContainerStarted","Data":"6102bf4401edc4ae72620fe1eb0e9502f04709ad9e62432f708fffba2894e54c"} Mar 07 15:10:09 crc kubenswrapper[4943]: I0307 15:10:09.086173 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"93b86e6f-5b8a-45fe-ac00-62b667d92a76","Type":"ContainerStarted","Data":"1a65eb594370437ee928435f464519e6fec795959f551e65a7bd19de60174781"} Mar 07 15:10:09 crc kubenswrapper[4943]: I0307 15:10:09.086188 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"93b86e6f-5b8a-45fe-ac00-62b667d92a76","Type":"ContainerStarted","Data":"b3dc134a1e971fdc89931785b5c42a63094257a4090b86a928fdc3fbbada6e50"} Mar 07 15:10:09 crc kubenswrapper[4943]: I0307 15:10:09.086200 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"93b86e6f-5b8a-45fe-ac00-62b667d92a76","Type":"ContainerStarted","Data":"4e42e1be4c23f0a38cea6d25ab4a3b2b9e9f6a2432fac4953b3c2c7bdd917621"} Mar 07 15:10:09 crc kubenswrapper[4943]: I0307 15:10:09.086211 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"93b86e6f-5b8a-45fe-ac00-62b667d92a76","Type":"ContainerStarted","Data":"556228e19c38348d428a3bba93ef89abec336bce942c0a05165050c54576dc97"} Mar 07 15:10:09 crc kubenswrapper[4943]: I0307 15:10:09.086221 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"93b86e6f-5b8a-45fe-ac00-62b667d92a76","Type":"ContainerStarted","Data":"2b906d17a88e011c936fbe9392806e44114b8d7536172f976afd16cf280a6aee"} Mar 07 15:10:09 crc kubenswrapper[4943]: I0307 15:10:09.086233 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"93b86e6f-5b8a-45fe-ac00-62b667d92a76","Type":"ContainerStarted","Data":"68fed155eb2402c7e01408cc00e17665715b74a236110e6fcebfaf827b016064"} Mar 07 15:10:10 crc kubenswrapper[4943]: I0307 15:10:10.100397 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"93b86e6f-5b8a-45fe-ac00-62b667d92a76","Type":"ContainerStarted","Data":"fbfa0e2ceb5c3913af746567a20fa18924e268408aa1e136f6afb07a1401ea12"} Mar 07 15:10:10 crc kubenswrapper[4943]: I0307 15:10:10.100448 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"93b86e6f-5b8a-45fe-ac00-62b667d92a76","Type":"ContainerStarted","Data":"8ed0ee5e433b696166a7f9c0c84faad973e0e400a4fcb0f9f40399116613a01b"} Mar 07 15:10:10 crc kubenswrapper[4943]: I0307 15:10:10.100462 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"93b86e6f-5b8a-45fe-ac00-62b667d92a76","Type":"ContainerStarted","Data":"7ee6a11f8c3d7cb343a7a8bab0a68d3c19c5d6e1290b679d16030c70972f95d7"} Mar 07 15:10:10 crc kubenswrapper[4943]: I0307 15:10:10.100473 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"93b86e6f-5b8a-45fe-ac00-62b667d92a76","Type":"ContainerStarted","Data":"16b8247841b80ad1133a201dfb7624599f5e108609d2d8e3a20b6aadbc176734"} Mar 07 15:10:10 crc kubenswrapper[4943]: I0307 15:10:10.100485 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"93b86e6f-5b8a-45fe-ac00-62b667d92a76","Type":"ContainerStarted","Data":"cb8922a42d3b9a18ce039306aa4661d2a8a940d6c5c8388cab28bc2c2917eb3e"} Mar 07 15:10:10 crc kubenswrapper[4943]: I0307 15:10:10.100498 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"93b86e6f-5b8a-45fe-ac00-62b667d92a76","Type":"ContainerStarted","Data":"a250c8924b7ef3b246aae7a1ad9492f923c065afd60ab0381bfb7ebd0bcbe26f"} Mar 07 15:10:11 crc kubenswrapper[4943]: I0307 15:10:11.120010 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"93b86e6f-5b8a-45fe-ac00-62b667d92a76","Type":"ContainerStarted","Data":"a49e000c5c089372422f621f31103f9271e15d2682bff33d9b0063f0e0beb75b"} Mar 07 15:10:11 crc kubenswrapper[4943]: I0307 15:10:11.176372 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-storage-0" podStartSLOduration=4.176345977 podStartE2EDuration="4.176345977s" podCreationTimestamp="2026-03-07 15:10:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:10:11.168091965 +0000 UTC m=+1853.120228513" watchObservedRunningTime="2026-03-07 15:10:11.176345977 +0000 UTC m=+1853.128482515" Mar 07 15:10:12 crc kubenswrapper[4943]: I0307 15:10:12.758068 4943 scope.go:117] "RemoveContainer" containerID="26f35e1f941cca0c34e3650d7aa047834e769dd76d2a61c3a39cad5bb2d090ae" Mar 07 15:10:13 crc kubenswrapper[4943]: I0307 15:10:13.138588 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" event={"ID":"4c1f18ea-e557-41fc-95dd-ce4e1722a61e","Type":"ContainerStarted","Data":"796971a8f0687fda2fdc2038d9ae78273781f30a32f7c084798a95ee29d28df7"} Mar 07 15:10:15 crc kubenswrapper[4943]: I0307 15:10:15.159340 4943 generic.go:334] "Generic (PLEG): container finished" podID="072c2e48-01cc-4912-8771-81c4e2713582" containerID="6b2b3d26dfb9fc84ca17101f4e32b166dab99cf4d078b30e536a2eb783f0b9cb" exitCode=0 Mar 07 15:10:15 crc kubenswrapper[4943]: I0307 15:10:15.159621 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-tc4dt" event={"ID":"072c2e48-01cc-4912-8771-81c4e2713582","Type":"ContainerDied","Data":"6b2b3d26dfb9fc84ca17101f4e32b166dab99cf4d078b30e536a2eb783f0b9cb"} Mar 07 15:10:16 crc kubenswrapper[4943]: I0307 15:10:16.489358 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-tc4dt" Mar 07 15:10:16 crc kubenswrapper[4943]: I0307 15:10:16.534647 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-tc4dt"] Mar 07 15:10:16 crc kubenswrapper[4943]: I0307 15:10:16.541125 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-tc4dt"] Mar 07 15:10:16 crc kubenswrapper[4943]: I0307 15:10:16.619432 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/072c2e48-01cc-4912-8771-81c4e2713582-ring-data-devices\") pod \"072c2e48-01cc-4912-8771-81c4e2713582\" (UID: \"072c2e48-01cc-4912-8771-81c4e2713582\") " Mar 07 15:10:16 crc kubenswrapper[4943]: I0307 15:10:16.619535 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/072c2e48-01cc-4912-8771-81c4e2713582-swiftconf\") pod \"072c2e48-01cc-4912-8771-81c4e2713582\" (UID: \"072c2e48-01cc-4912-8771-81c4e2713582\") " Mar 07 15:10:16 crc kubenswrapper[4943]: I0307 15:10:16.619585 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kp5pf\" (UniqueName: \"kubernetes.io/projected/072c2e48-01cc-4912-8771-81c4e2713582-kube-api-access-kp5pf\") pod \"072c2e48-01cc-4912-8771-81c4e2713582\" (UID: \"072c2e48-01cc-4912-8771-81c4e2713582\") " Mar 07 15:10:16 crc kubenswrapper[4943]: I0307 15:10:16.619675 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/072c2e48-01cc-4912-8771-81c4e2713582-scripts\") pod \"072c2e48-01cc-4912-8771-81c4e2713582\" (UID: \"072c2e48-01cc-4912-8771-81c4e2713582\") " Mar 07 15:10:16 crc kubenswrapper[4943]: I0307 15:10:16.619723 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/072c2e48-01cc-4912-8771-81c4e2713582-etc-swift\") pod \"072c2e48-01cc-4912-8771-81c4e2713582\" (UID: \"072c2e48-01cc-4912-8771-81c4e2713582\") " Mar 07 15:10:16 crc kubenswrapper[4943]: I0307 15:10:16.619786 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/072c2e48-01cc-4912-8771-81c4e2713582-dispersionconf\") pod \"072c2e48-01cc-4912-8771-81c4e2713582\" (UID: \"072c2e48-01cc-4912-8771-81c4e2713582\") " Mar 07 15:10:16 crc kubenswrapper[4943]: I0307 15:10:16.620565 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/072c2e48-01cc-4912-8771-81c4e2713582-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "072c2e48-01cc-4912-8771-81c4e2713582" (UID: "072c2e48-01cc-4912-8771-81c4e2713582"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:10:16 crc kubenswrapper[4943]: I0307 15:10:16.620893 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/072c2e48-01cc-4912-8771-81c4e2713582-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "072c2e48-01cc-4912-8771-81c4e2713582" (UID: "072c2e48-01cc-4912-8771-81c4e2713582"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:10:16 crc kubenswrapper[4943]: I0307 15:10:16.627181 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/072c2e48-01cc-4912-8771-81c4e2713582-kube-api-access-kp5pf" (OuterVolumeSpecName: "kube-api-access-kp5pf") pod "072c2e48-01cc-4912-8771-81c4e2713582" (UID: "072c2e48-01cc-4912-8771-81c4e2713582"). InnerVolumeSpecName "kube-api-access-kp5pf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:10:16 crc kubenswrapper[4943]: I0307 15:10:16.641440 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/072c2e48-01cc-4912-8771-81c4e2713582-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "072c2e48-01cc-4912-8771-81c4e2713582" (UID: "072c2e48-01cc-4912-8771-81c4e2713582"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:10:16 crc kubenswrapper[4943]: I0307 15:10:16.642882 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/072c2e48-01cc-4912-8771-81c4e2713582-scripts" (OuterVolumeSpecName: "scripts") pod "072c2e48-01cc-4912-8771-81c4e2713582" (UID: "072c2e48-01cc-4912-8771-81c4e2713582"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:10:16 crc kubenswrapper[4943]: I0307 15:10:16.652428 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/072c2e48-01cc-4912-8771-81c4e2713582-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "072c2e48-01cc-4912-8771-81c4e2713582" (UID: "072c2e48-01cc-4912-8771-81c4e2713582"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:10:16 crc kubenswrapper[4943]: I0307 15:10:16.722091 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/072c2e48-01cc-4912-8771-81c4e2713582-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:16 crc kubenswrapper[4943]: I0307 15:10:16.722167 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/072c2e48-01cc-4912-8771-81c4e2713582-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:16 crc kubenswrapper[4943]: I0307 15:10:16.722187 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kp5pf\" (UniqueName: \"kubernetes.io/projected/072c2e48-01cc-4912-8771-81c4e2713582-kube-api-access-kp5pf\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:16 crc kubenswrapper[4943]: I0307 15:10:16.722204 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/072c2e48-01cc-4912-8771-81c4e2713582-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:16 crc kubenswrapper[4943]: I0307 15:10:16.722221 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/072c2e48-01cc-4912-8771-81c4e2713582-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:16 crc kubenswrapper[4943]: I0307 15:10:16.722236 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/072c2e48-01cc-4912-8771-81c4e2713582-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:16 crc kubenswrapper[4943]: I0307 15:10:16.775437 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="072c2e48-01cc-4912-8771-81c4e2713582" path="/var/lib/kubelet/pods/072c2e48-01cc-4912-8771-81c4e2713582/volumes" Mar 07 15:10:17 crc kubenswrapper[4943]: I0307 15:10:17.039169 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-j4ng8"] Mar 07 15:10:17 crc kubenswrapper[4943]: E0307 15:10:17.039510 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="072c2e48-01cc-4912-8771-81c4e2713582" containerName="swift-ring-rebalance" Mar 07 15:10:17 crc kubenswrapper[4943]: I0307 15:10:17.039529 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="072c2e48-01cc-4912-8771-81c4e2713582" containerName="swift-ring-rebalance" Mar 07 15:10:17 crc kubenswrapper[4943]: I0307 15:10:17.039725 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="072c2e48-01cc-4912-8771-81c4e2713582" containerName="swift-ring-rebalance" Mar 07 15:10:17 crc kubenswrapper[4943]: I0307 15:10:17.045114 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4ng8" Mar 07 15:10:17 crc kubenswrapper[4943]: I0307 15:10:17.061522 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-j4ng8"] Mar 07 15:10:17 crc kubenswrapper[4943]: I0307 15:10:17.174507 4943 scope.go:117] "RemoveContainer" containerID="6b2b3d26dfb9fc84ca17101f4e32b166dab99cf4d078b30e536a2eb783f0b9cb" Mar 07 15:10:17 crc kubenswrapper[4943]: I0307 15:10:17.174575 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-tc4dt" Mar 07 15:10:17 crc kubenswrapper[4943]: I0307 15:10:17.231197 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1a558107-3b00-40b8-9043-45dc89e51dc1-etc-swift\") pod \"swift-ring-rebalance-debug-j4ng8\" (UID: \"1a558107-3b00-40b8-9043-45dc89e51dc1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4ng8" Mar 07 15:10:17 crc kubenswrapper[4943]: I0307 15:10:17.231375 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1a558107-3b00-40b8-9043-45dc89e51dc1-scripts\") pod \"swift-ring-rebalance-debug-j4ng8\" (UID: \"1a558107-3b00-40b8-9043-45dc89e51dc1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4ng8" Mar 07 15:10:17 crc kubenswrapper[4943]: I0307 15:10:17.231562 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1a558107-3b00-40b8-9043-45dc89e51dc1-swiftconf\") pod \"swift-ring-rebalance-debug-j4ng8\" (UID: \"1a558107-3b00-40b8-9043-45dc89e51dc1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4ng8" Mar 07 15:10:17 crc kubenswrapper[4943]: I0307 15:10:17.231647 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkrlp\" (UniqueName: \"kubernetes.io/projected/1a558107-3b00-40b8-9043-45dc89e51dc1-kube-api-access-bkrlp\") pod \"swift-ring-rebalance-debug-j4ng8\" (UID: \"1a558107-3b00-40b8-9043-45dc89e51dc1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4ng8" Mar 07 15:10:17 crc kubenswrapper[4943]: I0307 15:10:17.231723 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1a558107-3b00-40b8-9043-45dc89e51dc1-dispersionconf\") pod \"swift-ring-rebalance-debug-j4ng8\" (UID: \"1a558107-3b00-40b8-9043-45dc89e51dc1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4ng8" Mar 07 15:10:17 crc kubenswrapper[4943]: I0307 15:10:17.231960 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1a558107-3b00-40b8-9043-45dc89e51dc1-ring-data-devices\") pod \"swift-ring-rebalance-debug-j4ng8\" (UID: \"1a558107-3b00-40b8-9043-45dc89e51dc1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4ng8" Mar 07 15:10:17 crc kubenswrapper[4943]: I0307 15:10:17.333350 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1a558107-3b00-40b8-9043-45dc89e51dc1-ring-data-devices\") pod \"swift-ring-rebalance-debug-j4ng8\" (UID: \"1a558107-3b00-40b8-9043-45dc89e51dc1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4ng8" Mar 07 15:10:17 crc kubenswrapper[4943]: I0307 15:10:17.333780 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1a558107-3b00-40b8-9043-45dc89e51dc1-etc-swift\") pod \"swift-ring-rebalance-debug-j4ng8\" (UID: \"1a558107-3b00-40b8-9043-45dc89e51dc1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4ng8" Mar 07 15:10:17 crc kubenswrapper[4943]: I0307 15:10:17.333848 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1a558107-3b00-40b8-9043-45dc89e51dc1-scripts\") pod \"swift-ring-rebalance-debug-j4ng8\" (UID: \"1a558107-3b00-40b8-9043-45dc89e51dc1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4ng8" Mar 07 15:10:17 crc kubenswrapper[4943]: I0307 15:10:17.333903 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1a558107-3b00-40b8-9043-45dc89e51dc1-swiftconf\") pod \"swift-ring-rebalance-debug-j4ng8\" (UID: \"1a558107-3b00-40b8-9043-45dc89e51dc1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4ng8" Mar 07 15:10:17 crc kubenswrapper[4943]: I0307 15:10:17.333979 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkrlp\" (UniqueName: \"kubernetes.io/projected/1a558107-3b00-40b8-9043-45dc89e51dc1-kube-api-access-bkrlp\") pod \"swift-ring-rebalance-debug-j4ng8\" (UID: \"1a558107-3b00-40b8-9043-45dc89e51dc1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4ng8" Mar 07 15:10:17 crc kubenswrapper[4943]: I0307 15:10:17.334028 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1a558107-3b00-40b8-9043-45dc89e51dc1-dispersionconf\") pod \"swift-ring-rebalance-debug-j4ng8\" (UID: \"1a558107-3b00-40b8-9043-45dc89e51dc1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4ng8" Mar 07 15:10:17 crc kubenswrapper[4943]: I0307 15:10:17.334025 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1a558107-3b00-40b8-9043-45dc89e51dc1-ring-data-devices\") pod \"swift-ring-rebalance-debug-j4ng8\" (UID: \"1a558107-3b00-40b8-9043-45dc89e51dc1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4ng8" Mar 07 15:10:17 crc kubenswrapper[4943]: I0307 15:10:17.334315 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1a558107-3b00-40b8-9043-45dc89e51dc1-etc-swift\") pod \"swift-ring-rebalance-debug-j4ng8\" (UID: \"1a558107-3b00-40b8-9043-45dc89e51dc1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4ng8" Mar 07 15:10:17 crc kubenswrapper[4943]: I0307 15:10:17.336962 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1a558107-3b00-40b8-9043-45dc89e51dc1-scripts\") pod \"swift-ring-rebalance-debug-j4ng8\" (UID: \"1a558107-3b00-40b8-9043-45dc89e51dc1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4ng8" Mar 07 15:10:17 crc kubenswrapper[4943]: I0307 15:10:17.339957 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1a558107-3b00-40b8-9043-45dc89e51dc1-swiftconf\") pod \"swift-ring-rebalance-debug-j4ng8\" (UID: \"1a558107-3b00-40b8-9043-45dc89e51dc1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4ng8" Mar 07 15:10:17 crc kubenswrapper[4943]: I0307 15:10:17.341019 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1a558107-3b00-40b8-9043-45dc89e51dc1-dispersionconf\") pod \"swift-ring-rebalance-debug-j4ng8\" (UID: \"1a558107-3b00-40b8-9043-45dc89e51dc1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4ng8" Mar 07 15:10:17 crc kubenswrapper[4943]: I0307 15:10:17.360586 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkrlp\" (UniqueName: \"kubernetes.io/projected/1a558107-3b00-40b8-9043-45dc89e51dc1-kube-api-access-bkrlp\") pod \"swift-ring-rebalance-debug-j4ng8\" (UID: \"1a558107-3b00-40b8-9043-45dc89e51dc1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4ng8" Mar 07 15:10:17 crc kubenswrapper[4943]: I0307 15:10:17.418341 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4ng8" Mar 07 15:10:17 crc kubenswrapper[4943]: I0307 15:10:17.663792 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-j4ng8"] Mar 07 15:10:17 crc kubenswrapper[4943]: W0307 15:10:17.669467 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1a558107_3b00_40b8_9043_45dc89e51dc1.slice/crio-96271b1df8672db5151a7bbe3d5217a7a1b46b10bb1279f028a63fbe6acf67a8 WatchSource:0}: Error finding container 96271b1df8672db5151a7bbe3d5217a7a1b46b10bb1279f028a63fbe6acf67a8: Status 404 returned error can't find the container with id 96271b1df8672db5151a7bbe3d5217a7a1b46b10bb1279f028a63fbe6acf67a8 Mar 07 15:10:18 crc kubenswrapper[4943]: I0307 15:10:18.188113 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4ng8" event={"ID":"1a558107-3b00-40b8-9043-45dc89e51dc1","Type":"ContainerStarted","Data":"be5eb23beadf94e639477c15837fe80c8e5410f72b6bfc4038ab40de44979f6d"} Mar 07 15:10:18 crc kubenswrapper[4943]: I0307 15:10:18.188382 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4ng8" event={"ID":"1a558107-3b00-40b8-9043-45dc89e51dc1","Type":"ContainerStarted","Data":"96271b1df8672db5151a7bbe3d5217a7a1b46b10bb1279f028a63fbe6acf67a8"} Mar 07 15:10:18 crc kubenswrapper[4943]: I0307 15:10:18.215600 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4ng8" podStartSLOduration=1.215568177 podStartE2EDuration="1.215568177s" podCreationTimestamp="2026-03-07 15:10:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:10:18.207358516 +0000 UTC m=+1860.159495044" watchObservedRunningTime="2026-03-07 15:10:18.215568177 +0000 UTC m=+1860.167704705" Mar 07 15:10:20 crc kubenswrapper[4943]: I0307 15:10:20.212109 4943 generic.go:334] "Generic (PLEG): container finished" podID="1a558107-3b00-40b8-9043-45dc89e51dc1" containerID="be5eb23beadf94e639477c15837fe80c8e5410f72b6bfc4038ab40de44979f6d" exitCode=0 Mar 07 15:10:20 crc kubenswrapper[4943]: I0307 15:10:20.212221 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4ng8" event={"ID":"1a558107-3b00-40b8-9043-45dc89e51dc1","Type":"ContainerDied","Data":"be5eb23beadf94e639477c15837fe80c8e5410f72b6bfc4038ab40de44979f6d"} Mar 07 15:10:21 crc kubenswrapper[4943]: I0307 15:10:21.096465 4943 scope.go:117] "RemoveContainer" containerID="0894765bb00937a71e749c9f473984ddcb66ad8efbd3f954b5ea0513b71a5a6a" Mar 07 15:10:21 crc kubenswrapper[4943]: I0307 15:10:21.131023 4943 scope.go:117] "RemoveContainer" containerID="4ed71004898eb8cd3a17623fb532ec001eedf0b02e3f35dee7ff33da101e22db" Mar 07 15:10:21 crc kubenswrapper[4943]: I0307 15:10:21.518341 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4ng8" Mar 07 15:10:21 crc kubenswrapper[4943]: I0307 15:10:21.558890 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-j4ng8"] Mar 07 15:10:21 crc kubenswrapper[4943]: I0307 15:10:21.568664 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-j4ng8"] Mar 07 15:10:21 crc kubenswrapper[4943]: I0307 15:10:21.703688 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1a558107-3b00-40b8-9043-45dc89e51dc1-swiftconf\") pod \"1a558107-3b00-40b8-9043-45dc89e51dc1\" (UID: \"1a558107-3b00-40b8-9043-45dc89e51dc1\") " Mar 07 15:10:21 crc kubenswrapper[4943]: I0307 15:10:21.703796 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1a558107-3b00-40b8-9043-45dc89e51dc1-etc-swift\") pod \"1a558107-3b00-40b8-9043-45dc89e51dc1\" (UID: \"1a558107-3b00-40b8-9043-45dc89e51dc1\") " Mar 07 15:10:21 crc kubenswrapper[4943]: I0307 15:10:21.703852 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bkrlp\" (UniqueName: \"kubernetes.io/projected/1a558107-3b00-40b8-9043-45dc89e51dc1-kube-api-access-bkrlp\") pod \"1a558107-3b00-40b8-9043-45dc89e51dc1\" (UID: \"1a558107-3b00-40b8-9043-45dc89e51dc1\") " Mar 07 15:10:21 crc kubenswrapper[4943]: I0307 15:10:21.703904 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1a558107-3b00-40b8-9043-45dc89e51dc1-ring-data-devices\") pod \"1a558107-3b00-40b8-9043-45dc89e51dc1\" (UID: \"1a558107-3b00-40b8-9043-45dc89e51dc1\") " Mar 07 15:10:21 crc kubenswrapper[4943]: I0307 15:10:21.703986 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1a558107-3b00-40b8-9043-45dc89e51dc1-scripts\") pod \"1a558107-3b00-40b8-9043-45dc89e51dc1\" (UID: \"1a558107-3b00-40b8-9043-45dc89e51dc1\") " Mar 07 15:10:21 crc kubenswrapper[4943]: I0307 15:10:21.704040 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1a558107-3b00-40b8-9043-45dc89e51dc1-dispersionconf\") pod \"1a558107-3b00-40b8-9043-45dc89e51dc1\" (UID: \"1a558107-3b00-40b8-9043-45dc89e51dc1\") " Mar 07 15:10:21 crc kubenswrapper[4943]: I0307 15:10:21.705703 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a558107-3b00-40b8-9043-45dc89e51dc1-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "1a558107-3b00-40b8-9043-45dc89e51dc1" (UID: "1a558107-3b00-40b8-9043-45dc89e51dc1"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:10:21 crc kubenswrapper[4943]: I0307 15:10:21.705809 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a558107-3b00-40b8-9043-45dc89e51dc1-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "1a558107-3b00-40b8-9043-45dc89e51dc1" (UID: "1a558107-3b00-40b8-9043-45dc89e51dc1"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:10:21 crc kubenswrapper[4943]: I0307 15:10:21.712045 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a558107-3b00-40b8-9043-45dc89e51dc1-kube-api-access-bkrlp" (OuterVolumeSpecName: "kube-api-access-bkrlp") pod "1a558107-3b00-40b8-9043-45dc89e51dc1" (UID: "1a558107-3b00-40b8-9043-45dc89e51dc1"). InnerVolumeSpecName "kube-api-access-bkrlp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:10:21 crc kubenswrapper[4943]: I0307 15:10:21.731584 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a558107-3b00-40b8-9043-45dc89e51dc1-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "1a558107-3b00-40b8-9043-45dc89e51dc1" (UID: "1a558107-3b00-40b8-9043-45dc89e51dc1"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:10:21 crc kubenswrapper[4943]: I0307 15:10:21.733378 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a558107-3b00-40b8-9043-45dc89e51dc1-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "1a558107-3b00-40b8-9043-45dc89e51dc1" (UID: "1a558107-3b00-40b8-9043-45dc89e51dc1"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:10:21 crc kubenswrapper[4943]: I0307 15:10:21.739227 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a558107-3b00-40b8-9043-45dc89e51dc1-scripts" (OuterVolumeSpecName: "scripts") pod "1a558107-3b00-40b8-9043-45dc89e51dc1" (UID: "1a558107-3b00-40b8-9043-45dc89e51dc1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:10:21 crc kubenswrapper[4943]: I0307 15:10:21.806183 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1a558107-3b00-40b8-9043-45dc89e51dc1-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:21 crc kubenswrapper[4943]: I0307 15:10:21.806232 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1a558107-3b00-40b8-9043-45dc89e51dc1-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:21 crc kubenswrapper[4943]: I0307 15:10:21.806255 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bkrlp\" (UniqueName: \"kubernetes.io/projected/1a558107-3b00-40b8-9043-45dc89e51dc1-kube-api-access-bkrlp\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:21 crc kubenswrapper[4943]: I0307 15:10:21.806276 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1a558107-3b00-40b8-9043-45dc89e51dc1-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:21 crc kubenswrapper[4943]: I0307 15:10:21.806297 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1a558107-3b00-40b8-9043-45dc89e51dc1-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:21 crc kubenswrapper[4943]: I0307 15:10:21.806313 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1a558107-3b00-40b8-9043-45dc89e51dc1-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:22 crc kubenswrapper[4943]: I0307 15:10:22.244867 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="96271b1df8672db5151a7bbe3d5217a7a1b46b10bb1279f028a63fbe6acf67a8" Mar 07 15:10:22 crc kubenswrapper[4943]: I0307 15:10:22.245372 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4ng8" Mar 07 15:10:22 crc kubenswrapper[4943]: I0307 15:10:22.704427 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-9fmwb"] Mar 07 15:10:22 crc kubenswrapper[4943]: E0307 15:10:22.704750 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a558107-3b00-40b8-9043-45dc89e51dc1" containerName="swift-ring-rebalance" Mar 07 15:10:22 crc kubenswrapper[4943]: I0307 15:10:22.704761 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a558107-3b00-40b8-9043-45dc89e51dc1" containerName="swift-ring-rebalance" Mar 07 15:10:22 crc kubenswrapper[4943]: I0307 15:10:22.704918 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a558107-3b00-40b8-9043-45dc89e51dc1" containerName="swift-ring-rebalance" Mar 07 15:10:22 crc kubenswrapper[4943]: I0307 15:10:22.705500 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-9fmwb" Mar 07 15:10:22 crc kubenswrapper[4943]: I0307 15:10:22.708490 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:10:22 crc kubenswrapper[4943]: I0307 15:10:22.708572 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:10:22 crc kubenswrapper[4943]: I0307 15:10:22.722764 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-9fmwb"] Mar 07 15:10:22 crc kubenswrapper[4943]: I0307 15:10:22.765556 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a558107-3b00-40b8-9043-45dc89e51dc1" path="/var/lib/kubelet/pods/1a558107-3b00-40b8-9043-45dc89e51dc1/volumes" Mar 07 15:10:22 crc kubenswrapper[4943]: I0307 15:10:22.822747 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdjsk\" (UniqueName: \"kubernetes.io/projected/4287b075-3bd0-46df-a06b-c8e9699e9ff3-kube-api-access-mdjsk\") pod \"swift-ring-rebalance-debug-9fmwb\" (UID: \"4287b075-3bd0-46df-a06b-c8e9699e9ff3\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-9fmwb" Mar 07 15:10:22 crc kubenswrapper[4943]: I0307 15:10:22.823045 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/4287b075-3bd0-46df-a06b-c8e9699e9ff3-dispersionconf\") pod \"swift-ring-rebalance-debug-9fmwb\" (UID: \"4287b075-3bd0-46df-a06b-c8e9699e9ff3\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-9fmwb" Mar 07 15:10:22 crc kubenswrapper[4943]: I0307 15:10:22.823129 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/4287b075-3bd0-46df-a06b-c8e9699e9ff3-ring-data-devices\") pod \"swift-ring-rebalance-debug-9fmwb\" (UID: \"4287b075-3bd0-46df-a06b-c8e9699e9ff3\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-9fmwb" Mar 07 15:10:22 crc kubenswrapper[4943]: I0307 15:10:22.823174 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/4287b075-3bd0-46df-a06b-c8e9699e9ff3-etc-swift\") pod \"swift-ring-rebalance-debug-9fmwb\" (UID: \"4287b075-3bd0-46df-a06b-c8e9699e9ff3\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-9fmwb" Mar 07 15:10:22 crc kubenswrapper[4943]: I0307 15:10:22.823238 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/4287b075-3bd0-46df-a06b-c8e9699e9ff3-swiftconf\") pod \"swift-ring-rebalance-debug-9fmwb\" (UID: \"4287b075-3bd0-46df-a06b-c8e9699e9ff3\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-9fmwb" Mar 07 15:10:22 crc kubenswrapper[4943]: I0307 15:10:22.823413 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4287b075-3bd0-46df-a06b-c8e9699e9ff3-scripts\") pod \"swift-ring-rebalance-debug-9fmwb\" (UID: \"4287b075-3bd0-46df-a06b-c8e9699e9ff3\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-9fmwb" Mar 07 15:10:22 crc kubenswrapper[4943]: I0307 15:10:22.925362 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdjsk\" (UniqueName: \"kubernetes.io/projected/4287b075-3bd0-46df-a06b-c8e9699e9ff3-kube-api-access-mdjsk\") pod \"swift-ring-rebalance-debug-9fmwb\" (UID: \"4287b075-3bd0-46df-a06b-c8e9699e9ff3\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-9fmwb" Mar 07 15:10:22 crc kubenswrapper[4943]: I0307 15:10:22.925460 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/4287b075-3bd0-46df-a06b-c8e9699e9ff3-dispersionconf\") pod \"swift-ring-rebalance-debug-9fmwb\" (UID: \"4287b075-3bd0-46df-a06b-c8e9699e9ff3\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-9fmwb" Mar 07 15:10:22 crc kubenswrapper[4943]: I0307 15:10:22.925486 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/4287b075-3bd0-46df-a06b-c8e9699e9ff3-ring-data-devices\") pod \"swift-ring-rebalance-debug-9fmwb\" (UID: \"4287b075-3bd0-46df-a06b-c8e9699e9ff3\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-9fmwb" Mar 07 15:10:22 crc kubenswrapper[4943]: I0307 15:10:22.925512 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/4287b075-3bd0-46df-a06b-c8e9699e9ff3-etc-swift\") pod \"swift-ring-rebalance-debug-9fmwb\" (UID: \"4287b075-3bd0-46df-a06b-c8e9699e9ff3\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-9fmwb" Mar 07 15:10:22 crc kubenswrapper[4943]: I0307 15:10:22.925539 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/4287b075-3bd0-46df-a06b-c8e9699e9ff3-swiftconf\") pod \"swift-ring-rebalance-debug-9fmwb\" (UID: \"4287b075-3bd0-46df-a06b-c8e9699e9ff3\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-9fmwb" Mar 07 15:10:22 crc kubenswrapper[4943]: I0307 15:10:22.925585 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4287b075-3bd0-46df-a06b-c8e9699e9ff3-scripts\") pod \"swift-ring-rebalance-debug-9fmwb\" (UID: \"4287b075-3bd0-46df-a06b-c8e9699e9ff3\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-9fmwb" Mar 07 15:10:22 crc kubenswrapper[4943]: I0307 15:10:22.925975 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/4287b075-3bd0-46df-a06b-c8e9699e9ff3-etc-swift\") pod \"swift-ring-rebalance-debug-9fmwb\" (UID: \"4287b075-3bd0-46df-a06b-c8e9699e9ff3\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-9fmwb" Mar 07 15:10:22 crc kubenswrapper[4943]: I0307 15:10:22.926571 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4287b075-3bd0-46df-a06b-c8e9699e9ff3-scripts\") pod \"swift-ring-rebalance-debug-9fmwb\" (UID: \"4287b075-3bd0-46df-a06b-c8e9699e9ff3\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-9fmwb" Mar 07 15:10:22 crc kubenswrapper[4943]: I0307 15:10:22.926970 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/4287b075-3bd0-46df-a06b-c8e9699e9ff3-ring-data-devices\") pod \"swift-ring-rebalance-debug-9fmwb\" (UID: \"4287b075-3bd0-46df-a06b-c8e9699e9ff3\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-9fmwb" Mar 07 15:10:22 crc kubenswrapper[4943]: I0307 15:10:22.929874 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/4287b075-3bd0-46df-a06b-c8e9699e9ff3-dispersionconf\") pod \"swift-ring-rebalance-debug-9fmwb\" (UID: \"4287b075-3bd0-46df-a06b-c8e9699e9ff3\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-9fmwb" Mar 07 15:10:22 crc kubenswrapper[4943]: I0307 15:10:22.933900 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/4287b075-3bd0-46df-a06b-c8e9699e9ff3-swiftconf\") pod \"swift-ring-rebalance-debug-9fmwb\" (UID: \"4287b075-3bd0-46df-a06b-c8e9699e9ff3\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-9fmwb" Mar 07 15:10:22 crc kubenswrapper[4943]: I0307 15:10:22.945895 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdjsk\" (UniqueName: \"kubernetes.io/projected/4287b075-3bd0-46df-a06b-c8e9699e9ff3-kube-api-access-mdjsk\") pod \"swift-ring-rebalance-debug-9fmwb\" (UID: \"4287b075-3bd0-46df-a06b-c8e9699e9ff3\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-9fmwb" Mar 07 15:10:23 crc kubenswrapper[4943]: I0307 15:10:23.033513 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-9fmwb" Mar 07 15:10:23 crc kubenswrapper[4943]: I0307 15:10:23.295858 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-9fmwb"] Mar 07 15:10:24 crc kubenswrapper[4943]: I0307 15:10:24.271960 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-9fmwb" event={"ID":"4287b075-3bd0-46df-a06b-c8e9699e9ff3","Type":"ContainerStarted","Data":"7b29381c6714b878aa56e314ad179792ca8c473e39c9853b2c64065b9ce0bbfd"} Mar 07 15:10:24 crc kubenswrapper[4943]: I0307 15:10:24.272341 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-9fmwb" event={"ID":"4287b075-3bd0-46df-a06b-c8e9699e9ff3","Type":"ContainerStarted","Data":"b9d320e7d9271bfa101b8104eadef1c72edfaf0a6deefc526907b48d71b6851a"} Mar 07 15:10:24 crc kubenswrapper[4943]: I0307 15:10:24.312078 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-9fmwb" podStartSLOduration=2.312046436 podStartE2EDuration="2.312046436s" podCreationTimestamp="2026-03-07 15:10:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:10:24.30528919 +0000 UTC m=+1866.257425728" watchObservedRunningTime="2026-03-07 15:10:24.312046436 +0000 UTC m=+1866.264182974" Mar 07 15:10:25 crc kubenswrapper[4943]: I0307 15:10:25.287406 4943 generic.go:334] "Generic (PLEG): container finished" podID="4287b075-3bd0-46df-a06b-c8e9699e9ff3" containerID="7b29381c6714b878aa56e314ad179792ca8c473e39c9853b2c64065b9ce0bbfd" exitCode=0 Mar 07 15:10:25 crc kubenswrapper[4943]: I0307 15:10:25.287476 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-9fmwb" event={"ID":"4287b075-3bd0-46df-a06b-c8e9699e9ff3","Type":"ContainerDied","Data":"7b29381c6714b878aa56e314ad179792ca8c473e39c9853b2c64065b9ce0bbfd"} Mar 07 15:10:26 crc kubenswrapper[4943]: I0307 15:10:26.692827 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-9fmwb" Mar 07 15:10:26 crc kubenswrapper[4943]: I0307 15:10:26.776428 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-9fmwb"] Mar 07 15:10:26 crc kubenswrapper[4943]: I0307 15:10:26.776487 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-9fmwb"] Mar 07 15:10:26 crc kubenswrapper[4943]: I0307 15:10:26.798667 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4287b075-3bd0-46df-a06b-c8e9699e9ff3-scripts\") pod \"4287b075-3bd0-46df-a06b-c8e9699e9ff3\" (UID: \"4287b075-3bd0-46df-a06b-c8e9699e9ff3\") " Mar 07 15:10:26 crc kubenswrapper[4943]: I0307 15:10:26.798891 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/4287b075-3bd0-46df-a06b-c8e9699e9ff3-swiftconf\") pod \"4287b075-3bd0-46df-a06b-c8e9699e9ff3\" (UID: \"4287b075-3bd0-46df-a06b-c8e9699e9ff3\") " Mar 07 15:10:26 crc kubenswrapper[4943]: I0307 15:10:26.799561 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/4287b075-3bd0-46df-a06b-c8e9699e9ff3-ring-data-devices\") pod \"4287b075-3bd0-46df-a06b-c8e9699e9ff3\" (UID: \"4287b075-3bd0-46df-a06b-c8e9699e9ff3\") " Mar 07 15:10:26 crc kubenswrapper[4943]: I0307 15:10:26.799737 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/4287b075-3bd0-46df-a06b-c8e9699e9ff3-dispersionconf\") pod \"4287b075-3bd0-46df-a06b-c8e9699e9ff3\" (UID: \"4287b075-3bd0-46df-a06b-c8e9699e9ff3\") " Mar 07 15:10:26 crc kubenswrapper[4943]: I0307 15:10:26.799881 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mdjsk\" (UniqueName: \"kubernetes.io/projected/4287b075-3bd0-46df-a06b-c8e9699e9ff3-kube-api-access-mdjsk\") pod \"4287b075-3bd0-46df-a06b-c8e9699e9ff3\" (UID: \"4287b075-3bd0-46df-a06b-c8e9699e9ff3\") " Mar 07 15:10:26 crc kubenswrapper[4943]: I0307 15:10:26.800105 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/4287b075-3bd0-46df-a06b-c8e9699e9ff3-etc-swift\") pod \"4287b075-3bd0-46df-a06b-c8e9699e9ff3\" (UID: \"4287b075-3bd0-46df-a06b-c8e9699e9ff3\") " Mar 07 15:10:26 crc kubenswrapper[4943]: I0307 15:10:26.800567 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4287b075-3bd0-46df-a06b-c8e9699e9ff3-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "4287b075-3bd0-46df-a06b-c8e9699e9ff3" (UID: "4287b075-3bd0-46df-a06b-c8e9699e9ff3"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:10:26 crc kubenswrapper[4943]: I0307 15:10:26.800793 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/4287b075-3bd0-46df-a06b-c8e9699e9ff3-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:26 crc kubenswrapper[4943]: I0307 15:10:26.801314 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4287b075-3bd0-46df-a06b-c8e9699e9ff3-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "4287b075-3bd0-46df-a06b-c8e9699e9ff3" (UID: "4287b075-3bd0-46df-a06b-c8e9699e9ff3"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:10:26 crc kubenswrapper[4943]: I0307 15:10:26.808018 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4287b075-3bd0-46df-a06b-c8e9699e9ff3-kube-api-access-mdjsk" (OuterVolumeSpecName: "kube-api-access-mdjsk") pod "4287b075-3bd0-46df-a06b-c8e9699e9ff3" (UID: "4287b075-3bd0-46df-a06b-c8e9699e9ff3"). InnerVolumeSpecName "kube-api-access-mdjsk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:10:26 crc kubenswrapper[4943]: I0307 15:10:26.821077 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4287b075-3bd0-46df-a06b-c8e9699e9ff3-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "4287b075-3bd0-46df-a06b-c8e9699e9ff3" (UID: "4287b075-3bd0-46df-a06b-c8e9699e9ff3"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:10:26 crc kubenswrapper[4943]: I0307 15:10:26.830766 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4287b075-3bd0-46df-a06b-c8e9699e9ff3-scripts" (OuterVolumeSpecName: "scripts") pod "4287b075-3bd0-46df-a06b-c8e9699e9ff3" (UID: "4287b075-3bd0-46df-a06b-c8e9699e9ff3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:10:26 crc kubenswrapper[4943]: I0307 15:10:26.838767 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4287b075-3bd0-46df-a06b-c8e9699e9ff3-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "4287b075-3bd0-46df-a06b-c8e9699e9ff3" (UID: "4287b075-3bd0-46df-a06b-c8e9699e9ff3"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:10:26 crc kubenswrapper[4943]: I0307 15:10:26.902718 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4287b075-3bd0-46df-a06b-c8e9699e9ff3-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:26 crc kubenswrapper[4943]: I0307 15:10:26.902775 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/4287b075-3bd0-46df-a06b-c8e9699e9ff3-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:26 crc kubenswrapper[4943]: I0307 15:10:26.902796 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/4287b075-3bd0-46df-a06b-c8e9699e9ff3-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:26 crc kubenswrapper[4943]: I0307 15:10:26.902816 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mdjsk\" (UniqueName: \"kubernetes.io/projected/4287b075-3bd0-46df-a06b-c8e9699e9ff3-kube-api-access-mdjsk\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:26 crc kubenswrapper[4943]: I0307 15:10:26.902830 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/4287b075-3bd0-46df-a06b-c8e9699e9ff3-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:27 crc kubenswrapper[4943]: I0307 15:10:27.305440 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b9d320e7d9271bfa101b8104eadef1c72edfaf0a6deefc526907b48d71b6851a" Mar 07 15:10:27 crc kubenswrapper[4943]: I0307 15:10:27.305501 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-9fmwb" Mar 07 15:10:27 crc kubenswrapper[4943]: I0307 15:10:27.928234 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-pkncg"] Mar 07 15:10:27 crc kubenswrapper[4943]: E0307 15:10:27.930125 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4287b075-3bd0-46df-a06b-c8e9699e9ff3" containerName="swift-ring-rebalance" Mar 07 15:10:27 crc kubenswrapper[4943]: I0307 15:10:27.930272 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="4287b075-3bd0-46df-a06b-c8e9699e9ff3" containerName="swift-ring-rebalance" Mar 07 15:10:27 crc kubenswrapper[4943]: I0307 15:10:27.930873 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="4287b075-3bd0-46df-a06b-c8e9699e9ff3" containerName="swift-ring-rebalance" Mar 07 15:10:27 crc kubenswrapper[4943]: I0307 15:10:27.931793 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pkncg" Mar 07 15:10:27 crc kubenswrapper[4943]: I0307 15:10:27.936893 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:10:27 crc kubenswrapper[4943]: I0307 15:10:27.938213 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:10:27 crc kubenswrapper[4943]: I0307 15:10:27.938583 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-pkncg"] Mar 07 15:10:28 crc kubenswrapper[4943]: I0307 15:10:28.122196 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/176fd40b-9bf9-4774-afba-58cc9cea7704-etc-swift\") pod \"swift-ring-rebalance-debug-pkncg\" (UID: \"176fd40b-9bf9-4774-afba-58cc9cea7704\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pkncg" Mar 07 15:10:28 crc kubenswrapper[4943]: I0307 15:10:28.122560 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/176fd40b-9bf9-4774-afba-58cc9cea7704-scripts\") pod \"swift-ring-rebalance-debug-pkncg\" (UID: \"176fd40b-9bf9-4774-afba-58cc9cea7704\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pkncg" Mar 07 15:10:28 crc kubenswrapper[4943]: I0307 15:10:28.122652 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/176fd40b-9bf9-4774-afba-58cc9cea7704-dispersionconf\") pod \"swift-ring-rebalance-debug-pkncg\" (UID: \"176fd40b-9bf9-4774-afba-58cc9cea7704\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pkncg" Mar 07 15:10:28 crc kubenswrapper[4943]: I0307 15:10:28.122678 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/176fd40b-9bf9-4774-afba-58cc9cea7704-swiftconf\") pod \"swift-ring-rebalance-debug-pkncg\" (UID: \"176fd40b-9bf9-4774-afba-58cc9cea7704\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pkncg" Mar 07 15:10:28 crc kubenswrapper[4943]: I0307 15:10:28.122779 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvvxg\" (UniqueName: \"kubernetes.io/projected/176fd40b-9bf9-4774-afba-58cc9cea7704-kube-api-access-vvvxg\") pod \"swift-ring-rebalance-debug-pkncg\" (UID: \"176fd40b-9bf9-4774-afba-58cc9cea7704\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pkncg" Mar 07 15:10:28 crc kubenswrapper[4943]: I0307 15:10:28.122820 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/176fd40b-9bf9-4774-afba-58cc9cea7704-ring-data-devices\") pod \"swift-ring-rebalance-debug-pkncg\" (UID: \"176fd40b-9bf9-4774-afba-58cc9cea7704\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pkncg" Mar 07 15:10:28 crc kubenswrapper[4943]: I0307 15:10:28.224117 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/176fd40b-9bf9-4774-afba-58cc9cea7704-ring-data-devices\") pod \"swift-ring-rebalance-debug-pkncg\" (UID: \"176fd40b-9bf9-4774-afba-58cc9cea7704\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pkncg" Mar 07 15:10:28 crc kubenswrapper[4943]: I0307 15:10:28.224182 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/176fd40b-9bf9-4774-afba-58cc9cea7704-etc-swift\") pod \"swift-ring-rebalance-debug-pkncg\" (UID: \"176fd40b-9bf9-4774-afba-58cc9cea7704\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pkncg" Mar 07 15:10:28 crc kubenswrapper[4943]: I0307 15:10:28.224221 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/176fd40b-9bf9-4774-afba-58cc9cea7704-scripts\") pod \"swift-ring-rebalance-debug-pkncg\" (UID: \"176fd40b-9bf9-4774-afba-58cc9cea7704\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pkncg" Mar 07 15:10:28 crc kubenswrapper[4943]: I0307 15:10:28.224301 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/176fd40b-9bf9-4774-afba-58cc9cea7704-dispersionconf\") pod \"swift-ring-rebalance-debug-pkncg\" (UID: \"176fd40b-9bf9-4774-afba-58cc9cea7704\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pkncg" Mar 07 15:10:28 crc kubenswrapper[4943]: I0307 15:10:28.224321 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/176fd40b-9bf9-4774-afba-58cc9cea7704-swiftconf\") pod \"swift-ring-rebalance-debug-pkncg\" (UID: \"176fd40b-9bf9-4774-afba-58cc9cea7704\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pkncg" Mar 07 15:10:28 crc kubenswrapper[4943]: I0307 15:10:28.224365 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvvxg\" (UniqueName: \"kubernetes.io/projected/176fd40b-9bf9-4774-afba-58cc9cea7704-kube-api-access-vvvxg\") pod \"swift-ring-rebalance-debug-pkncg\" (UID: \"176fd40b-9bf9-4774-afba-58cc9cea7704\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pkncg" Mar 07 15:10:28 crc kubenswrapper[4943]: I0307 15:10:28.226226 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/176fd40b-9bf9-4774-afba-58cc9cea7704-etc-swift\") pod \"swift-ring-rebalance-debug-pkncg\" (UID: \"176fd40b-9bf9-4774-afba-58cc9cea7704\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pkncg" Mar 07 15:10:28 crc kubenswrapper[4943]: I0307 15:10:28.226384 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/176fd40b-9bf9-4774-afba-58cc9cea7704-scripts\") pod \"swift-ring-rebalance-debug-pkncg\" (UID: \"176fd40b-9bf9-4774-afba-58cc9cea7704\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pkncg" Mar 07 15:10:28 crc kubenswrapper[4943]: I0307 15:10:28.227030 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/176fd40b-9bf9-4774-afba-58cc9cea7704-ring-data-devices\") pod \"swift-ring-rebalance-debug-pkncg\" (UID: \"176fd40b-9bf9-4774-afba-58cc9cea7704\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pkncg" Mar 07 15:10:28 crc kubenswrapper[4943]: I0307 15:10:28.232658 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/176fd40b-9bf9-4774-afba-58cc9cea7704-dispersionconf\") pod \"swift-ring-rebalance-debug-pkncg\" (UID: \"176fd40b-9bf9-4774-afba-58cc9cea7704\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pkncg" Mar 07 15:10:28 crc kubenswrapper[4943]: I0307 15:10:28.235204 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/176fd40b-9bf9-4774-afba-58cc9cea7704-swiftconf\") pod \"swift-ring-rebalance-debug-pkncg\" (UID: \"176fd40b-9bf9-4774-afba-58cc9cea7704\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pkncg" Mar 07 15:10:28 crc kubenswrapper[4943]: I0307 15:10:28.247868 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvvxg\" (UniqueName: \"kubernetes.io/projected/176fd40b-9bf9-4774-afba-58cc9cea7704-kube-api-access-vvvxg\") pod \"swift-ring-rebalance-debug-pkncg\" (UID: \"176fd40b-9bf9-4774-afba-58cc9cea7704\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pkncg" Mar 07 15:10:28 crc kubenswrapper[4943]: I0307 15:10:28.282605 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pkncg" Mar 07 15:10:28 crc kubenswrapper[4943]: W0307 15:10:28.753643 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod176fd40b_9bf9_4774_afba_58cc9cea7704.slice/crio-bf3a144988e7f43c21196ccecaadbd67eb8d5e933ea7ffe136de8227d71311e3 WatchSource:0}: Error finding container bf3a144988e7f43c21196ccecaadbd67eb8d5e933ea7ffe136de8227d71311e3: Status 404 returned error can't find the container with id bf3a144988e7f43c21196ccecaadbd67eb8d5e933ea7ffe136de8227d71311e3 Mar 07 15:10:28 crc kubenswrapper[4943]: I0307 15:10:28.770171 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4287b075-3bd0-46df-a06b-c8e9699e9ff3" path="/var/lib/kubelet/pods/4287b075-3bd0-46df-a06b-c8e9699e9ff3/volumes" Mar 07 15:10:28 crc kubenswrapper[4943]: I0307 15:10:28.771479 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-pkncg"] Mar 07 15:10:29 crc kubenswrapper[4943]: I0307 15:10:29.340885 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pkncg" event={"ID":"176fd40b-9bf9-4774-afba-58cc9cea7704","Type":"ContainerStarted","Data":"5856b9b9b3ebd822cf233d14dab3b104f8c51a4851c2cd4234792e319f5b72ff"} Mar 07 15:10:29 crc kubenswrapper[4943]: I0307 15:10:29.341241 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pkncg" event={"ID":"176fd40b-9bf9-4774-afba-58cc9cea7704","Type":"ContainerStarted","Data":"bf3a144988e7f43c21196ccecaadbd67eb8d5e933ea7ffe136de8227d71311e3"} Mar 07 15:10:29 crc kubenswrapper[4943]: I0307 15:10:29.370781 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pkncg" podStartSLOduration=2.370758014 podStartE2EDuration="2.370758014s" podCreationTimestamp="2026-03-07 15:10:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:10:29.360799959 +0000 UTC m=+1871.312936497" watchObservedRunningTime="2026-03-07 15:10:29.370758014 +0000 UTC m=+1871.322894542" Mar 07 15:10:30 crc kubenswrapper[4943]: I0307 15:10:30.352659 4943 generic.go:334] "Generic (PLEG): container finished" podID="176fd40b-9bf9-4774-afba-58cc9cea7704" containerID="5856b9b9b3ebd822cf233d14dab3b104f8c51a4851c2cd4234792e319f5b72ff" exitCode=0 Mar 07 15:10:30 crc kubenswrapper[4943]: I0307 15:10:30.352708 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pkncg" event={"ID":"176fd40b-9bf9-4774-afba-58cc9cea7704","Type":"ContainerDied","Data":"5856b9b9b3ebd822cf233d14dab3b104f8c51a4851c2cd4234792e319f5b72ff"} Mar 07 15:10:31 crc kubenswrapper[4943]: I0307 15:10:31.744221 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pkncg" Mar 07 15:10:31 crc kubenswrapper[4943]: I0307 15:10:31.799058 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-pkncg"] Mar 07 15:10:31 crc kubenswrapper[4943]: I0307 15:10:31.807055 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-pkncg"] Mar 07 15:10:31 crc kubenswrapper[4943]: I0307 15:10:31.883837 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/176fd40b-9bf9-4774-afba-58cc9cea7704-ring-data-devices\") pod \"176fd40b-9bf9-4774-afba-58cc9cea7704\" (UID: \"176fd40b-9bf9-4774-afba-58cc9cea7704\") " Mar 07 15:10:31 crc kubenswrapper[4943]: I0307 15:10:31.883957 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/176fd40b-9bf9-4774-afba-58cc9cea7704-swiftconf\") pod \"176fd40b-9bf9-4774-afba-58cc9cea7704\" (UID: \"176fd40b-9bf9-4774-afba-58cc9cea7704\") " Mar 07 15:10:31 crc kubenswrapper[4943]: I0307 15:10:31.884061 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/176fd40b-9bf9-4774-afba-58cc9cea7704-etc-swift\") pod \"176fd40b-9bf9-4774-afba-58cc9cea7704\" (UID: \"176fd40b-9bf9-4774-afba-58cc9cea7704\") " Mar 07 15:10:31 crc kubenswrapper[4943]: I0307 15:10:31.884118 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/176fd40b-9bf9-4774-afba-58cc9cea7704-dispersionconf\") pod \"176fd40b-9bf9-4774-afba-58cc9cea7704\" (UID: \"176fd40b-9bf9-4774-afba-58cc9cea7704\") " Mar 07 15:10:31 crc kubenswrapper[4943]: I0307 15:10:31.884187 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/176fd40b-9bf9-4774-afba-58cc9cea7704-scripts\") pod \"176fd40b-9bf9-4774-afba-58cc9cea7704\" (UID: \"176fd40b-9bf9-4774-afba-58cc9cea7704\") " Mar 07 15:10:31 crc kubenswrapper[4943]: I0307 15:10:31.884242 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vvvxg\" (UniqueName: \"kubernetes.io/projected/176fd40b-9bf9-4774-afba-58cc9cea7704-kube-api-access-vvvxg\") pod \"176fd40b-9bf9-4774-afba-58cc9cea7704\" (UID: \"176fd40b-9bf9-4774-afba-58cc9cea7704\") " Mar 07 15:10:31 crc kubenswrapper[4943]: I0307 15:10:31.885028 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/176fd40b-9bf9-4774-afba-58cc9cea7704-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "176fd40b-9bf9-4774-afba-58cc9cea7704" (UID: "176fd40b-9bf9-4774-afba-58cc9cea7704"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:10:31 crc kubenswrapper[4943]: I0307 15:10:31.885888 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/176fd40b-9bf9-4774-afba-58cc9cea7704-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "176fd40b-9bf9-4774-afba-58cc9cea7704" (UID: "176fd40b-9bf9-4774-afba-58cc9cea7704"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:10:31 crc kubenswrapper[4943]: I0307 15:10:31.891775 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/176fd40b-9bf9-4774-afba-58cc9cea7704-kube-api-access-vvvxg" (OuterVolumeSpecName: "kube-api-access-vvvxg") pod "176fd40b-9bf9-4774-afba-58cc9cea7704" (UID: "176fd40b-9bf9-4774-afba-58cc9cea7704"). InnerVolumeSpecName "kube-api-access-vvvxg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:10:31 crc kubenswrapper[4943]: I0307 15:10:31.911889 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/176fd40b-9bf9-4774-afba-58cc9cea7704-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "176fd40b-9bf9-4774-afba-58cc9cea7704" (UID: "176fd40b-9bf9-4774-afba-58cc9cea7704"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:10:31 crc kubenswrapper[4943]: I0307 15:10:31.918644 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/176fd40b-9bf9-4774-afba-58cc9cea7704-scripts" (OuterVolumeSpecName: "scripts") pod "176fd40b-9bf9-4774-afba-58cc9cea7704" (UID: "176fd40b-9bf9-4774-afba-58cc9cea7704"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:10:31 crc kubenswrapper[4943]: I0307 15:10:31.920340 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/176fd40b-9bf9-4774-afba-58cc9cea7704-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "176fd40b-9bf9-4774-afba-58cc9cea7704" (UID: "176fd40b-9bf9-4774-afba-58cc9cea7704"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:10:31 crc kubenswrapper[4943]: I0307 15:10:31.986180 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/176fd40b-9bf9-4774-afba-58cc9cea7704-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:31 crc kubenswrapper[4943]: I0307 15:10:31.986243 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/176fd40b-9bf9-4774-afba-58cc9cea7704-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:31 crc kubenswrapper[4943]: I0307 15:10:31.986265 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/176fd40b-9bf9-4774-afba-58cc9cea7704-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:31 crc kubenswrapper[4943]: I0307 15:10:31.986286 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vvvxg\" (UniqueName: \"kubernetes.io/projected/176fd40b-9bf9-4774-afba-58cc9cea7704-kube-api-access-vvvxg\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:31 crc kubenswrapper[4943]: I0307 15:10:31.986304 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/176fd40b-9bf9-4774-afba-58cc9cea7704-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:31 crc kubenswrapper[4943]: I0307 15:10:31.986321 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/176fd40b-9bf9-4774-afba-58cc9cea7704-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:32 crc kubenswrapper[4943]: I0307 15:10:32.374823 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bf3a144988e7f43c21196ccecaadbd67eb8d5e933ea7ffe136de8227d71311e3" Mar 07 15:10:32 crc kubenswrapper[4943]: I0307 15:10:32.374904 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pkncg" Mar 07 15:10:32 crc kubenswrapper[4943]: I0307 15:10:32.769175 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="176fd40b-9bf9-4774-afba-58cc9cea7704" path="/var/lib/kubelet/pods/176fd40b-9bf9-4774-afba-58cc9cea7704/volumes" Mar 07 15:10:33 crc kubenswrapper[4943]: I0307 15:10:33.022794 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7jsd9"] Mar 07 15:10:33 crc kubenswrapper[4943]: E0307 15:10:33.023338 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="176fd40b-9bf9-4774-afba-58cc9cea7704" containerName="swift-ring-rebalance" Mar 07 15:10:33 crc kubenswrapper[4943]: I0307 15:10:33.023362 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="176fd40b-9bf9-4774-afba-58cc9cea7704" containerName="swift-ring-rebalance" Mar 07 15:10:33 crc kubenswrapper[4943]: I0307 15:10:33.023636 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="176fd40b-9bf9-4774-afba-58cc9cea7704" containerName="swift-ring-rebalance" Mar 07 15:10:33 crc kubenswrapper[4943]: I0307 15:10:33.024557 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7jsd9" Mar 07 15:10:33 crc kubenswrapper[4943]: I0307 15:10:33.026583 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:10:33 crc kubenswrapper[4943]: I0307 15:10:33.026606 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:10:33 crc kubenswrapper[4943]: I0307 15:10:33.036445 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7jsd9"] Mar 07 15:10:33 crc kubenswrapper[4943]: I0307 15:10:33.112324 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2a29040d-69a6-42ac-bbb8-e1d69c64c166-swiftconf\") pod \"swift-ring-rebalance-debug-7jsd9\" (UID: \"2a29040d-69a6-42ac-bbb8-e1d69c64c166\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7jsd9" Mar 07 15:10:33 crc kubenswrapper[4943]: I0307 15:10:33.112404 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2a29040d-69a6-42ac-bbb8-e1d69c64c166-dispersionconf\") pod \"swift-ring-rebalance-debug-7jsd9\" (UID: \"2a29040d-69a6-42ac-bbb8-e1d69c64c166\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7jsd9" Mar 07 15:10:33 crc kubenswrapper[4943]: I0307 15:10:33.112439 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2a29040d-69a6-42ac-bbb8-e1d69c64c166-scripts\") pod \"swift-ring-rebalance-debug-7jsd9\" (UID: \"2a29040d-69a6-42ac-bbb8-e1d69c64c166\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7jsd9" Mar 07 15:10:33 crc kubenswrapper[4943]: I0307 15:10:33.112474 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2a29040d-69a6-42ac-bbb8-e1d69c64c166-etc-swift\") pod \"swift-ring-rebalance-debug-7jsd9\" (UID: \"2a29040d-69a6-42ac-bbb8-e1d69c64c166\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7jsd9" Mar 07 15:10:33 crc kubenswrapper[4943]: I0307 15:10:33.112557 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97q8k\" (UniqueName: \"kubernetes.io/projected/2a29040d-69a6-42ac-bbb8-e1d69c64c166-kube-api-access-97q8k\") pod \"swift-ring-rebalance-debug-7jsd9\" (UID: \"2a29040d-69a6-42ac-bbb8-e1d69c64c166\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7jsd9" Mar 07 15:10:33 crc kubenswrapper[4943]: I0307 15:10:33.112583 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2a29040d-69a6-42ac-bbb8-e1d69c64c166-ring-data-devices\") pod \"swift-ring-rebalance-debug-7jsd9\" (UID: \"2a29040d-69a6-42ac-bbb8-e1d69c64c166\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7jsd9" Mar 07 15:10:33 crc kubenswrapper[4943]: I0307 15:10:33.213434 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2a29040d-69a6-42ac-bbb8-e1d69c64c166-swiftconf\") pod \"swift-ring-rebalance-debug-7jsd9\" (UID: \"2a29040d-69a6-42ac-bbb8-e1d69c64c166\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7jsd9" Mar 07 15:10:33 crc kubenswrapper[4943]: I0307 15:10:33.213565 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2a29040d-69a6-42ac-bbb8-e1d69c64c166-dispersionconf\") pod \"swift-ring-rebalance-debug-7jsd9\" (UID: \"2a29040d-69a6-42ac-bbb8-e1d69c64c166\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7jsd9" Mar 07 15:10:33 crc kubenswrapper[4943]: I0307 15:10:33.213610 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2a29040d-69a6-42ac-bbb8-e1d69c64c166-scripts\") pod \"swift-ring-rebalance-debug-7jsd9\" (UID: \"2a29040d-69a6-42ac-bbb8-e1d69c64c166\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7jsd9" Mar 07 15:10:33 crc kubenswrapper[4943]: I0307 15:10:33.213654 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2a29040d-69a6-42ac-bbb8-e1d69c64c166-etc-swift\") pod \"swift-ring-rebalance-debug-7jsd9\" (UID: \"2a29040d-69a6-42ac-bbb8-e1d69c64c166\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7jsd9" Mar 07 15:10:33 crc kubenswrapper[4943]: I0307 15:10:33.213826 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97q8k\" (UniqueName: \"kubernetes.io/projected/2a29040d-69a6-42ac-bbb8-e1d69c64c166-kube-api-access-97q8k\") pod \"swift-ring-rebalance-debug-7jsd9\" (UID: \"2a29040d-69a6-42ac-bbb8-e1d69c64c166\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7jsd9" Mar 07 15:10:33 crc kubenswrapper[4943]: I0307 15:10:33.213884 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2a29040d-69a6-42ac-bbb8-e1d69c64c166-ring-data-devices\") pod \"swift-ring-rebalance-debug-7jsd9\" (UID: \"2a29040d-69a6-42ac-bbb8-e1d69c64c166\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7jsd9" Mar 07 15:10:33 crc kubenswrapper[4943]: I0307 15:10:33.214611 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2a29040d-69a6-42ac-bbb8-e1d69c64c166-scripts\") pod \"swift-ring-rebalance-debug-7jsd9\" (UID: \"2a29040d-69a6-42ac-bbb8-e1d69c64c166\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7jsd9" Mar 07 15:10:33 crc kubenswrapper[4943]: I0307 15:10:33.214659 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2a29040d-69a6-42ac-bbb8-e1d69c64c166-etc-swift\") pod \"swift-ring-rebalance-debug-7jsd9\" (UID: \"2a29040d-69a6-42ac-bbb8-e1d69c64c166\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7jsd9" Mar 07 15:10:33 crc kubenswrapper[4943]: I0307 15:10:33.215181 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2a29040d-69a6-42ac-bbb8-e1d69c64c166-ring-data-devices\") pod \"swift-ring-rebalance-debug-7jsd9\" (UID: \"2a29040d-69a6-42ac-bbb8-e1d69c64c166\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7jsd9" Mar 07 15:10:33 crc kubenswrapper[4943]: I0307 15:10:33.218370 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2a29040d-69a6-42ac-bbb8-e1d69c64c166-swiftconf\") pod \"swift-ring-rebalance-debug-7jsd9\" (UID: \"2a29040d-69a6-42ac-bbb8-e1d69c64c166\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7jsd9" Mar 07 15:10:33 crc kubenswrapper[4943]: I0307 15:10:33.222119 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2a29040d-69a6-42ac-bbb8-e1d69c64c166-dispersionconf\") pod \"swift-ring-rebalance-debug-7jsd9\" (UID: \"2a29040d-69a6-42ac-bbb8-e1d69c64c166\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7jsd9" Mar 07 15:10:33 crc kubenswrapper[4943]: I0307 15:10:33.245606 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97q8k\" (UniqueName: \"kubernetes.io/projected/2a29040d-69a6-42ac-bbb8-e1d69c64c166-kube-api-access-97q8k\") pod \"swift-ring-rebalance-debug-7jsd9\" (UID: \"2a29040d-69a6-42ac-bbb8-e1d69c64c166\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7jsd9" Mar 07 15:10:33 crc kubenswrapper[4943]: I0307 15:10:33.354038 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7jsd9" Mar 07 15:10:33 crc kubenswrapper[4943]: W0307 15:10:33.609722 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2a29040d_69a6_42ac_bbb8_e1d69c64c166.slice/crio-6f35ca2db2623ef7b34c38cd84a934e675654137d1309a9edca2dcd1ede8cada WatchSource:0}: Error finding container 6f35ca2db2623ef7b34c38cd84a934e675654137d1309a9edca2dcd1ede8cada: Status 404 returned error can't find the container with id 6f35ca2db2623ef7b34c38cd84a934e675654137d1309a9edca2dcd1ede8cada Mar 07 15:10:33 crc kubenswrapper[4943]: I0307 15:10:33.615740 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7jsd9"] Mar 07 15:10:34 crc kubenswrapper[4943]: I0307 15:10:34.392028 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7jsd9" event={"ID":"2a29040d-69a6-42ac-bbb8-e1d69c64c166","Type":"ContainerStarted","Data":"897e9786c93077b536da1b7c8d55d5b6d0d17ae15c9c7d354cf6ca12256d7e5c"} Mar 07 15:10:34 crc kubenswrapper[4943]: I0307 15:10:34.392312 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7jsd9" event={"ID":"2a29040d-69a6-42ac-bbb8-e1d69c64c166","Type":"ContainerStarted","Data":"6f35ca2db2623ef7b34c38cd84a934e675654137d1309a9edca2dcd1ede8cada"} Mar 07 15:10:34 crc kubenswrapper[4943]: I0307 15:10:34.416371 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7jsd9" podStartSLOduration=2.416347178 podStartE2EDuration="2.416347178s" podCreationTimestamp="2026-03-07 15:10:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:10:34.407155823 +0000 UTC m=+1876.359292341" watchObservedRunningTime="2026-03-07 15:10:34.416347178 +0000 UTC m=+1876.368483696" Mar 07 15:10:36 crc kubenswrapper[4943]: I0307 15:10:36.418400 4943 generic.go:334] "Generic (PLEG): container finished" podID="2a29040d-69a6-42ac-bbb8-e1d69c64c166" containerID="897e9786c93077b536da1b7c8d55d5b6d0d17ae15c9c7d354cf6ca12256d7e5c" exitCode=0 Mar 07 15:10:36 crc kubenswrapper[4943]: I0307 15:10:36.418518 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7jsd9" event={"ID":"2a29040d-69a6-42ac-bbb8-e1d69c64c166","Type":"ContainerDied","Data":"897e9786c93077b536da1b7c8d55d5b6d0d17ae15c9c7d354cf6ca12256d7e5c"} Mar 07 15:10:37 crc kubenswrapper[4943]: I0307 15:10:37.815980 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7jsd9" Mar 07 15:10:37 crc kubenswrapper[4943]: I0307 15:10:37.855194 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7jsd9"] Mar 07 15:10:37 crc kubenswrapper[4943]: I0307 15:10:37.861243 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7jsd9"] Mar 07 15:10:37 crc kubenswrapper[4943]: I0307 15:10:37.886536 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97q8k\" (UniqueName: \"kubernetes.io/projected/2a29040d-69a6-42ac-bbb8-e1d69c64c166-kube-api-access-97q8k\") pod \"2a29040d-69a6-42ac-bbb8-e1d69c64c166\" (UID: \"2a29040d-69a6-42ac-bbb8-e1d69c64c166\") " Mar 07 15:10:37 crc kubenswrapper[4943]: I0307 15:10:37.886630 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2a29040d-69a6-42ac-bbb8-e1d69c64c166-ring-data-devices\") pod \"2a29040d-69a6-42ac-bbb8-e1d69c64c166\" (UID: \"2a29040d-69a6-42ac-bbb8-e1d69c64c166\") " Mar 07 15:10:37 crc kubenswrapper[4943]: I0307 15:10:37.886693 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2a29040d-69a6-42ac-bbb8-e1d69c64c166-scripts\") pod \"2a29040d-69a6-42ac-bbb8-e1d69c64c166\" (UID: \"2a29040d-69a6-42ac-bbb8-e1d69c64c166\") " Mar 07 15:10:37 crc kubenswrapper[4943]: I0307 15:10:37.886742 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2a29040d-69a6-42ac-bbb8-e1d69c64c166-dispersionconf\") pod \"2a29040d-69a6-42ac-bbb8-e1d69c64c166\" (UID: \"2a29040d-69a6-42ac-bbb8-e1d69c64c166\") " Mar 07 15:10:37 crc kubenswrapper[4943]: I0307 15:10:37.886777 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2a29040d-69a6-42ac-bbb8-e1d69c64c166-swiftconf\") pod \"2a29040d-69a6-42ac-bbb8-e1d69c64c166\" (UID: \"2a29040d-69a6-42ac-bbb8-e1d69c64c166\") " Mar 07 15:10:37 crc kubenswrapper[4943]: I0307 15:10:37.886902 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2a29040d-69a6-42ac-bbb8-e1d69c64c166-etc-swift\") pod \"2a29040d-69a6-42ac-bbb8-e1d69c64c166\" (UID: \"2a29040d-69a6-42ac-bbb8-e1d69c64c166\") " Mar 07 15:10:37 crc kubenswrapper[4943]: I0307 15:10:37.888275 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a29040d-69a6-42ac-bbb8-e1d69c64c166-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "2a29040d-69a6-42ac-bbb8-e1d69c64c166" (UID: "2a29040d-69a6-42ac-bbb8-e1d69c64c166"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:10:37 crc kubenswrapper[4943]: I0307 15:10:37.888724 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a29040d-69a6-42ac-bbb8-e1d69c64c166-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "2a29040d-69a6-42ac-bbb8-e1d69c64c166" (UID: "2a29040d-69a6-42ac-bbb8-e1d69c64c166"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:10:37 crc kubenswrapper[4943]: I0307 15:10:37.895508 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a29040d-69a6-42ac-bbb8-e1d69c64c166-kube-api-access-97q8k" (OuterVolumeSpecName: "kube-api-access-97q8k") pod "2a29040d-69a6-42ac-bbb8-e1d69c64c166" (UID: "2a29040d-69a6-42ac-bbb8-e1d69c64c166"). InnerVolumeSpecName "kube-api-access-97q8k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:10:37 crc kubenswrapper[4943]: I0307 15:10:37.915181 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a29040d-69a6-42ac-bbb8-e1d69c64c166-scripts" (OuterVolumeSpecName: "scripts") pod "2a29040d-69a6-42ac-bbb8-e1d69c64c166" (UID: "2a29040d-69a6-42ac-bbb8-e1d69c64c166"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:10:37 crc kubenswrapper[4943]: I0307 15:10:37.932391 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a29040d-69a6-42ac-bbb8-e1d69c64c166-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "2a29040d-69a6-42ac-bbb8-e1d69c64c166" (UID: "2a29040d-69a6-42ac-bbb8-e1d69c64c166"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:10:37 crc kubenswrapper[4943]: I0307 15:10:37.940905 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a29040d-69a6-42ac-bbb8-e1d69c64c166-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "2a29040d-69a6-42ac-bbb8-e1d69c64c166" (UID: "2a29040d-69a6-42ac-bbb8-e1d69c64c166"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:10:37 crc kubenswrapper[4943]: I0307 15:10:37.988433 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2a29040d-69a6-42ac-bbb8-e1d69c64c166-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:37 crc kubenswrapper[4943]: I0307 15:10:37.988483 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2a29040d-69a6-42ac-bbb8-e1d69c64c166-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:37 crc kubenswrapper[4943]: I0307 15:10:37.988504 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2a29040d-69a6-42ac-bbb8-e1d69c64c166-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:37 crc kubenswrapper[4943]: I0307 15:10:37.988524 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2a29040d-69a6-42ac-bbb8-e1d69c64c166-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:37 crc kubenswrapper[4943]: I0307 15:10:37.988647 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2a29040d-69a6-42ac-bbb8-e1d69c64c166-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:37 crc kubenswrapper[4943]: I0307 15:10:37.988672 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97q8k\" (UniqueName: \"kubernetes.io/projected/2a29040d-69a6-42ac-bbb8-e1d69c64c166-kube-api-access-97q8k\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:38 crc kubenswrapper[4943]: I0307 15:10:38.440868 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6f35ca2db2623ef7b34c38cd84a934e675654137d1309a9edca2dcd1ede8cada" Mar 07 15:10:38 crc kubenswrapper[4943]: I0307 15:10:38.440948 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7jsd9" Mar 07 15:10:38 crc kubenswrapper[4943]: I0307 15:10:38.766708 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a29040d-69a6-42ac-bbb8-e1d69c64c166" path="/var/lib/kubelet/pods/2a29040d-69a6-42ac-bbb8-e1d69c64c166/volumes" Mar 07 15:10:39 crc kubenswrapper[4943]: I0307 15:10:39.057514 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-nl4hw"] Mar 07 15:10:39 crc kubenswrapper[4943]: E0307 15:10:39.057790 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a29040d-69a6-42ac-bbb8-e1d69c64c166" containerName="swift-ring-rebalance" Mar 07 15:10:39 crc kubenswrapper[4943]: I0307 15:10:39.057802 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a29040d-69a6-42ac-bbb8-e1d69c64c166" containerName="swift-ring-rebalance" Mar 07 15:10:39 crc kubenswrapper[4943]: I0307 15:10:39.057966 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a29040d-69a6-42ac-bbb8-e1d69c64c166" containerName="swift-ring-rebalance" Mar 07 15:10:39 crc kubenswrapper[4943]: I0307 15:10:39.058423 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-nl4hw" Mar 07 15:10:39 crc kubenswrapper[4943]: I0307 15:10:39.059991 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:10:39 crc kubenswrapper[4943]: I0307 15:10:39.070774 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:10:39 crc kubenswrapper[4943]: I0307 15:10:39.076604 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-nl4hw"] Mar 07 15:10:39 crc kubenswrapper[4943]: I0307 15:10:39.205999 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d49deefc-a177-4d58-a075-a24ef44ebce1-ring-data-devices\") pod \"swift-ring-rebalance-debug-nl4hw\" (UID: \"d49deefc-a177-4d58-a075-a24ef44ebce1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-nl4hw" Mar 07 15:10:39 crc kubenswrapper[4943]: I0307 15:10:39.206577 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d49deefc-a177-4d58-a075-a24ef44ebce1-dispersionconf\") pod \"swift-ring-rebalance-debug-nl4hw\" (UID: \"d49deefc-a177-4d58-a075-a24ef44ebce1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-nl4hw" Mar 07 15:10:39 crc kubenswrapper[4943]: I0307 15:10:39.206627 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9kbvj\" (UniqueName: \"kubernetes.io/projected/d49deefc-a177-4d58-a075-a24ef44ebce1-kube-api-access-9kbvj\") pod \"swift-ring-rebalance-debug-nl4hw\" (UID: \"d49deefc-a177-4d58-a075-a24ef44ebce1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-nl4hw" Mar 07 15:10:39 crc kubenswrapper[4943]: I0307 15:10:39.206652 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d49deefc-a177-4d58-a075-a24ef44ebce1-etc-swift\") pod \"swift-ring-rebalance-debug-nl4hw\" (UID: \"d49deefc-a177-4d58-a075-a24ef44ebce1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-nl4hw" Mar 07 15:10:39 crc kubenswrapper[4943]: I0307 15:10:39.206808 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d49deefc-a177-4d58-a075-a24ef44ebce1-swiftconf\") pod \"swift-ring-rebalance-debug-nl4hw\" (UID: \"d49deefc-a177-4d58-a075-a24ef44ebce1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-nl4hw" Mar 07 15:10:39 crc kubenswrapper[4943]: I0307 15:10:39.206953 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d49deefc-a177-4d58-a075-a24ef44ebce1-scripts\") pod \"swift-ring-rebalance-debug-nl4hw\" (UID: \"d49deefc-a177-4d58-a075-a24ef44ebce1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-nl4hw" Mar 07 15:10:39 crc kubenswrapper[4943]: I0307 15:10:39.308496 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d49deefc-a177-4d58-a075-a24ef44ebce1-scripts\") pod \"swift-ring-rebalance-debug-nl4hw\" (UID: \"d49deefc-a177-4d58-a075-a24ef44ebce1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-nl4hw" Mar 07 15:10:39 crc kubenswrapper[4943]: I0307 15:10:39.308851 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d49deefc-a177-4d58-a075-a24ef44ebce1-ring-data-devices\") pod \"swift-ring-rebalance-debug-nl4hw\" (UID: \"d49deefc-a177-4d58-a075-a24ef44ebce1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-nl4hw" Mar 07 15:10:39 crc kubenswrapper[4943]: I0307 15:10:39.308959 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d49deefc-a177-4d58-a075-a24ef44ebce1-dispersionconf\") pod \"swift-ring-rebalance-debug-nl4hw\" (UID: \"d49deefc-a177-4d58-a075-a24ef44ebce1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-nl4hw" Mar 07 15:10:39 crc kubenswrapper[4943]: I0307 15:10:39.309051 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9kbvj\" (UniqueName: \"kubernetes.io/projected/d49deefc-a177-4d58-a075-a24ef44ebce1-kube-api-access-9kbvj\") pod \"swift-ring-rebalance-debug-nl4hw\" (UID: \"d49deefc-a177-4d58-a075-a24ef44ebce1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-nl4hw" Mar 07 15:10:39 crc kubenswrapper[4943]: I0307 15:10:39.309133 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d49deefc-a177-4d58-a075-a24ef44ebce1-etc-swift\") pod \"swift-ring-rebalance-debug-nl4hw\" (UID: \"d49deefc-a177-4d58-a075-a24ef44ebce1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-nl4hw" Mar 07 15:10:39 crc kubenswrapper[4943]: I0307 15:10:39.309234 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d49deefc-a177-4d58-a075-a24ef44ebce1-swiftconf\") pod \"swift-ring-rebalance-debug-nl4hw\" (UID: \"d49deefc-a177-4d58-a075-a24ef44ebce1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-nl4hw" Mar 07 15:10:39 crc kubenswrapper[4943]: I0307 15:10:39.309298 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d49deefc-a177-4d58-a075-a24ef44ebce1-scripts\") pod \"swift-ring-rebalance-debug-nl4hw\" (UID: \"d49deefc-a177-4d58-a075-a24ef44ebce1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-nl4hw" Mar 07 15:10:39 crc kubenswrapper[4943]: I0307 15:10:39.309476 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d49deefc-a177-4d58-a075-a24ef44ebce1-etc-swift\") pod \"swift-ring-rebalance-debug-nl4hw\" (UID: \"d49deefc-a177-4d58-a075-a24ef44ebce1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-nl4hw" Mar 07 15:10:39 crc kubenswrapper[4943]: I0307 15:10:39.309633 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d49deefc-a177-4d58-a075-a24ef44ebce1-ring-data-devices\") pod \"swift-ring-rebalance-debug-nl4hw\" (UID: \"d49deefc-a177-4d58-a075-a24ef44ebce1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-nl4hw" Mar 07 15:10:39 crc kubenswrapper[4943]: I0307 15:10:39.312994 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d49deefc-a177-4d58-a075-a24ef44ebce1-dispersionconf\") pod \"swift-ring-rebalance-debug-nl4hw\" (UID: \"d49deefc-a177-4d58-a075-a24ef44ebce1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-nl4hw" Mar 07 15:10:39 crc kubenswrapper[4943]: I0307 15:10:39.320553 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d49deefc-a177-4d58-a075-a24ef44ebce1-swiftconf\") pod \"swift-ring-rebalance-debug-nl4hw\" (UID: \"d49deefc-a177-4d58-a075-a24ef44ebce1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-nl4hw" Mar 07 15:10:39 crc kubenswrapper[4943]: I0307 15:10:39.324619 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9kbvj\" (UniqueName: \"kubernetes.io/projected/d49deefc-a177-4d58-a075-a24ef44ebce1-kube-api-access-9kbvj\") pod \"swift-ring-rebalance-debug-nl4hw\" (UID: \"d49deefc-a177-4d58-a075-a24ef44ebce1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-nl4hw" Mar 07 15:10:39 crc kubenswrapper[4943]: I0307 15:10:39.374537 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-nl4hw" Mar 07 15:10:39 crc kubenswrapper[4943]: I0307 15:10:39.578082 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-nl4hw"] Mar 07 15:10:40 crc kubenswrapper[4943]: I0307 15:10:40.463705 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-nl4hw" event={"ID":"d49deefc-a177-4d58-a075-a24ef44ebce1","Type":"ContainerStarted","Data":"4f054df4e375fb2c1e152506750925311e103138ee9c7e4f5dfe2a8284ff67ce"} Mar 07 15:10:40 crc kubenswrapper[4943]: I0307 15:10:40.464224 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-nl4hw" event={"ID":"d49deefc-a177-4d58-a075-a24ef44ebce1","Type":"ContainerStarted","Data":"cc072e3436ca706bb9623d316c7175bacc6679d0c69d6c8c3c70de1eb20b407f"} Mar 07 15:10:40 crc kubenswrapper[4943]: I0307 15:10:40.503546 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-nl4hw" podStartSLOduration=1.503525526 podStartE2EDuration="1.503525526s" podCreationTimestamp="2026-03-07 15:10:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:10:40.486104919 +0000 UTC m=+1882.438241467" watchObservedRunningTime="2026-03-07 15:10:40.503525526 +0000 UTC m=+1882.455662064" Mar 07 15:10:41 crc kubenswrapper[4943]: I0307 15:10:41.476224 4943 generic.go:334] "Generic (PLEG): container finished" podID="d49deefc-a177-4d58-a075-a24ef44ebce1" containerID="4f054df4e375fb2c1e152506750925311e103138ee9c7e4f5dfe2a8284ff67ce" exitCode=0 Mar 07 15:10:41 crc kubenswrapper[4943]: I0307 15:10:41.476276 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-nl4hw" event={"ID":"d49deefc-a177-4d58-a075-a24ef44ebce1","Type":"ContainerDied","Data":"4f054df4e375fb2c1e152506750925311e103138ee9c7e4f5dfe2a8284ff67ce"} Mar 07 15:10:42 crc kubenswrapper[4943]: I0307 15:10:42.823819 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-nl4hw" Mar 07 15:10:42 crc kubenswrapper[4943]: I0307 15:10:42.873099 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-nl4hw"] Mar 07 15:10:42 crc kubenswrapper[4943]: I0307 15:10:42.884956 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-nl4hw"] Mar 07 15:10:42 crc kubenswrapper[4943]: I0307 15:10:42.963636 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d49deefc-a177-4d58-a075-a24ef44ebce1-ring-data-devices\") pod \"d49deefc-a177-4d58-a075-a24ef44ebce1\" (UID: \"d49deefc-a177-4d58-a075-a24ef44ebce1\") " Mar 07 15:10:42 crc kubenswrapper[4943]: I0307 15:10:42.963689 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d49deefc-a177-4d58-a075-a24ef44ebce1-swiftconf\") pod \"d49deefc-a177-4d58-a075-a24ef44ebce1\" (UID: \"d49deefc-a177-4d58-a075-a24ef44ebce1\") " Mar 07 15:10:42 crc kubenswrapper[4943]: I0307 15:10:42.963750 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d49deefc-a177-4d58-a075-a24ef44ebce1-dispersionconf\") pod \"d49deefc-a177-4d58-a075-a24ef44ebce1\" (UID: \"d49deefc-a177-4d58-a075-a24ef44ebce1\") " Mar 07 15:10:42 crc kubenswrapper[4943]: I0307 15:10:42.963861 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9kbvj\" (UniqueName: \"kubernetes.io/projected/d49deefc-a177-4d58-a075-a24ef44ebce1-kube-api-access-9kbvj\") pod \"d49deefc-a177-4d58-a075-a24ef44ebce1\" (UID: \"d49deefc-a177-4d58-a075-a24ef44ebce1\") " Mar 07 15:10:42 crc kubenswrapper[4943]: I0307 15:10:42.963901 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d49deefc-a177-4d58-a075-a24ef44ebce1-scripts\") pod \"d49deefc-a177-4d58-a075-a24ef44ebce1\" (UID: \"d49deefc-a177-4d58-a075-a24ef44ebce1\") " Mar 07 15:10:42 crc kubenswrapper[4943]: I0307 15:10:42.963962 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d49deefc-a177-4d58-a075-a24ef44ebce1-etc-swift\") pod \"d49deefc-a177-4d58-a075-a24ef44ebce1\" (UID: \"d49deefc-a177-4d58-a075-a24ef44ebce1\") " Mar 07 15:10:42 crc kubenswrapper[4943]: I0307 15:10:42.964888 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d49deefc-a177-4d58-a075-a24ef44ebce1-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "d49deefc-a177-4d58-a075-a24ef44ebce1" (UID: "d49deefc-a177-4d58-a075-a24ef44ebce1"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:10:42 crc kubenswrapper[4943]: I0307 15:10:42.965276 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d49deefc-a177-4d58-a075-a24ef44ebce1-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "d49deefc-a177-4d58-a075-a24ef44ebce1" (UID: "d49deefc-a177-4d58-a075-a24ef44ebce1"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:10:42 crc kubenswrapper[4943]: I0307 15:10:42.970911 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d49deefc-a177-4d58-a075-a24ef44ebce1-kube-api-access-9kbvj" (OuterVolumeSpecName: "kube-api-access-9kbvj") pod "d49deefc-a177-4d58-a075-a24ef44ebce1" (UID: "d49deefc-a177-4d58-a075-a24ef44ebce1"). InnerVolumeSpecName "kube-api-access-9kbvj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:10:42 crc kubenswrapper[4943]: I0307 15:10:42.989405 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d49deefc-a177-4d58-a075-a24ef44ebce1-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "d49deefc-a177-4d58-a075-a24ef44ebce1" (UID: "d49deefc-a177-4d58-a075-a24ef44ebce1"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:10:42 crc kubenswrapper[4943]: I0307 15:10:42.996182 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d49deefc-a177-4d58-a075-a24ef44ebce1-scripts" (OuterVolumeSpecName: "scripts") pod "d49deefc-a177-4d58-a075-a24ef44ebce1" (UID: "d49deefc-a177-4d58-a075-a24ef44ebce1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:10:43 crc kubenswrapper[4943]: I0307 15:10:43.001245 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d49deefc-a177-4d58-a075-a24ef44ebce1-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "d49deefc-a177-4d58-a075-a24ef44ebce1" (UID: "d49deefc-a177-4d58-a075-a24ef44ebce1"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:10:43 crc kubenswrapper[4943]: I0307 15:10:43.066825 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9kbvj\" (UniqueName: \"kubernetes.io/projected/d49deefc-a177-4d58-a075-a24ef44ebce1-kube-api-access-9kbvj\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:43 crc kubenswrapper[4943]: I0307 15:10:43.066885 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d49deefc-a177-4d58-a075-a24ef44ebce1-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:43 crc kubenswrapper[4943]: I0307 15:10:43.066905 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d49deefc-a177-4d58-a075-a24ef44ebce1-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:43 crc kubenswrapper[4943]: I0307 15:10:43.066923 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d49deefc-a177-4d58-a075-a24ef44ebce1-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:43 crc kubenswrapper[4943]: I0307 15:10:43.066973 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d49deefc-a177-4d58-a075-a24ef44ebce1-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:43 crc kubenswrapper[4943]: I0307 15:10:43.066990 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d49deefc-a177-4d58-a075-a24ef44ebce1-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:43 crc kubenswrapper[4943]: I0307 15:10:43.499568 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc072e3436ca706bb9623d316c7175bacc6679d0c69d6c8c3c70de1eb20b407f" Mar 07 15:10:43 crc kubenswrapper[4943]: I0307 15:10:43.499691 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-nl4hw" Mar 07 15:10:44 crc kubenswrapper[4943]: I0307 15:10:44.059777 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-c4x8k"] Mar 07 15:10:44 crc kubenswrapper[4943]: E0307 15:10:44.060475 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d49deefc-a177-4d58-a075-a24ef44ebce1" containerName="swift-ring-rebalance" Mar 07 15:10:44 crc kubenswrapper[4943]: I0307 15:10:44.060508 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="d49deefc-a177-4d58-a075-a24ef44ebce1" containerName="swift-ring-rebalance" Mar 07 15:10:44 crc kubenswrapper[4943]: I0307 15:10:44.060812 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="d49deefc-a177-4d58-a075-a24ef44ebce1" containerName="swift-ring-rebalance" Mar 07 15:10:44 crc kubenswrapper[4943]: I0307 15:10:44.061986 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-c4x8k" Mar 07 15:10:44 crc kubenswrapper[4943]: I0307 15:10:44.068345 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:10:44 crc kubenswrapper[4943]: I0307 15:10:44.068674 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:10:44 crc kubenswrapper[4943]: I0307 15:10:44.075665 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-c4x8k"] Mar 07 15:10:44 crc kubenswrapper[4943]: I0307 15:10:44.093812 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/68fa23df-ae3a-4145-aeb4-a41d1e160d63-scripts\") pod \"swift-ring-rebalance-debug-c4x8k\" (UID: \"68fa23df-ae3a-4145-aeb4-a41d1e160d63\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-c4x8k" Mar 07 15:10:44 crc kubenswrapper[4943]: I0307 15:10:44.093958 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/68fa23df-ae3a-4145-aeb4-a41d1e160d63-swiftconf\") pod \"swift-ring-rebalance-debug-c4x8k\" (UID: \"68fa23df-ae3a-4145-aeb4-a41d1e160d63\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-c4x8k" Mar 07 15:10:44 crc kubenswrapper[4943]: I0307 15:10:44.094076 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcj5g\" (UniqueName: \"kubernetes.io/projected/68fa23df-ae3a-4145-aeb4-a41d1e160d63-kube-api-access-mcj5g\") pod \"swift-ring-rebalance-debug-c4x8k\" (UID: \"68fa23df-ae3a-4145-aeb4-a41d1e160d63\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-c4x8k" Mar 07 15:10:44 crc kubenswrapper[4943]: I0307 15:10:44.094195 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/68fa23df-ae3a-4145-aeb4-a41d1e160d63-etc-swift\") pod \"swift-ring-rebalance-debug-c4x8k\" (UID: \"68fa23df-ae3a-4145-aeb4-a41d1e160d63\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-c4x8k" Mar 07 15:10:44 crc kubenswrapper[4943]: I0307 15:10:44.094266 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/68fa23df-ae3a-4145-aeb4-a41d1e160d63-ring-data-devices\") pod \"swift-ring-rebalance-debug-c4x8k\" (UID: \"68fa23df-ae3a-4145-aeb4-a41d1e160d63\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-c4x8k" Mar 07 15:10:44 crc kubenswrapper[4943]: I0307 15:10:44.094325 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/68fa23df-ae3a-4145-aeb4-a41d1e160d63-dispersionconf\") pod \"swift-ring-rebalance-debug-c4x8k\" (UID: \"68fa23df-ae3a-4145-aeb4-a41d1e160d63\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-c4x8k" Mar 07 15:10:44 crc kubenswrapper[4943]: I0307 15:10:44.196468 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/68fa23df-ae3a-4145-aeb4-a41d1e160d63-etc-swift\") pod \"swift-ring-rebalance-debug-c4x8k\" (UID: \"68fa23df-ae3a-4145-aeb4-a41d1e160d63\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-c4x8k" Mar 07 15:10:44 crc kubenswrapper[4943]: I0307 15:10:44.196538 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/68fa23df-ae3a-4145-aeb4-a41d1e160d63-ring-data-devices\") pod \"swift-ring-rebalance-debug-c4x8k\" (UID: \"68fa23df-ae3a-4145-aeb4-a41d1e160d63\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-c4x8k" Mar 07 15:10:44 crc kubenswrapper[4943]: I0307 15:10:44.196576 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/68fa23df-ae3a-4145-aeb4-a41d1e160d63-dispersionconf\") pod \"swift-ring-rebalance-debug-c4x8k\" (UID: \"68fa23df-ae3a-4145-aeb4-a41d1e160d63\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-c4x8k" Mar 07 15:10:44 crc kubenswrapper[4943]: I0307 15:10:44.196713 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/68fa23df-ae3a-4145-aeb4-a41d1e160d63-scripts\") pod \"swift-ring-rebalance-debug-c4x8k\" (UID: \"68fa23df-ae3a-4145-aeb4-a41d1e160d63\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-c4x8k" Mar 07 15:10:44 crc kubenswrapper[4943]: I0307 15:10:44.196746 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/68fa23df-ae3a-4145-aeb4-a41d1e160d63-swiftconf\") pod \"swift-ring-rebalance-debug-c4x8k\" (UID: \"68fa23df-ae3a-4145-aeb4-a41d1e160d63\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-c4x8k" Mar 07 15:10:44 crc kubenswrapper[4943]: I0307 15:10:44.196790 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcj5g\" (UniqueName: \"kubernetes.io/projected/68fa23df-ae3a-4145-aeb4-a41d1e160d63-kube-api-access-mcj5g\") pod \"swift-ring-rebalance-debug-c4x8k\" (UID: \"68fa23df-ae3a-4145-aeb4-a41d1e160d63\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-c4x8k" Mar 07 15:10:44 crc kubenswrapper[4943]: I0307 15:10:44.197295 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/68fa23df-ae3a-4145-aeb4-a41d1e160d63-etc-swift\") pod \"swift-ring-rebalance-debug-c4x8k\" (UID: \"68fa23df-ae3a-4145-aeb4-a41d1e160d63\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-c4x8k" Mar 07 15:10:44 crc kubenswrapper[4943]: I0307 15:10:44.197578 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/68fa23df-ae3a-4145-aeb4-a41d1e160d63-ring-data-devices\") pod \"swift-ring-rebalance-debug-c4x8k\" (UID: \"68fa23df-ae3a-4145-aeb4-a41d1e160d63\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-c4x8k" Mar 07 15:10:44 crc kubenswrapper[4943]: I0307 15:10:44.197684 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/68fa23df-ae3a-4145-aeb4-a41d1e160d63-scripts\") pod \"swift-ring-rebalance-debug-c4x8k\" (UID: \"68fa23df-ae3a-4145-aeb4-a41d1e160d63\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-c4x8k" Mar 07 15:10:44 crc kubenswrapper[4943]: I0307 15:10:44.203619 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/68fa23df-ae3a-4145-aeb4-a41d1e160d63-swiftconf\") pod \"swift-ring-rebalance-debug-c4x8k\" (UID: \"68fa23df-ae3a-4145-aeb4-a41d1e160d63\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-c4x8k" Mar 07 15:10:44 crc kubenswrapper[4943]: I0307 15:10:44.205150 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/68fa23df-ae3a-4145-aeb4-a41d1e160d63-dispersionconf\") pod \"swift-ring-rebalance-debug-c4x8k\" (UID: \"68fa23df-ae3a-4145-aeb4-a41d1e160d63\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-c4x8k" Mar 07 15:10:44 crc kubenswrapper[4943]: I0307 15:10:44.226313 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcj5g\" (UniqueName: \"kubernetes.io/projected/68fa23df-ae3a-4145-aeb4-a41d1e160d63-kube-api-access-mcj5g\") pod \"swift-ring-rebalance-debug-c4x8k\" (UID: \"68fa23df-ae3a-4145-aeb4-a41d1e160d63\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-c4x8k" Mar 07 15:10:44 crc kubenswrapper[4943]: I0307 15:10:44.398536 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-c4x8k" Mar 07 15:10:44 crc kubenswrapper[4943]: I0307 15:10:44.698760 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-c4x8k"] Mar 07 15:10:44 crc kubenswrapper[4943]: I0307 15:10:44.765680 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d49deefc-a177-4d58-a075-a24ef44ebce1" path="/var/lib/kubelet/pods/d49deefc-a177-4d58-a075-a24ef44ebce1/volumes" Mar 07 15:10:45 crc kubenswrapper[4943]: I0307 15:10:45.522324 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-c4x8k" event={"ID":"68fa23df-ae3a-4145-aeb4-a41d1e160d63","Type":"ContainerStarted","Data":"1581ae5bf3af5855b7ab2b0c264ff111dca284eedfa2e49a4cf21a6b80de367a"} Mar 07 15:10:45 crc kubenswrapper[4943]: I0307 15:10:45.522611 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-c4x8k" event={"ID":"68fa23df-ae3a-4145-aeb4-a41d1e160d63","Type":"ContainerStarted","Data":"0439f1265383e1f7551b7c250cdd58c07c97ecbe12ba09561e131c1fcf234f03"} Mar 07 15:10:45 crc kubenswrapper[4943]: I0307 15:10:45.548579 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-c4x8k" podStartSLOduration=1.548496817 podStartE2EDuration="1.548496817s" podCreationTimestamp="2026-03-07 15:10:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:10:45.544280003 +0000 UTC m=+1887.496416551" watchObservedRunningTime="2026-03-07 15:10:45.548496817 +0000 UTC m=+1887.500633345" Mar 07 15:10:46 crc kubenswrapper[4943]: I0307 15:10:46.535921 4943 generic.go:334] "Generic (PLEG): container finished" podID="68fa23df-ae3a-4145-aeb4-a41d1e160d63" containerID="1581ae5bf3af5855b7ab2b0c264ff111dca284eedfa2e49a4cf21a6b80de367a" exitCode=0 Mar 07 15:10:46 crc kubenswrapper[4943]: I0307 15:10:46.536085 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-c4x8k" event={"ID":"68fa23df-ae3a-4145-aeb4-a41d1e160d63","Type":"ContainerDied","Data":"1581ae5bf3af5855b7ab2b0c264ff111dca284eedfa2e49a4cf21a6b80de367a"} Mar 07 15:10:47 crc kubenswrapper[4943]: I0307 15:10:47.894195 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-c4x8k" Mar 07 15:10:47 crc kubenswrapper[4943]: I0307 15:10:47.946056 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-c4x8k"] Mar 07 15:10:47 crc kubenswrapper[4943]: I0307 15:10:47.952836 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-c4x8k"] Mar 07 15:10:47 crc kubenswrapper[4943]: I0307 15:10:47.956146 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/68fa23df-ae3a-4145-aeb4-a41d1e160d63-scripts\") pod \"68fa23df-ae3a-4145-aeb4-a41d1e160d63\" (UID: \"68fa23df-ae3a-4145-aeb4-a41d1e160d63\") " Mar 07 15:10:47 crc kubenswrapper[4943]: I0307 15:10:47.956187 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/68fa23df-ae3a-4145-aeb4-a41d1e160d63-swiftconf\") pod \"68fa23df-ae3a-4145-aeb4-a41d1e160d63\" (UID: \"68fa23df-ae3a-4145-aeb4-a41d1e160d63\") " Mar 07 15:10:47 crc kubenswrapper[4943]: I0307 15:10:47.956320 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/68fa23df-ae3a-4145-aeb4-a41d1e160d63-etc-swift\") pod \"68fa23df-ae3a-4145-aeb4-a41d1e160d63\" (UID: \"68fa23df-ae3a-4145-aeb4-a41d1e160d63\") " Mar 07 15:10:47 crc kubenswrapper[4943]: I0307 15:10:47.956378 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/68fa23df-ae3a-4145-aeb4-a41d1e160d63-dispersionconf\") pod \"68fa23df-ae3a-4145-aeb4-a41d1e160d63\" (UID: \"68fa23df-ae3a-4145-aeb4-a41d1e160d63\") " Mar 07 15:10:47 crc kubenswrapper[4943]: I0307 15:10:47.956454 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mcj5g\" (UniqueName: \"kubernetes.io/projected/68fa23df-ae3a-4145-aeb4-a41d1e160d63-kube-api-access-mcj5g\") pod \"68fa23df-ae3a-4145-aeb4-a41d1e160d63\" (UID: \"68fa23df-ae3a-4145-aeb4-a41d1e160d63\") " Mar 07 15:10:47 crc kubenswrapper[4943]: I0307 15:10:47.956523 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/68fa23df-ae3a-4145-aeb4-a41d1e160d63-ring-data-devices\") pod \"68fa23df-ae3a-4145-aeb4-a41d1e160d63\" (UID: \"68fa23df-ae3a-4145-aeb4-a41d1e160d63\") " Mar 07 15:10:47 crc kubenswrapper[4943]: I0307 15:10:47.958037 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68fa23df-ae3a-4145-aeb4-a41d1e160d63-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "68fa23df-ae3a-4145-aeb4-a41d1e160d63" (UID: "68fa23df-ae3a-4145-aeb4-a41d1e160d63"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:10:47 crc kubenswrapper[4943]: I0307 15:10:47.958042 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68fa23df-ae3a-4145-aeb4-a41d1e160d63-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "68fa23df-ae3a-4145-aeb4-a41d1e160d63" (UID: "68fa23df-ae3a-4145-aeb4-a41d1e160d63"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:10:47 crc kubenswrapper[4943]: I0307 15:10:47.963003 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68fa23df-ae3a-4145-aeb4-a41d1e160d63-kube-api-access-mcj5g" (OuterVolumeSpecName: "kube-api-access-mcj5g") pod "68fa23df-ae3a-4145-aeb4-a41d1e160d63" (UID: "68fa23df-ae3a-4145-aeb4-a41d1e160d63"). InnerVolumeSpecName "kube-api-access-mcj5g". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:10:47 crc kubenswrapper[4943]: I0307 15:10:47.981670 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68fa23df-ae3a-4145-aeb4-a41d1e160d63-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "68fa23df-ae3a-4145-aeb4-a41d1e160d63" (UID: "68fa23df-ae3a-4145-aeb4-a41d1e160d63"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:10:47 crc kubenswrapper[4943]: I0307 15:10:47.982504 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68fa23df-ae3a-4145-aeb4-a41d1e160d63-scripts" (OuterVolumeSpecName: "scripts") pod "68fa23df-ae3a-4145-aeb4-a41d1e160d63" (UID: "68fa23df-ae3a-4145-aeb4-a41d1e160d63"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:10:47 crc kubenswrapper[4943]: I0307 15:10:47.988977 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68fa23df-ae3a-4145-aeb4-a41d1e160d63-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "68fa23df-ae3a-4145-aeb4-a41d1e160d63" (UID: "68fa23df-ae3a-4145-aeb4-a41d1e160d63"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:10:48 crc kubenswrapper[4943]: I0307 15:10:48.059159 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mcj5g\" (UniqueName: \"kubernetes.io/projected/68fa23df-ae3a-4145-aeb4-a41d1e160d63-kube-api-access-mcj5g\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:48 crc kubenswrapper[4943]: I0307 15:10:48.059492 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/68fa23df-ae3a-4145-aeb4-a41d1e160d63-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:48 crc kubenswrapper[4943]: I0307 15:10:48.059514 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/68fa23df-ae3a-4145-aeb4-a41d1e160d63-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:48 crc kubenswrapper[4943]: I0307 15:10:48.059532 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/68fa23df-ae3a-4145-aeb4-a41d1e160d63-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:48 crc kubenswrapper[4943]: I0307 15:10:48.059552 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/68fa23df-ae3a-4145-aeb4-a41d1e160d63-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:48 crc kubenswrapper[4943]: I0307 15:10:48.059569 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/68fa23df-ae3a-4145-aeb4-a41d1e160d63-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:48 crc kubenswrapper[4943]: I0307 15:10:48.572556 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0439f1265383e1f7551b7c250cdd58c07c97ecbe12ba09561e131c1fcf234f03" Mar 07 15:10:48 crc kubenswrapper[4943]: I0307 15:10:48.572651 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-c4x8k" Mar 07 15:10:48 crc kubenswrapper[4943]: I0307 15:10:48.776788 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68fa23df-ae3a-4145-aeb4-a41d1e160d63" path="/var/lib/kubelet/pods/68fa23df-ae3a-4145-aeb4-a41d1e160d63/volumes" Mar 07 15:10:49 crc kubenswrapper[4943]: I0307 15:10:49.173491 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-6n8pt"] Mar 07 15:10:49 crc kubenswrapper[4943]: E0307 15:10:49.173952 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68fa23df-ae3a-4145-aeb4-a41d1e160d63" containerName="swift-ring-rebalance" Mar 07 15:10:49 crc kubenswrapper[4943]: I0307 15:10:49.173975 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="68fa23df-ae3a-4145-aeb4-a41d1e160d63" containerName="swift-ring-rebalance" Mar 07 15:10:49 crc kubenswrapper[4943]: I0307 15:10:49.174240 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="68fa23df-ae3a-4145-aeb4-a41d1e160d63" containerName="swift-ring-rebalance" Mar 07 15:10:49 crc kubenswrapper[4943]: I0307 15:10:49.175035 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6n8pt" Mar 07 15:10:49 crc kubenswrapper[4943]: I0307 15:10:49.178603 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:10:49 crc kubenswrapper[4943]: I0307 15:10:49.179304 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:10:49 crc kubenswrapper[4943]: I0307 15:10:49.179480 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-6n8pt"] Mar 07 15:10:49 crc kubenswrapper[4943]: I0307 15:10:49.281103 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/0b4f0306-c5bc-4bf8-9399-4edabd0ab411-etc-swift\") pod \"swift-ring-rebalance-debug-6n8pt\" (UID: \"0b4f0306-c5bc-4bf8-9399-4edabd0ab411\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6n8pt" Mar 07 15:10:49 crc kubenswrapper[4943]: I0307 15:10:49.281163 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/0b4f0306-c5bc-4bf8-9399-4edabd0ab411-ring-data-devices\") pod \"swift-ring-rebalance-debug-6n8pt\" (UID: \"0b4f0306-c5bc-4bf8-9399-4edabd0ab411\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6n8pt" Mar 07 15:10:49 crc kubenswrapper[4943]: I0307 15:10:49.281211 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/0b4f0306-c5bc-4bf8-9399-4edabd0ab411-swiftconf\") pod \"swift-ring-rebalance-debug-6n8pt\" (UID: \"0b4f0306-c5bc-4bf8-9399-4edabd0ab411\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6n8pt" Mar 07 15:10:49 crc kubenswrapper[4943]: I0307 15:10:49.281386 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0b4f0306-c5bc-4bf8-9399-4edabd0ab411-scripts\") pod \"swift-ring-rebalance-debug-6n8pt\" (UID: \"0b4f0306-c5bc-4bf8-9399-4edabd0ab411\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6n8pt" Mar 07 15:10:49 crc kubenswrapper[4943]: I0307 15:10:49.281628 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4z9xg\" (UniqueName: \"kubernetes.io/projected/0b4f0306-c5bc-4bf8-9399-4edabd0ab411-kube-api-access-4z9xg\") pod \"swift-ring-rebalance-debug-6n8pt\" (UID: \"0b4f0306-c5bc-4bf8-9399-4edabd0ab411\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6n8pt" Mar 07 15:10:49 crc kubenswrapper[4943]: I0307 15:10:49.281796 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/0b4f0306-c5bc-4bf8-9399-4edabd0ab411-dispersionconf\") pod \"swift-ring-rebalance-debug-6n8pt\" (UID: \"0b4f0306-c5bc-4bf8-9399-4edabd0ab411\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6n8pt" Mar 07 15:10:49 crc kubenswrapper[4943]: I0307 15:10:49.383706 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4z9xg\" (UniqueName: \"kubernetes.io/projected/0b4f0306-c5bc-4bf8-9399-4edabd0ab411-kube-api-access-4z9xg\") pod \"swift-ring-rebalance-debug-6n8pt\" (UID: \"0b4f0306-c5bc-4bf8-9399-4edabd0ab411\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6n8pt" Mar 07 15:10:49 crc kubenswrapper[4943]: I0307 15:10:49.383832 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/0b4f0306-c5bc-4bf8-9399-4edabd0ab411-dispersionconf\") pod \"swift-ring-rebalance-debug-6n8pt\" (UID: \"0b4f0306-c5bc-4bf8-9399-4edabd0ab411\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6n8pt" Mar 07 15:10:49 crc kubenswrapper[4943]: I0307 15:10:49.383906 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/0b4f0306-c5bc-4bf8-9399-4edabd0ab411-etc-swift\") pod \"swift-ring-rebalance-debug-6n8pt\" (UID: \"0b4f0306-c5bc-4bf8-9399-4edabd0ab411\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6n8pt" Mar 07 15:10:49 crc kubenswrapper[4943]: I0307 15:10:49.383964 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/0b4f0306-c5bc-4bf8-9399-4edabd0ab411-ring-data-devices\") pod \"swift-ring-rebalance-debug-6n8pt\" (UID: \"0b4f0306-c5bc-4bf8-9399-4edabd0ab411\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6n8pt" Mar 07 15:10:49 crc kubenswrapper[4943]: I0307 15:10:49.384006 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/0b4f0306-c5bc-4bf8-9399-4edabd0ab411-swiftconf\") pod \"swift-ring-rebalance-debug-6n8pt\" (UID: \"0b4f0306-c5bc-4bf8-9399-4edabd0ab411\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6n8pt" Mar 07 15:10:49 crc kubenswrapper[4943]: I0307 15:10:49.384131 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0b4f0306-c5bc-4bf8-9399-4edabd0ab411-scripts\") pod \"swift-ring-rebalance-debug-6n8pt\" (UID: \"0b4f0306-c5bc-4bf8-9399-4edabd0ab411\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6n8pt" Mar 07 15:10:49 crc kubenswrapper[4943]: I0307 15:10:49.384738 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/0b4f0306-c5bc-4bf8-9399-4edabd0ab411-etc-swift\") pod \"swift-ring-rebalance-debug-6n8pt\" (UID: \"0b4f0306-c5bc-4bf8-9399-4edabd0ab411\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6n8pt" Mar 07 15:10:49 crc kubenswrapper[4943]: I0307 15:10:49.385417 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/0b4f0306-c5bc-4bf8-9399-4edabd0ab411-ring-data-devices\") pod \"swift-ring-rebalance-debug-6n8pt\" (UID: \"0b4f0306-c5bc-4bf8-9399-4edabd0ab411\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6n8pt" Mar 07 15:10:49 crc kubenswrapper[4943]: I0307 15:10:49.385490 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0b4f0306-c5bc-4bf8-9399-4edabd0ab411-scripts\") pod \"swift-ring-rebalance-debug-6n8pt\" (UID: \"0b4f0306-c5bc-4bf8-9399-4edabd0ab411\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6n8pt" Mar 07 15:10:49 crc kubenswrapper[4943]: I0307 15:10:49.393417 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/0b4f0306-c5bc-4bf8-9399-4edabd0ab411-swiftconf\") pod \"swift-ring-rebalance-debug-6n8pt\" (UID: \"0b4f0306-c5bc-4bf8-9399-4edabd0ab411\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6n8pt" Mar 07 15:10:49 crc kubenswrapper[4943]: I0307 15:10:49.393605 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/0b4f0306-c5bc-4bf8-9399-4edabd0ab411-dispersionconf\") pod \"swift-ring-rebalance-debug-6n8pt\" (UID: \"0b4f0306-c5bc-4bf8-9399-4edabd0ab411\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6n8pt" Mar 07 15:10:49 crc kubenswrapper[4943]: I0307 15:10:49.414818 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4z9xg\" (UniqueName: \"kubernetes.io/projected/0b4f0306-c5bc-4bf8-9399-4edabd0ab411-kube-api-access-4z9xg\") pod \"swift-ring-rebalance-debug-6n8pt\" (UID: \"0b4f0306-c5bc-4bf8-9399-4edabd0ab411\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6n8pt" Mar 07 15:10:49 crc kubenswrapper[4943]: I0307 15:10:49.517631 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6n8pt" Mar 07 15:10:49 crc kubenswrapper[4943]: I0307 15:10:49.867665 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-6n8pt"] Mar 07 15:10:50 crc kubenswrapper[4943]: I0307 15:10:50.593672 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6n8pt" event={"ID":"0b4f0306-c5bc-4bf8-9399-4edabd0ab411","Type":"ContainerStarted","Data":"7df5bfaa2e10322b574e07b3024a8090cf867a6ba151ab94ed79f3a1da20d2d4"} Mar 07 15:10:50 crc kubenswrapper[4943]: I0307 15:10:50.593954 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6n8pt" event={"ID":"0b4f0306-c5bc-4bf8-9399-4edabd0ab411","Type":"ContainerStarted","Data":"437ff7e14983b56303398be68f4936360ffe056d10c299fd945928551523b255"} Mar 07 15:10:50 crc kubenswrapper[4943]: I0307 15:10:50.625819 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6n8pt" podStartSLOduration=1.62579801 podStartE2EDuration="1.62579801s" podCreationTimestamp="2026-03-07 15:10:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:10:50.620535631 +0000 UTC m=+1892.572672159" watchObservedRunningTime="2026-03-07 15:10:50.62579801 +0000 UTC m=+1892.577934508" Mar 07 15:10:51 crc kubenswrapper[4943]: I0307 15:10:51.607450 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6n8pt" event={"ID":"0b4f0306-c5bc-4bf8-9399-4edabd0ab411","Type":"ContainerDied","Data":"7df5bfaa2e10322b574e07b3024a8090cf867a6ba151ab94ed79f3a1da20d2d4"} Mar 07 15:10:51 crc kubenswrapper[4943]: I0307 15:10:51.608534 4943 generic.go:334] "Generic (PLEG): container finished" podID="0b4f0306-c5bc-4bf8-9399-4edabd0ab411" containerID="7df5bfaa2e10322b574e07b3024a8090cf867a6ba151ab94ed79f3a1da20d2d4" exitCode=0 Mar 07 15:10:52 crc kubenswrapper[4943]: I0307 15:10:52.979228 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6n8pt" Mar 07 15:10:53 crc kubenswrapper[4943]: I0307 15:10:53.009621 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-6n8pt"] Mar 07 15:10:53 crc kubenswrapper[4943]: I0307 15:10:53.016888 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-6n8pt"] Mar 07 15:10:53 crc kubenswrapper[4943]: I0307 15:10:53.143949 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/0b4f0306-c5bc-4bf8-9399-4edabd0ab411-swiftconf\") pod \"0b4f0306-c5bc-4bf8-9399-4edabd0ab411\" (UID: \"0b4f0306-c5bc-4bf8-9399-4edabd0ab411\") " Mar 07 15:10:53 crc kubenswrapper[4943]: I0307 15:10:53.144092 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4z9xg\" (UniqueName: \"kubernetes.io/projected/0b4f0306-c5bc-4bf8-9399-4edabd0ab411-kube-api-access-4z9xg\") pod \"0b4f0306-c5bc-4bf8-9399-4edabd0ab411\" (UID: \"0b4f0306-c5bc-4bf8-9399-4edabd0ab411\") " Mar 07 15:10:53 crc kubenswrapper[4943]: I0307 15:10:53.144158 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/0b4f0306-c5bc-4bf8-9399-4edabd0ab411-ring-data-devices\") pod \"0b4f0306-c5bc-4bf8-9399-4edabd0ab411\" (UID: \"0b4f0306-c5bc-4bf8-9399-4edabd0ab411\") " Mar 07 15:10:53 crc kubenswrapper[4943]: I0307 15:10:53.144188 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/0b4f0306-c5bc-4bf8-9399-4edabd0ab411-dispersionconf\") pod \"0b4f0306-c5bc-4bf8-9399-4edabd0ab411\" (UID: \"0b4f0306-c5bc-4bf8-9399-4edabd0ab411\") " Mar 07 15:10:53 crc kubenswrapper[4943]: I0307 15:10:53.144223 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/0b4f0306-c5bc-4bf8-9399-4edabd0ab411-etc-swift\") pod \"0b4f0306-c5bc-4bf8-9399-4edabd0ab411\" (UID: \"0b4f0306-c5bc-4bf8-9399-4edabd0ab411\") " Mar 07 15:10:53 crc kubenswrapper[4943]: I0307 15:10:53.144299 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0b4f0306-c5bc-4bf8-9399-4edabd0ab411-scripts\") pod \"0b4f0306-c5bc-4bf8-9399-4edabd0ab411\" (UID: \"0b4f0306-c5bc-4bf8-9399-4edabd0ab411\") " Mar 07 15:10:53 crc kubenswrapper[4943]: I0307 15:10:53.145459 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b4f0306-c5bc-4bf8-9399-4edabd0ab411-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "0b4f0306-c5bc-4bf8-9399-4edabd0ab411" (UID: "0b4f0306-c5bc-4bf8-9399-4edabd0ab411"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:10:53 crc kubenswrapper[4943]: I0307 15:10:53.146385 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b4f0306-c5bc-4bf8-9399-4edabd0ab411-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "0b4f0306-c5bc-4bf8-9399-4edabd0ab411" (UID: "0b4f0306-c5bc-4bf8-9399-4edabd0ab411"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:10:53 crc kubenswrapper[4943]: I0307 15:10:53.153145 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b4f0306-c5bc-4bf8-9399-4edabd0ab411-kube-api-access-4z9xg" (OuterVolumeSpecName: "kube-api-access-4z9xg") pod "0b4f0306-c5bc-4bf8-9399-4edabd0ab411" (UID: "0b4f0306-c5bc-4bf8-9399-4edabd0ab411"). InnerVolumeSpecName "kube-api-access-4z9xg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:10:53 crc kubenswrapper[4943]: I0307 15:10:53.186454 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b4f0306-c5bc-4bf8-9399-4edabd0ab411-scripts" (OuterVolumeSpecName: "scripts") pod "0b4f0306-c5bc-4bf8-9399-4edabd0ab411" (UID: "0b4f0306-c5bc-4bf8-9399-4edabd0ab411"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:10:53 crc kubenswrapper[4943]: I0307 15:10:53.188279 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b4f0306-c5bc-4bf8-9399-4edabd0ab411-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "0b4f0306-c5bc-4bf8-9399-4edabd0ab411" (UID: "0b4f0306-c5bc-4bf8-9399-4edabd0ab411"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:10:53 crc kubenswrapper[4943]: I0307 15:10:53.198588 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b4f0306-c5bc-4bf8-9399-4edabd0ab411-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "0b4f0306-c5bc-4bf8-9399-4edabd0ab411" (UID: "0b4f0306-c5bc-4bf8-9399-4edabd0ab411"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:10:53 crc kubenswrapper[4943]: I0307 15:10:53.246316 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/0b4f0306-c5bc-4bf8-9399-4edabd0ab411-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:53 crc kubenswrapper[4943]: I0307 15:10:53.246356 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4z9xg\" (UniqueName: \"kubernetes.io/projected/0b4f0306-c5bc-4bf8-9399-4edabd0ab411-kube-api-access-4z9xg\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:53 crc kubenswrapper[4943]: I0307 15:10:53.246376 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/0b4f0306-c5bc-4bf8-9399-4edabd0ab411-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:53 crc kubenswrapper[4943]: I0307 15:10:53.246397 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/0b4f0306-c5bc-4bf8-9399-4edabd0ab411-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:53 crc kubenswrapper[4943]: I0307 15:10:53.246417 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/0b4f0306-c5bc-4bf8-9399-4edabd0ab411-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:53 crc kubenswrapper[4943]: I0307 15:10:53.246436 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0b4f0306-c5bc-4bf8-9399-4edabd0ab411-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:53 crc kubenswrapper[4943]: I0307 15:10:53.632022 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="437ff7e14983b56303398be68f4936360ffe056d10c299fd945928551523b255" Mar 07 15:10:53 crc kubenswrapper[4943]: I0307 15:10:53.632162 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6n8pt" Mar 07 15:10:54 crc kubenswrapper[4943]: I0307 15:10:54.168034 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-hqb5w"] Mar 07 15:10:54 crc kubenswrapper[4943]: E0307 15:10:54.168495 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b4f0306-c5bc-4bf8-9399-4edabd0ab411" containerName="swift-ring-rebalance" Mar 07 15:10:54 crc kubenswrapper[4943]: I0307 15:10:54.168518 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b4f0306-c5bc-4bf8-9399-4edabd0ab411" containerName="swift-ring-rebalance" Mar 07 15:10:54 crc kubenswrapper[4943]: I0307 15:10:54.168794 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b4f0306-c5bc-4bf8-9399-4edabd0ab411" containerName="swift-ring-rebalance" Mar 07 15:10:54 crc kubenswrapper[4943]: I0307 15:10:54.170475 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-hqb5w" Mar 07 15:10:54 crc kubenswrapper[4943]: I0307 15:10:54.172563 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:10:54 crc kubenswrapper[4943]: I0307 15:10:54.172907 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:10:54 crc kubenswrapper[4943]: I0307 15:10:54.181464 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-hqb5w"] Mar 07 15:10:54 crc kubenswrapper[4943]: I0307 15:10:54.263595 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/db1ce9d8-45b0-4485-981b-ca88ec0c9dba-ring-data-devices\") pod \"swift-ring-rebalance-debug-hqb5w\" (UID: \"db1ce9d8-45b0-4485-981b-ca88ec0c9dba\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hqb5w" Mar 07 15:10:54 crc kubenswrapper[4943]: I0307 15:10:54.263711 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mb64p\" (UniqueName: \"kubernetes.io/projected/db1ce9d8-45b0-4485-981b-ca88ec0c9dba-kube-api-access-mb64p\") pod \"swift-ring-rebalance-debug-hqb5w\" (UID: \"db1ce9d8-45b0-4485-981b-ca88ec0c9dba\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hqb5w" Mar 07 15:10:54 crc kubenswrapper[4943]: I0307 15:10:54.263751 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/db1ce9d8-45b0-4485-981b-ca88ec0c9dba-dispersionconf\") pod \"swift-ring-rebalance-debug-hqb5w\" (UID: \"db1ce9d8-45b0-4485-981b-ca88ec0c9dba\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hqb5w" Mar 07 15:10:54 crc kubenswrapper[4943]: I0307 15:10:54.263786 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/db1ce9d8-45b0-4485-981b-ca88ec0c9dba-etc-swift\") pod \"swift-ring-rebalance-debug-hqb5w\" (UID: \"db1ce9d8-45b0-4485-981b-ca88ec0c9dba\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hqb5w" Mar 07 15:10:54 crc kubenswrapper[4943]: I0307 15:10:54.263897 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/db1ce9d8-45b0-4485-981b-ca88ec0c9dba-swiftconf\") pod \"swift-ring-rebalance-debug-hqb5w\" (UID: \"db1ce9d8-45b0-4485-981b-ca88ec0c9dba\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hqb5w" Mar 07 15:10:54 crc kubenswrapper[4943]: I0307 15:10:54.263983 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/db1ce9d8-45b0-4485-981b-ca88ec0c9dba-scripts\") pod \"swift-ring-rebalance-debug-hqb5w\" (UID: \"db1ce9d8-45b0-4485-981b-ca88ec0c9dba\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hqb5w" Mar 07 15:10:54 crc kubenswrapper[4943]: I0307 15:10:54.366157 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/db1ce9d8-45b0-4485-981b-ca88ec0c9dba-ring-data-devices\") pod \"swift-ring-rebalance-debug-hqb5w\" (UID: \"db1ce9d8-45b0-4485-981b-ca88ec0c9dba\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hqb5w" Mar 07 15:10:54 crc kubenswrapper[4943]: I0307 15:10:54.366252 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mb64p\" (UniqueName: \"kubernetes.io/projected/db1ce9d8-45b0-4485-981b-ca88ec0c9dba-kube-api-access-mb64p\") pod \"swift-ring-rebalance-debug-hqb5w\" (UID: \"db1ce9d8-45b0-4485-981b-ca88ec0c9dba\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hqb5w" Mar 07 15:10:54 crc kubenswrapper[4943]: I0307 15:10:54.366295 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/db1ce9d8-45b0-4485-981b-ca88ec0c9dba-dispersionconf\") pod \"swift-ring-rebalance-debug-hqb5w\" (UID: \"db1ce9d8-45b0-4485-981b-ca88ec0c9dba\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hqb5w" Mar 07 15:10:54 crc kubenswrapper[4943]: I0307 15:10:54.366333 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/db1ce9d8-45b0-4485-981b-ca88ec0c9dba-etc-swift\") pod \"swift-ring-rebalance-debug-hqb5w\" (UID: \"db1ce9d8-45b0-4485-981b-ca88ec0c9dba\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hqb5w" Mar 07 15:10:54 crc kubenswrapper[4943]: I0307 15:10:54.366465 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/db1ce9d8-45b0-4485-981b-ca88ec0c9dba-swiftconf\") pod \"swift-ring-rebalance-debug-hqb5w\" (UID: \"db1ce9d8-45b0-4485-981b-ca88ec0c9dba\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hqb5w" Mar 07 15:10:54 crc kubenswrapper[4943]: I0307 15:10:54.366507 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/db1ce9d8-45b0-4485-981b-ca88ec0c9dba-scripts\") pod \"swift-ring-rebalance-debug-hqb5w\" (UID: \"db1ce9d8-45b0-4485-981b-ca88ec0c9dba\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hqb5w" Mar 07 15:10:54 crc kubenswrapper[4943]: I0307 15:10:54.367107 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/db1ce9d8-45b0-4485-981b-ca88ec0c9dba-etc-swift\") pod \"swift-ring-rebalance-debug-hqb5w\" (UID: \"db1ce9d8-45b0-4485-981b-ca88ec0c9dba\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hqb5w" Mar 07 15:10:54 crc kubenswrapper[4943]: I0307 15:10:54.367702 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/db1ce9d8-45b0-4485-981b-ca88ec0c9dba-ring-data-devices\") pod \"swift-ring-rebalance-debug-hqb5w\" (UID: \"db1ce9d8-45b0-4485-981b-ca88ec0c9dba\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hqb5w" Mar 07 15:10:54 crc kubenswrapper[4943]: I0307 15:10:54.367747 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/db1ce9d8-45b0-4485-981b-ca88ec0c9dba-scripts\") pod \"swift-ring-rebalance-debug-hqb5w\" (UID: \"db1ce9d8-45b0-4485-981b-ca88ec0c9dba\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hqb5w" Mar 07 15:10:54 crc kubenswrapper[4943]: I0307 15:10:54.371337 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/db1ce9d8-45b0-4485-981b-ca88ec0c9dba-dispersionconf\") pod \"swift-ring-rebalance-debug-hqb5w\" (UID: \"db1ce9d8-45b0-4485-981b-ca88ec0c9dba\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hqb5w" Mar 07 15:10:54 crc kubenswrapper[4943]: I0307 15:10:54.372183 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/db1ce9d8-45b0-4485-981b-ca88ec0c9dba-swiftconf\") pod \"swift-ring-rebalance-debug-hqb5w\" (UID: \"db1ce9d8-45b0-4485-981b-ca88ec0c9dba\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hqb5w" Mar 07 15:10:54 crc kubenswrapper[4943]: I0307 15:10:54.395644 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mb64p\" (UniqueName: \"kubernetes.io/projected/db1ce9d8-45b0-4485-981b-ca88ec0c9dba-kube-api-access-mb64p\") pod \"swift-ring-rebalance-debug-hqb5w\" (UID: \"db1ce9d8-45b0-4485-981b-ca88ec0c9dba\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hqb5w" Mar 07 15:10:54 crc kubenswrapper[4943]: I0307 15:10:54.552321 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-hqb5w" Mar 07 15:10:54 crc kubenswrapper[4943]: I0307 15:10:54.777052 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b4f0306-c5bc-4bf8-9399-4edabd0ab411" path="/var/lib/kubelet/pods/0b4f0306-c5bc-4bf8-9399-4edabd0ab411/volumes" Mar 07 15:10:54 crc kubenswrapper[4943]: I0307 15:10:54.824773 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-hqb5w"] Mar 07 15:10:54 crc kubenswrapper[4943]: W0307 15:10:54.871035 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddb1ce9d8_45b0_4485_981b_ca88ec0c9dba.slice/crio-bbe1a697072029460b2310dbb9c41a8377d206483afe404bdf4676d21c3b8009 WatchSource:0}: Error finding container bbe1a697072029460b2310dbb9c41a8377d206483afe404bdf4676d21c3b8009: Status 404 returned error can't find the container with id bbe1a697072029460b2310dbb9c41a8377d206483afe404bdf4676d21c3b8009 Mar 07 15:10:55 crc kubenswrapper[4943]: I0307 15:10:55.652169 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-hqb5w" event={"ID":"db1ce9d8-45b0-4485-981b-ca88ec0c9dba","Type":"ContainerStarted","Data":"c2616fec6a6b33d05df21ca67ab7a148ccc3c6558a0b33d4d9e5af6e21ec9ec0"} Mar 07 15:10:55 crc kubenswrapper[4943]: I0307 15:10:55.652238 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-hqb5w" event={"ID":"db1ce9d8-45b0-4485-981b-ca88ec0c9dba","Type":"ContainerStarted","Data":"bbe1a697072029460b2310dbb9c41a8377d206483afe404bdf4676d21c3b8009"} Mar 07 15:10:55 crc kubenswrapper[4943]: I0307 15:10:55.680818 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-hqb5w" podStartSLOduration=1.6808026150000002 podStartE2EDuration="1.680802615s" podCreationTimestamp="2026-03-07 15:10:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:10:55.676007648 +0000 UTC m=+1897.628144186" watchObservedRunningTime="2026-03-07 15:10:55.680802615 +0000 UTC m=+1897.632939113" Mar 07 15:10:56 crc kubenswrapper[4943]: E0307 15:10:56.403099 4943 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddb1ce9d8_45b0_4485_981b_ca88ec0c9dba.slice/crio-c2616fec6a6b33d05df21ca67ab7a148ccc3c6558a0b33d4d9e5af6e21ec9ec0.scope\": RecentStats: unable to find data in memory cache]" Mar 07 15:10:56 crc kubenswrapper[4943]: I0307 15:10:56.662605 4943 generic.go:334] "Generic (PLEG): container finished" podID="db1ce9d8-45b0-4485-981b-ca88ec0c9dba" containerID="c2616fec6a6b33d05df21ca67ab7a148ccc3c6558a0b33d4d9e5af6e21ec9ec0" exitCode=0 Mar 07 15:10:56 crc kubenswrapper[4943]: I0307 15:10:56.662651 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-hqb5w" event={"ID":"db1ce9d8-45b0-4485-981b-ca88ec0c9dba","Type":"ContainerDied","Data":"c2616fec6a6b33d05df21ca67ab7a148ccc3c6558a0b33d4d9e5af6e21ec9ec0"} Mar 07 15:10:58 crc kubenswrapper[4943]: I0307 15:10:58.034500 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-hqb5w" Mar 07 15:10:58 crc kubenswrapper[4943]: I0307 15:10:58.081345 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-hqb5w"] Mar 07 15:10:58 crc kubenswrapper[4943]: I0307 15:10:58.087392 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-hqb5w"] Mar 07 15:10:58 crc kubenswrapper[4943]: I0307 15:10:58.227200 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/db1ce9d8-45b0-4485-981b-ca88ec0c9dba-ring-data-devices\") pod \"db1ce9d8-45b0-4485-981b-ca88ec0c9dba\" (UID: \"db1ce9d8-45b0-4485-981b-ca88ec0c9dba\") " Mar 07 15:10:58 crc kubenswrapper[4943]: I0307 15:10:58.227332 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/db1ce9d8-45b0-4485-981b-ca88ec0c9dba-dispersionconf\") pod \"db1ce9d8-45b0-4485-981b-ca88ec0c9dba\" (UID: \"db1ce9d8-45b0-4485-981b-ca88ec0c9dba\") " Mar 07 15:10:58 crc kubenswrapper[4943]: I0307 15:10:58.227395 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/db1ce9d8-45b0-4485-981b-ca88ec0c9dba-swiftconf\") pod \"db1ce9d8-45b0-4485-981b-ca88ec0c9dba\" (UID: \"db1ce9d8-45b0-4485-981b-ca88ec0c9dba\") " Mar 07 15:10:58 crc kubenswrapper[4943]: I0307 15:10:58.227473 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/db1ce9d8-45b0-4485-981b-ca88ec0c9dba-etc-swift\") pod \"db1ce9d8-45b0-4485-981b-ca88ec0c9dba\" (UID: \"db1ce9d8-45b0-4485-981b-ca88ec0c9dba\") " Mar 07 15:10:58 crc kubenswrapper[4943]: I0307 15:10:58.227555 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mb64p\" (UniqueName: \"kubernetes.io/projected/db1ce9d8-45b0-4485-981b-ca88ec0c9dba-kube-api-access-mb64p\") pod \"db1ce9d8-45b0-4485-981b-ca88ec0c9dba\" (UID: \"db1ce9d8-45b0-4485-981b-ca88ec0c9dba\") " Mar 07 15:10:58 crc kubenswrapper[4943]: I0307 15:10:58.227696 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/db1ce9d8-45b0-4485-981b-ca88ec0c9dba-scripts\") pod \"db1ce9d8-45b0-4485-981b-ca88ec0c9dba\" (UID: \"db1ce9d8-45b0-4485-981b-ca88ec0c9dba\") " Mar 07 15:10:58 crc kubenswrapper[4943]: I0307 15:10:58.228208 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db1ce9d8-45b0-4485-981b-ca88ec0c9dba-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "db1ce9d8-45b0-4485-981b-ca88ec0c9dba" (UID: "db1ce9d8-45b0-4485-981b-ca88ec0c9dba"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:10:58 crc kubenswrapper[4943]: I0307 15:10:58.228197 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db1ce9d8-45b0-4485-981b-ca88ec0c9dba-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "db1ce9d8-45b0-4485-981b-ca88ec0c9dba" (UID: "db1ce9d8-45b0-4485-981b-ca88ec0c9dba"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:10:58 crc kubenswrapper[4943]: I0307 15:10:58.228658 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/db1ce9d8-45b0-4485-981b-ca88ec0c9dba-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:58 crc kubenswrapper[4943]: I0307 15:10:58.228706 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/db1ce9d8-45b0-4485-981b-ca88ec0c9dba-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:58 crc kubenswrapper[4943]: I0307 15:10:58.233655 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db1ce9d8-45b0-4485-981b-ca88ec0c9dba-kube-api-access-mb64p" (OuterVolumeSpecName: "kube-api-access-mb64p") pod "db1ce9d8-45b0-4485-981b-ca88ec0c9dba" (UID: "db1ce9d8-45b0-4485-981b-ca88ec0c9dba"). InnerVolumeSpecName "kube-api-access-mb64p". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:10:58 crc kubenswrapper[4943]: I0307 15:10:58.249357 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db1ce9d8-45b0-4485-981b-ca88ec0c9dba-scripts" (OuterVolumeSpecName: "scripts") pod "db1ce9d8-45b0-4485-981b-ca88ec0c9dba" (UID: "db1ce9d8-45b0-4485-981b-ca88ec0c9dba"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:10:58 crc kubenswrapper[4943]: I0307 15:10:58.250686 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db1ce9d8-45b0-4485-981b-ca88ec0c9dba-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "db1ce9d8-45b0-4485-981b-ca88ec0c9dba" (UID: "db1ce9d8-45b0-4485-981b-ca88ec0c9dba"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:10:58 crc kubenswrapper[4943]: I0307 15:10:58.269498 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db1ce9d8-45b0-4485-981b-ca88ec0c9dba-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "db1ce9d8-45b0-4485-981b-ca88ec0c9dba" (UID: "db1ce9d8-45b0-4485-981b-ca88ec0c9dba"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:10:58 crc kubenswrapper[4943]: I0307 15:10:58.330073 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/db1ce9d8-45b0-4485-981b-ca88ec0c9dba-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:58 crc kubenswrapper[4943]: I0307 15:10:58.330105 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/db1ce9d8-45b0-4485-981b-ca88ec0c9dba-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:58 crc kubenswrapper[4943]: I0307 15:10:58.330119 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/db1ce9d8-45b0-4485-981b-ca88ec0c9dba-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:58 crc kubenswrapper[4943]: I0307 15:10:58.330135 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mb64p\" (UniqueName: \"kubernetes.io/projected/db1ce9d8-45b0-4485-981b-ca88ec0c9dba-kube-api-access-mb64p\") on node \"crc\" DevicePath \"\"" Mar 07 15:10:58 crc kubenswrapper[4943]: I0307 15:10:58.687331 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bbe1a697072029460b2310dbb9c41a8377d206483afe404bdf4676d21c3b8009" Mar 07 15:10:58 crc kubenswrapper[4943]: I0307 15:10:58.687420 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-hqb5w" Mar 07 15:10:58 crc kubenswrapper[4943]: I0307 15:10:58.791439 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db1ce9d8-45b0-4485-981b-ca88ec0c9dba" path="/var/lib/kubelet/pods/db1ce9d8-45b0-4485-981b-ca88ec0c9dba/volumes" Mar 07 15:10:59 crc kubenswrapper[4943]: I0307 15:10:59.298385 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-t9xkd"] Mar 07 15:10:59 crc kubenswrapper[4943]: E0307 15:10:59.298862 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db1ce9d8-45b0-4485-981b-ca88ec0c9dba" containerName="swift-ring-rebalance" Mar 07 15:10:59 crc kubenswrapper[4943]: I0307 15:10:59.298884 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="db1ce9d8-45b0-4485-981b-ca88ec0c9dba" containerName="swift-ring-rebalance" Mar 07 15:10:59 crc kubenswrapper[4943]: I0307 15:10:59.299142 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="db1ce9d8-45b0-4485-981b-ca88ec0c9dba" containerName="swift-ring-rebalance" Mar 07 15:10:59 crc kubenswrapper[4943]: I0307 15:10:59.299873 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-t9xkd" Mar 07 15:10:59 crc kubenswrapper[4943]: I0307 15:10:59.302985 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:10:59 crc kubenswrapper[4943]: I0307 15:10:59.303015 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:10:59 crc kubenswrapper[4943]: I0307 15:10:59.327489 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-t9xkd"] Mar 07 15:10:59 crc kubenswrapper[4943]: I0307 15:10:59.445376 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/95e41dae-eba7-43ed-b92d-ce2a4cb7a297-etc-swift\") pod \"swift-ring-rebalance-debug-t9xkd\" (UID: \"95e41dae-eba7-43ed-b92d-ce2a4cb7a297\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-t9xkd" Mar 07 15:10:59 crc kubenswrapper[4943]: I0307 15:10:59.445428 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/95e41dae-eba7-43ed-b92d-ce2a4cb7a297-dispersionconf\") pod \"swift-ring-rebalance-debug-t9xkd\" (UID: \"95e41dae-eba7-43ed-b92d-ce2a4cb7a297\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-t9xkd" Mar 07 15:10:59 crc kubenswrapper[4943]: I0307 15:10:59.445497 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/95e41dae-eba7-43ed-b92d-ce2a4cb7a297-ring-data-devices\") pod \"swift-ring-rebalance-debug-t9xkd\" (UID: \"95e41dae-eba7-43ed-b92d-ce2a4cb7a297\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-t9xkd" Mar 07 15:10:59 crc kubenswrapper[4943]: I0307 15:10:59.445556 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/95e41dae-eba7-43ed-b92d-ce2a4cb7a297-scripts\") pod \"swift-ring-rebalance-debug-t9xkd\" (UID: \"95e41dae-eba7-43ed-b92d-ce2a4cb7a297\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-t9xkd" Mar 07 15:10:59 crc kubenswrapper[4943]: I0307 15:10:59.445750 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/95e41dae-eba7-43ed-b92d-ce2a4cb7a297-swiftconf\") pod \"swift-ring-rebalance-debug-t9xkd\" (UID: \"95e41dae-eba7-43ed-b92d-ce2a4cb7a297\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-t9xkd" Mar 07 15:10:59 crc kubenswrapper[4943]: I0307 15:10:59.445956 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l88fb\" (UniqueName: \"kubernetes.io/projected/95e41dae-eba7-43ed-b92d-ce2a4cb7a297-kube-api-access-l88fb\") pod \"swift-ring-rebalance-debug-t9xkd\" (UID: \"95e41dae-eba7-43ed-b92d-ce2a4cb7a297\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-t9xkd" Mar 07 15:10:59 crc kubenswrapper[4943]: I0307 15:10:59.547321 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/95e41dae-eba7-43ed-b92d-ce2a4cb7a297-swiftconf\") pod \"swift-ring-rebalance-debug-t9xkd\" (UID: \"95e41dae-eba7-43ed-b92d-ce2a4cb7a297\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-t9xkd" Mar 07 15:10:59 crc kubenswrapper[4943]: I0307 15:10:59.547423 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l88fb\" (UniqueName: \"kubernetes.io/projected/95e41dae-eba7-43ed-b92d-ce2a4cb7a297-kube-api-access-l88fb\") pod \"swift-ring-rebalance-debug-t9xkd\" (UID: \"95e41dae-eba7-43ed-b92d-ce2a4cb7a297\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-t9xkd" Mar 07 15:10:59 crc kubenswrapper[4943]: I0307 15:10:59.547477 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/95e41dae-eba7-43ed-b92d-ce2a4cb7a297-etc-swift\") pod \"swift-ring-rebalance-debug-t9xkd\" (UID: \"95e41dae-eba7-43ed-b92d-ce2a4cb7a297\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-t9xkd" Mar 07 15:10:59 crc kubenswrapper[4943]: I0307 15:10:59.547497 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/95e41dae-eba7-43ed-b92d-ce2a4cb7a297-dispersionconf\") pod \"swift-ring-rebalance-debug-t9xkd\" (UID: \"95e41dae-eba7-43ed-b92d-ce2a4cb7a297\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-t9xkd" Mar 07 15:10:59 crc kubenswrapper[4943]: I0307 15:10:59.547543 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/95e41dae-eba7-43ed-b92d-ce2a4cb7a297-ring-data-devices\") pod \"swift-ring-rebalance-debug-t9xkd\" (UID: \"95e41dae-eba7-43ed-b92d-ce2a4cb7a297\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-t9xkd" Mar 07 15:10:59 crc kubenswrapper[4943]: I0307 15:10:59.547588 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/95e41dae-eba7-43ed-b92d-ce2a4cb7a297-scripts\") pod \"swift-ring-rebalance-debug-t9xkd\" (UID: \"95e41dae-eba7-43ed-b92d-ce2a4cb7a297\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-t9xkd" Mar 07 15:10:59 crc kubenswrapper[4943]: I0307 15:10:59.548481 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/95e41dae-eba7-43ed-b92d-ce2a4cb7a297-scripts\") pod \"swift-ring-rebalance-debug-t9xkd\" (UID: \"95e41dae-eba7-43ed-b92d-ce2a4cb7a297\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-t9xkd" Mar 07 15:10:59 crc kubenswrapper[4943]: I0307 15:10:59.548795 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/95e41dae-eba7-43ed-b92d-ce2a4cb7a297-ring-data-devices\") pod \"swift-ring-rebalance-debug-t9xkd\" (UID: \"95e41dae-eba7-43ed-b92d-ce2a4cb7a297\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-t9xkd" Mar 07 15:10:59 crc kubenswrapper[4943]: I0307 15:10:59.549347 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/95e41dae-eba7-43ed-b92d-ce2a4cb7a297-etc-swift\") pod \"swift-ring-rebalance-debug-t9xkd\" (UID: \"95e41dae-eba7-43ed-b92d-ce2a4cb7a297\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-t9xkd" Mar 07 15:10:59 crc kubenswrapper[4943]: I0307 15:10:59.554708 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/95e41dae-eba7-43ed-b92d-ce2a4cb7a297-swiftconf\") pod \"swift-ring-rebalance-debug-t9xkd\" (UID: \"95e41dae-eba7-43ed-b92d-ce2a4cb7a297\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-t9xkd" Mar 07 15:10:59 crc kubenswrapper[4943]: I0307 15:10:59.554792 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/95e41dae-eba7-43ed-b92d-ce2a4cb7a297-dispersionconf\") pod \"swift-ring-rebalance-debug-t9xkd\" (UID: \"95e41dae-eba7-43ed-b92d-ce2a4cb7a297\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-t9xkd" Mar 07 15:10:59 crc kubenswrapper[4943]: I0307 15:10:59.582853 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l88fb\" (UniqueName: \"kubernetes.io/projected/95e41dae-eba7-43ed-b92d-ce2a4cb7a297-kube-api-access-l88fb\") pod \"swift-ring-rebalance-debug-t9xkd\" (UID: \"95e41dae-eba7-43ed-b92d-ce2a4cb7a297\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-t9xkd" Mar 07 15:10:59 crc kubenswrapper[4943]: I0307 15:10:59.631822 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-t9xkd" Mar 07 15:10:59 crc kubenswrapper[4943]: I0307 15:10:59.877088 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-t9xkd"] Mar 07 15:11:00 crc kubenswrapper[4943]: I0307 15:11:00.709386 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-t9xkd" event={"ID":"95e41dae-eba7-43ed-b92d-ce2a4cb7a297","Type":"ContainerStarted","Data":"68e7707f6cdf208c71b188376640fe9988be69744d8938ac436f849086359f57"} Mar 07 15:11:00 crc kubenswrapper[4943]: I0307 15:11:00.709473 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-t9xkd" event={"ID":"95e41dae-eba7-43ed-b92d-ce2a4cb7a297","Type":"ContainerStarted","Data":"4ac34e93287d698fb6bfe33a62002b5dc350d35300f3f5d38caee2d3f0ab38d2"} Mar 07 15:11:00 crc kubenswrapper[4943]: I0307 15:11:00.738965 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-t9xkd" podStartSLOduration=1.738900337 podStartE2EDuration="1.738900337s" podCreationTimestamp="2026-03-07 15:10:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:11:00.729741822 +0000 UTC m=+1902.681878360" watchObservedRunningTime="2026-03-07 15:11:00.738900337 +0000 UTC m=+1902.691036875" Mar 07 15:11:01 crc kubenswrapper[4943]: I0307 15:11:01.727665 4943 generic.go:334] "Generic (PLEG): container finished" podID="95e41dae-eba7-43ed-b92d-ce2a4cb7a297" containerID="68e7707f6cdf208c71b188376640fe9988be69744d8938ac436f849086359f57" exitCode=0 Mar 07 15:11:01 crc kubenswrapper[4943]: I0307 15:11:01.727796 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-t9xkd" event={"ID":"95e41dae-eba7-43ed-b92d-ce2a4cb7a297","Type":"ContainerDied","Data":"68e7707f6cdf208c71b188376640fe9988be69744d8938ac436f849086359f57"} Mar 07 15:11:03 crc kubenswrapper[4943]: I0307 15:11:03.101065 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-t9xkd" Mar 07 15:11:03 crc kubenswrapper[4943]: I0307 15:11:03.154598 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-t9xkd"] Mar 07 15:11:03 crc kubenswrapper[4943]: I0307 15:11:03.163109 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-t9xkd"] Mar 07 15:11:03 crc kubenswrapper[4943]: I0307 15:11:03.229419 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/95e41dae-eba7-43ed-b92d-ce2a4cb7a297-swiftconf\") pod \"95e41dae-eba7-43ed-b92d-ce2a4cb7a297\" (UID: \"95e41dae-eba7-43ed-b92d-ce2a4cb7a297\") " Mar 07 15:11:03 crc kubenswrapper[4943]: I0307 15:11:03.229530 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/95e41dae-eba7-43ed-b92d-ce2a4cb7a297-ring-data-devices\") pod \"95e41dae-eba7-43ed-b92d-ce2a4cb7a297\" (UID: \"95e41dae-eba7-43ed-b92d-ce2a4cb7a297\") " Mar 07 15:11:03 crc kubenswrapper[4943]: I0307 15:11:03.229605 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/95e41dae-eba7-43ed-b92d-ce2a4cb7a297-scripts\") pod \"95e41dae-eba7-43ed-b92d-ce2a4cb7a297\" (UID: \"95e41dae-eba7-43ed-b92d-ce2a4cb7a297\") " Mar 07 15:11:03 crc kubenswrapper[4943]: I0307 15:11:03.229651 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l88fb\" (UniqueName: \"kubernetes.io/projected/95e41dae-eba7-43ed-b92d-ce2a4cb7a297-kube-api-access-l88fb\") pod \"95e41dae-eba7-43ed-b92d-ce2a4cb7a297\" (UID: \"95e41dae-eba7-43ed-b92d-ce2a4cb7a297\") " Mar 07 15:11:03 crc kubenswrapper[4943]: I0307 15:11:03.229679 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/95e41dae-eba7-43ed-b92d-ce2a4cb7a297-dispersionconf\") pod \"95e41dae-eba7-43ed-b92d-ce2a4cb7a297\" (UID: \"95e41dae-eba7-43ed-b92d-ce2a4cb7a297\") " Mar 07 15:11:03 crc kubenswrapper[4943]: I0307 15:11:03.229700 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/95e41dae-eba7-43ed-b92d-ce2a4cb7a297-etc-swift\") pod \"95e41dae-eba7-43ed-b92d-ce2a4cb7a297\" (UID: \"95e41dae-eba7-43ed-b92d-ce2a4cb7a297\") " Mar 07 15:11:03 crc kubenswrapper[4943]: I0307 15:11:03.231018 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95e41dae-eba7-43ed-b92d-ce2a4cb7a297-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "95e41dae-eba7-43ed-b92d-ce2a4cb7a297" (UID: "95e41dae-eba7-43ed-b92d-ce2a4cb7a297"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:11:03 crc kubenswrapper[4943]: I0307 15:11:03.232193 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95e41dae-eba7-43ed-b92d-ce2a4cb7a297-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "95e41dae-eba7-43ed-b92d-ce2a4cb7a297" (UID: "95e41dae-eba7-43ed-b92d-ce2a4cb7a297"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:11:03 crc kubenswrapper[4943]: I0307 15:11:03.238875 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95e41dae-eba7-43ed-b92d-ce2a4cb7a297-kube-api-access-l88fb" (OuterVolumeSpecName: "kube-api-access-l88fb") pod "95e41dae-eba7-43ed-b92d-ce2a4cb7a297" (UID: "95e41dae-eba7-43ed-b92d-ce2a4cb7a297"). InnerVolumeSpecName "kube-api-access-l88fb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:11:03 crc kubenswrapper[4943]: I0307 15:11:03.262500 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95e41dae-eba7-43ed-b92d-ce2a4cb7a297-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "95e41dae-eba7-43ed-b92d-ce2a4cb7a297" (UID: "95e41dae-eba7-43ed-b92d-ce2a4cb7a297"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:11:03 crc kubenswrapper[4943]: I0307 15:11:03.271278 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95e41dae-eba7-43ed-b92d-ce2a4cb7a297-scripts" (OuterVolumeSpecName: "scripts") pod "95e41dae-eba7-43ed-b92d-ce2a4cb7a297" (UID: "95e41dae-eba7-43ed-b92d-ce2a4cb7a297"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:11:03 crc kubenswrapper[4943]: I0307 15:11:03.271622 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95e41dae-eba7-43ed-b92d-ce2a4cb7a297-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "95e41dae-eba7-43ed-b92d-ce2a4cb7a297" (UID: "95e41dae-eba7-43ed-b92d-ce2a4cb7a297"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:11:03 crc kubenswrapper[4943]: I0307 15:11:03.331493 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/95e41dae-eba7-43ed-b92d-ce2a4cb7a297-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:03 crc kubenswrapper[4943]: I0307 15:11:03.331532 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/95e41dae-eba7-43ed-b92d-ce2a4cb7a297-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:03 crc kubenswrapper[4943]: I0307 15:11:03.331546 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/95e41dae-eba7-43ed-b92d-ce2a4cb7a297-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:03 crc kubenswrapper[4943]: I0307 15:11:03.331557 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l88fb\" (UniqueName: \"kubernetes.io/projected/95e41dae-eba7-43ed-b92d-ce2a4cb7a297-kube-api-access-l88fb\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:03 crc kubenswrapper[4943]: I0307 15:11:03.331569 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/95e41dae-eba7-43ed-b92d-ce2a4cb7a297-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:03 crc kubenswrapper[4943]: I0307 15:11:03.331579 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/95e41dae-eba7-43ed-b92d-ce2a4cb7a297-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:03 crc kubenswrapper[4943]: I0307 15:11:03.756389 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4ac34e93287d698fb6bfe33a62002b5dc350d35300f3f5d38caee2d3f0ab38d2" Mar 07 15:11:03 crc kubenswrapper[4943]: I0307 15:11:03.756467 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-t9xkd" Mar 07 15:11:04 crc kubenswrapper[4943]: I0307 15:11:04.327101 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-kqvrl"] Mar 07 15:11:04 crc kubenswrapper[4943]: E0307 15:11:04.327533 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95e41dae-eba7-43ed-b92d-ce2a4cb7a297" containerName="swift-ring-rebalance" Mar 07 15:11:04 crc kubenswrapper[4943]: I0307 15:11:04.327552 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="95e41dae-eba7-43ed-b92d-ce2a4cb7a297" containerName="swift-ring-rebalance" Mar 07 15:11:04 crc kubenswrapper[4943]: I0307 15:11:04.327733 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="95e41dae-eba7-43ed-b92d-ce2a4cb7a297" containerName="swift-ring-rebalance" Mar 07 15:11:04 crc kubenswrapper[4943]: I0307 15:11:04.328439 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kqvrl" Mar 07 15:11:04 crc kubenswrapper[4943]: I0307 15:11:04.334473 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:11:04 crc kubenswrapper[4943]: I0307 15:11:04.334968 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:11:04 crc kubenswrapper[4943]: I0307 15:11:04.343377 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-kqvrl"] Mar 07 15:11:04 crc kubenswrapper[4943]: I0307 15:11:04.394524 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4-swiftconf\") pod \"swift-ring-rebalance-debug-kqvrl\" (UID: \"8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kqvrl" Mar 07 15:11:04 crc kubenswrapper[4943]: I0307 15:11:04.394609 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4-ring-data-devices\") pod \"swift-ring-rebalance-debug-kqvrl\" (UID: \"8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kqvrl" Mar 07 15:11:04 crc kubenswrapper[4943]: I0307 15:11:04.394715 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4-scripts\") pod \"swift-ring-rebalance-debug-kqvrl\" (UID: \"8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kqvrl" Mar 07 15:11:04 crc kubenswrapper[4943]: I0307 15:11:04.394813 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4-etc-swift\") pod \"swift-ring-rebalance-debug-kqvrl\" (UID: \"8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kqvrl" Mar 07 15:11:04 crc kubenswrapper[4943]: I0307 15:11:04.394907 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9t5r2\" (UniqueName: \"kubernetes.io/projected/8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4-kube-api-access-9t5r2\") pod \"swift-ring-rebalance-debug-kqvrl\" (UID: \"8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kqvrl" Mar 07 15:11:04 crc kubenswrapper[4943]: I0307 15:11:04.394958 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4-dispersionconf\") pod \"swift-ring-rebalance-debug-kqvrl\" (UID: \"8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kqvrl" Mar 07 15:11:04 crc kubenswrapper[4943]: I0307 15:11:04.496656 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4-scripts\") pod \"swift-ring-rebalance-debug-kqvrl\" (UID: \"8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kqvrl" Mar 07 15:11:04 crc kubenswrapper[4943]: I0307 15:11:04.496761 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4-etc-swift\") pod \"swift-ring-rebalance-debug-kqvrl\" (UID: \"8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kqvrl" Mar 07 15:11:04 crc kubenswrapper[4943]: I0307 15:11:04.496856 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9t5r2\" (UniqueName: \"kubernetes.io/projected/8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4-kube-api-access-9t5r2\") pod \"swift-ring-rebalance-debug-kqvrl\" (UID: \"8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kqvrl" Mar 07 15:11:04 crc kubenswrapper[4943]: I0307 15:11:04.496900 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4-dispersionconf\") pod \"swift-ring-rebalance-debug-kqvrl\" (UID: \"8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kqvrl" Mar 07 15:11:04 crc kubenswrapper[4943]: I0307 15:11:04.496982 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4-swiftconf\") pod \"swift-ring-rebalance-debug-kqvrl\" (UID: \"8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kqvrl" Mar 07 15:11:04 crc kubenswrapper[4943]: I0307 15:11:04.497026 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4-ring-data-devices\") pod \"swift-ring-rebalance-debug-kqvrl\" (UID: \"8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kqvrl" Mar 07 15:11:04 crc kubenswrapper[4943]: I0307 15:11:04.497619 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4-etc-swift\") pod \"swift-ring-rebalance-debug-kqvrl\" (UID: \"8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kqvrl" Mar 07 15:11:04 crc kubenswrapper[4943]: I0307 15:11:04.498230 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4-ring-data-devices\") pod \"swift-ring-rebalance-debug-kqvrl\" (UID: \"8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kqvrl" Mar 07 15:11:04 crc kubenswrapper[4943]: I0307 15:11:04.498808 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4-scripts\") pod \"swift-ring-rebalance-debug-kqvrl\" (UID: \"8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kqvrl" Mar 07 15:11:04 crc kubenswrapper[4943]: I0307 15:11:04.509913 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4-swiftconf\") pod \"swift-ring-rebalance-debug-kqvrl\" (UID: \"8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kqvrl" Mar 07 15:11:04 crc kubenswrapper[4943]: I0307 15:11:04.510916 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4-dispersionconf\") pod \"swift-ring-rebalance-debug-kqvrl\" (UID: \"8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kqvrl" Mar 07 15:11:04 crc kubenswrapper[4943]: I0307 15:11:04.514892 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9t5r2\" (UniqueName: \"kubernetes.io/projected/8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4-kube-api-access-9t5r2\") pod \"swift-ring-rebalance-debug-kqvrl\" (UID: \"8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kqvrl" Mar 07 15:11:04 crc kubenswrapper[4943]: I0307 15:11:04.652159 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kqvrl" Mar 07 15:11:04 crc kubenswrapper[4943]: I0307 15:11:04.775065 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95e41dae-eba7-43ed-b92d-ce2a4cb7a297" path="/var/lib/kubelet/pods/95e41dae-eba7-43ed-b92d-ce2a4cb7a297/volumes" Mar 07 15:11:04 crc kubenswrapper[4943]: I0307 15:11:04.888157 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-kqvrl"] Mar 07 15:11:05 crc kubenswrapper[4943]: I0307 15:11:05.782365 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kqvrl" event={"ID":"8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4","Type":"ContainerStarted","Data":"aa9bf1456c0952d2717320faaaaba5b5277d81c2b5faac7bcd32617e1aca82b6"} Mar 07 15:11:05 crc kubenswrapper[4943]: I0307 15:11:05.782681 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kqvrl" event={"ID":"8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4","Type":"ContainerStarted","Data":"174abee5941416a6bfc056861770044fd2f3086a4e8a9ae22d9a96939f6017e5"} Mar 07 15:11:05 crc kubenswrapper[4943]: I0307 15:11:05.809674 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kqvrl" podStartSLOduration=1.809650867 podStartE2EDuration="1.809650867s" podCreationTimestamp="2026-03-07 15:11:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:11:05.80121601 +0000 UTC m=+1907.753352508" watchObservedRunningTime="2026-03-07 15:11:05.809650867 +0000 UTC m=+1907.761787395" Mar 07 15:11:06 crc kubenswrapper[4943]: I0307 15:11:06.791119 4943 generic.go:334] "Generic (PLEG): container finished" podID="8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4" containerID="aa9bf1456c0952d2717320faaaaba5b5277d81c2b5faac7bcd32617e1aca82b6" exitCode=0 Mar 07 15:11:06 crc kubenswrapper[4943]: I0307 15:11:06.791188 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kqvrl" event={"ID":"8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4","Type":"ContainerDied","Data":"aa9bf1456c0952d2717320faaaaba5b5277d81c2b5faac7bcd32617e1aca82b6"} Mar 07 15:11:08 crc kubenswrapper[4943]: I0307 15:11:08.124949 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kqvrl" Mar 07 15:11:08 crc kubenswrapper[4943]: I0307 15:11:08.153184 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4-ring-data-devices\") pod \"8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4\" (UID: \"8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4\") " Mar 07 15:11:08 crc kubenswrapper[4943]: I0307 15:11:08.153244 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4-dispersionconf\") pod \"8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4\" (UID: \"8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4\") " Mar 07 15:11:08 crc kubenswrapper[4943]: I0307 15:11:08.153287 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4-swiftconf\") pod \"8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4\" (UID: \"8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4\") " Mar 07 15:11:08 crc kubenswrapper[4943]: I0307 15:11:08.153359 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4-scripts\") pod \"8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4\" (UID: \"8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4\") " Mar 07 15:11:08 crc kubenswrapper[4943]: I0307 15:11:08.153449 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9t5r2\" (UniqueName: \"kubernetes.io/projected/8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4-kube-api-access-9t5r2\") pod \"8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4\" (UID: \"8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4\") " Mar 07 15:11:08 crc kubenswrapper[4943]: I0307 15:11:08.153518 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4-etc-swift\") pod \"8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4\" (UID: \"8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4\") " Mar 07 15:11:08 crc kubenswrapper[4943]: I0307 15:11:08.154520 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4" (UID: "8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:11:08 crc kubenswrapper[4943]: I0307 15:11:08.155071 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4" (UID: "8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:11:08 crc kubenswrapper[4943]: I0307 15:11:08.164334 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4-kube-api-access-9t5r2" (OuterVolumeSpecName: "kube-api-access-9t5r2") pod "8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4" (UID: "8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4"). InnerVolumeSpecName "kube-api-access-9t5r2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:11:08 crc kubenswrapper[4943]: I0307 15:11:08.171556 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-kqvrl"] Mar 07 15:11:08 crc kubenswrapper[4943]: I0307 15:11:08.179496 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4" (UID: "8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:11:08 crc kubenswrapper[4943]: I0307 15:11:08.183753 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-kqvrl"] Mar 07 15:11:08 crc kubenswrapper[4943]: I0307 15:11:08.196380 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4-scripts" (OuterVolumeSpecName: "scripts") pod "8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4" (UID: "8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:11:08 crc kubenswrapper[4943]: I0307 15:11:08.197868 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4" (UID: "8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:11:08 crc kubenswrapper[4943]: I0307 15:11:08.255335 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:08 crc kubenswrapper[4943]: I0307 15:11:08.255368 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:08 crc kubenswrapper[4943]: I0307 15:11:08.255383 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9t5r2\" (UniqueName: \"kubernetes.io/projected/8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4-kube-api-access-9t5r2\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:08 crc kubenswrapper[4943]: I0307 15:11:08.255400 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:08 crc kubenswrapper[4943]: I0307 15:11:08.255412 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:08 crc kubenswrapper[4943]: I0307 15:11:08.255423 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:08 crc kubenswrapper[4943]: I0307 15:11:08.770470 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4" path="/var/lib/kubelet/pods/8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4/volumes" Mar 07 15:11:08 crc kubenswrapper[4943]: I0307 15:11:08.812478 4943 scope.go:117] "RemoveContainer" containerID="aa9bf1456c0952d2717320faaaaba5b5277d81c2b5faac7bcd32617e1aca82b6" Mar 07 15:11:08 crc kubenswrapper[4943]: I0307 15:11:08.812592 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kqvrl" Mar 07 15:11:09 crc kubenswrapper[4943]: I0307 15:11:09.408242 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-86tkz"] Mar 07 15:11:09 crc kubenswrapper[4943]: E0307 15:11:09.408675 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4" containerName="swift-ring-rebalance" Mar 07 15:11:09 crc kubenswrapper[4943]: I0307 15:11:09.408694 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4" containerName="swift-ring-rebalance" Mar 07 15:11:09 crc kubenswrapper[4943]: I0307 15:11:09.408959 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b23b3ba-c661-46bd-9e15-4bbc0fd54cf4" containerName="swift-ring-rebalance" Mar 07 15:11:09 crc kubenswrapper[4943]: I0307 15:11:09.409682 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-86tkz" Mar 07 15:11:09 crc kubenswrapper[4943]: I0307 15:11:09.415201 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:11:09 crc kubenswrapper[4943]: I0307 15:11:09.416879 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:11:09 crc kubenswrapper[4943]: I0307 15:11:09.439837 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-86tkz"] Mar 07 15:11:09 crc kubenswrapper[4943]: I0307 15:11:09.475645 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/241288de-75c2-4fdf-b691-fa6b92607595-swiftconf\") pod \"swift-ring-rebalance-debug-86tkz\" (UID: \"241288de-75c2-4fdf-b691-fa6b92607595\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-86tkz" Mar 07 15:11:09 crc kubenswrapper[4943]: I0307 15:11:09.475737 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/241288de-75c2-4fdf-b691-fa6b92607595-etc-swift\") pod \"swift-ring-rebalance-debug-86tkz\" (UID: \"241288de-75c2-4fdf-b691-fa6b92607595\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-86tkz" Mar 07 15:11:09 crc kubenswrapper[4943]: I0307 15:11:09.475813 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/241288de-75c2-4fdf-b691-fa6b92607595-ring-data-devices\") pod \"swift-ring-rebalance-debug-86tkz\" (UID: \"241288de-75c2-4fdf-b691-fa6b92607595\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-86tkz" Mar 07 15:11:09 crc kubenswrapper[4943]: I0307 15:11:09.475973 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/241288de-75c2-4fdf-b691-fa6b92607595-scripts\") pod \"swift-ring-rebalance-debug-86tkz\" (UID: \"241288de-75c2-4fdf-b691-fa6b92607595\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-86tkz" Mar 07 15:11:09 crc kubenswrapper[4943]: I0307 15:11:09.476201 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/241288de-75c2-4fdf-b691-fa6b92607595-dispersionconf\") pod \"swift-ring-rebalance-debug-86tkz\" (UID: \"241288de-75c2-4fdf-b691-fa6b92607595\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-86tkz" Mar 07 15:11:09 crc kubenswrapper[4943]: I0307 15:11:09.476259 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6r4g\" (UniqueName: \"kubernetes.io/projected/241288de-75c2-4fdf-b691-fa6b92607595-kube-api-access-g6r4g\") pod \"swift-ring-rebalance-debug-86tkz\" (UID: \"241288de-75c2-4fdf-b691-fa6b92607595\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-86tkz" Mar 07 15:11:09 crc kubenswrapper[4943]: I0307 15:11:09.578096 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/241288de-75c2-4fdf-b691-fa6b92607595-scripts\") pod \"swift-ring-rebalance-debug-86tkz\" (UID: \"241288de-75c2-4fdf-b691-fa6b92607595\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-86tkz" Mar 07 15:11:09 crc kubenswrapper[4943]: I0307 15:11:09.578221 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/241288de-75c2-4fdf-b691-fa6b92607595-dispersionconf\") pod \"swift-ring-rebalance-debug-86tkz\" (UID: \"241288de-75c2-4fdf-b691-fa6b92607595\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-86tkz" Mar 07 15:11:09 crc kubenswrapper[4943]: I0307 15:11:09.578267 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6r4g\" (UniqueName: \"kubernetes.io/projected/241288de-75c2-4fdf-b691-fa6b92607595-kube-api-access-g6r4g\") pod \"swift-ring-rebalance-debug-86tkz\" (UID: \"241288de-75c2-4fdf-b691-fa6b92607595\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-86tkz" Mar 07 15:11:09 crc kubenswrapper[4943]: I0307 15:11:09.578342 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/241288de-75c2-4fdf-b691-fa6b92607595-swiftconf\") pod \"swift-ring-rebalance-debug-86tkz\" (UID: \"241288de-75c2-4fdf-b691-fa6b92607595\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-86tkz" Mar 07 15:11:09 crc kubenswrapper[4943]: I0307 15:11:09.578425 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/241288de-75c2-4fdf-b691-fa6b92607595-etc-swift\") pod \"swift-ring-rebalance-debug-86tkz\" (UID: \"241288de-75c2-4fdf-b691-fa6b92607595\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-86tkz" Mar 07 15:11:09 crc kubenswrapper[4943]: I0307 15:11:09.578514 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/241288de-75c2-4fdf-b691-fa6b92607595-ring-data-devices\") pod \"swift-ring-rebalance-debug-86tkz\" (UID: \"241288de-75c2-4fdf-b691-fa6b92607595\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-86tkz" Mar 07 15:11:09 crc kubenswrapper[4943]: I0307 15:11:09.579315 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/241288de-75c2-4fdf-b691-fa6b92607595-etc-swift\") pod \"swift-ring-rebalance-debug-86tkz\" (UID: \"241288de-75c2-4fdf-b691-fa6b92607595\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-86tkz" Mar 07 15:11:09 crc kubenswrapper[4943]: I0307 15:11:09.579646 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/241288de-75c2-4fdf-b691-fa6b92607595-scripts\") pod \"swift-ring-rebalance-debug-86tkz\" (UID: \"241288de-75c2-4fdf-b691-fa6b92607595\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-86tkz" Mar 07 15:11:09 crc kubenswrapper[4943]: I0307 15:11:09.579663 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/241288de-75c2-4fdf-b691-fa6b92607595-ring-data-devices\") pod \"swift-ring-rebalance-debug-86tkz\" (UID: \"241288de-75c2-4fdf-b691-fa6b92607595\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-86tkz" Mar 07 15:11:09 crc kubenswrapper[4943]: I0307 15:11:09.584116 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/241288de-75c2-4fdf-b691-fa6b92607595-swiftconf\") pod \"swift-ring-rebalance-debug-86tkz\" (UID: \"241288de-75c2-4fdf-b691-fa6b92607595\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-86tkz" Mar 07 15:11:09 crc kubenswrapper[4943]: I0307 15:11:09.585809 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/241288de-75c2-4fdf-b691-fa6b92607595-dispersionconf\") pod \"swift-ring-rebalance-debug-86tkz\" (UID: \"241288de-75c2-4fdf-b691-fa6b92607595\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-86tkz" Mar 07 15:11:09 crc kubenswrapper[4943]: I0307 15:11:09.600989 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6r4g\" (UniqueName: \"kubernetes.io/projected/241288de-75c2-4fdf-b691-fa6b92607595-kube-api-access-g6r4g\") pod \"swift-ring-rebalance-debug-86tkz\" (UID: \"241288de-75c2-4fdf-b691-fa6b92607595\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-86tkz" Mar 07 15:11:09 crc kubenswrapper[4943]: I0307 15:11:09.741050 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-86tkz" Mar 07 15:11:09 crc kubenswrapper[4943]: I0307 15:11:09.973092 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-86tkz"] Mar 07 15:11:09 crc kubenswrapper[4943]: W0307 15:11:09.980024 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod241288de_75c2_4fdf_b691_fa6b92607595.slice/crio-d80f5e504b2cc605c20655d516f30f48471c93738c34cc8a2b596ebd56765447 WatchSource:0}: Error finding container d80f5e504b2cc605c20655d516f30f48471c93738c34cc8a2b596ebd56765447: Status 404 returned error can't find the container with id d80f5e504b2cc605c20655d516f30f48471c93738c34cc8a2b596ebd56765447 Mar 07 15:11:10 crc kubenswrapper[4943]: I0307 15:11:10.838496 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-86tkz" event={"ID":"241288de-75c2-4fdf-b691-fa6b92607595","Type":"ContainerStarted","Data":"8787decac2665a4084b059ae56065aeeb268fd1a232ba5d141389c8e154e64ff"} Mar 07 15:11:10 crc kubenswrapper[4943]: I0307 15:11:10.838899 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-86tkz" event={"ID":"241288de-75c2-4fdf-b691-fa6b92607595","Type":"ContainerStarted","Data":"d80f5e504b2cc605c20655d516f30f48471c93738c34cc8a2b596ebd56765447"} Mar 07 15:11:10 crc kubenswrapper[4943]: I0307 15:11:10.875880 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-86tkz" podStartSLOduration=1.875855038 podStartE2EDuration="1.875855038s" podCreationTimestamp="2026-03-07 15:11:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:11:10.866490218 +0000 UTC m=+1912.818626756" watchObservedRunningTime="2026-03-07 15:11:10.875855038 +0000 UTC m=+1912.827991566" Mar 07 15:11:11 crc kubenswrapper[4943]: I0307 15:11:11.851338 4943 generic.go:334] "Generic (PLEG): container finished" podID="241288de-75c2-4fdf-b691-fa6b92607595" containerID="8787decac2665a4084b059ae56065aeeb268fd1a232ba5d141389c8e154e64ff" exitCode=0 Mar 07 15:11:11 crc kubenswrapper[4943]: I0307 15:11:11.851404 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-86tkz" event={"ID":"241288de-75c2-4fdf-b691-fa6b92607595","Type":"ContainerDied","Data":"8787decac2665a4084b059ae56065aeeb268fd1a232ba5d141389c8e154e64ff"} Mar 07 15:11:13 crc kubenswrapper[4943]: I0307 15:11:13.218503 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-86tkz" Mar 07 15:11:13 crc kubenswrapper[4943]: I0307 15:11:13.256172 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/241288de-75c2-4fdf-b691-fa6b92607595-ring-data-devices\") pod \"241288de-75c2-4fdf-b691-fa6b92607595\" (UID: \"241288de-75c2-4fdf-b691-fa6b92607595\") " Mar 07 15:11:13 crc kubenswrapper[4943]: I0307 15:11:13.256227 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/241288de-75c2-4fdf-b691-fa6b92607595-scripts\") pod \"241288de-75c2-4fdf-b691-fa6b92607595\" (UID: \"241288de-75c2-4fdf-b691-fa6b92607595\") " Mar 07 15:11:13 crc kubenswrapper[4943]: I0307 15:11:13.256263 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/241288de-75c2-4fdf-b691-fa6b92607595-swiftconf\") pod \"241288de-75c2-4fdf-b691-fa6b92607595\" (UID: \"241288de-75c2-4fdf-b691-fa6b92607595\") " Mar 07 15:11:13 crc kubenswrapper[4943]: I0307 15:11:13.256387 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/241288de-75c2-4fdf-b691-fa6b92607595-dispersionconf\") pod \"241288de-75c2-4fdf-b691-fa6b92607595\" (UID: \"241288de-75c2-4fdf-b691-fa6b92607595\") " Mar 07 15:11:13 crc kubenswrapper[4943]: I0307 15:11:13.256437 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/241288de-75c2-4fdf-b691-fa6b92607595-etc-swift\") pod \"241288de-75c2-4fdf-b691-fa6b92607595\" (UID: \"241288de-75c2-4fdf-b691-fa6b92607595\") " Mar 07 15:11:13 crc kubenswrapper[4943]: I0307 15:11:13.256480 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g6r4g\" (UniqueName: \"kubernetes.io/projected/241288de-75c2-4fdf-b691-fa6b92607595-kube-api-access-g6r4g\") pod \"241288de-75c2-4fdf-b691-fa6b92607595\" (UID: \"241288de-75c2-4fdf-b691-fa6b92607595\") " Mar 07 15:11:13 crc kubenswrapper[4943]: I0307 15:11:13.270848 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-86tkz"] Mar 07 15:11:13 crc kubenswrapper[4943]: I0307 15:11:13.271250 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-86tkz"] Mar 07 15:11:13 crc kubenswrapper[4943]: I0307 15:11:13.272116 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/241288de-75c2-4fdf-b691-fa6b92607595-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "241288de-75c2-4fdf-b691-fa6b92607595" (UID: "241288de-75c2-4fdf-b691-fa6b92607595"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:11:13 crc kubenswrapper[4943]: I0307 15:11:13.283624 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/241288de-75c2-4fdf-b691-fa6b92607595-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "241288de-75c2-4fdf-b691-fa6b92607595" (UID: "241288de-75c2-4fdf-b691-fa6b92607595"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:11:13 crc kubenswrapper[4943]: I0307 15:11:13.290468 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/241288de-75c2-4fdf-b691-fa6b92607595-kube-api-access-g6r4g" (OuterVolumeSpecName: "kube-api-access-g6r4g") pod "241288de-75c2-4fdf-b691-fa6b92607595" (UID: "241288de-75c2-4fdf-b691-fa6b92607595"). InnerVolumeSpecName "kube-api-access-g6r4g". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:11:13 crc kubenswrapper[4943]: I0307 15:11:13.294503 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/241288de-75c2-4fdf-b691-fa6b92607595-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "241288de-75c2-4fdf-b691-fa6b92607595" (UID: "241288de-75c2-4fdf-b691-fa6b92607595"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:11:13 crc kubenswrapper[4943]: I0307 15:11:13.315509 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/241288de-75c2-4fdf-b691-fa6b92607595-scripts" (OuterVolumeSpecName: "scripts") pod "241288de-75c2-4fdf-b691-fa6b92607595" (UID: "241288de-75c2-4fdf-b691-fa6b92607595"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:11:13 crc kubenswrapper[4943]: I0307 15:11:13.318793 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/241288de-75c2-4fdf-b691-fa6b92607595-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "241288de-75c2-4fdf-b691-fa6b92607595" (UID: "241288de-75c2-4fdf-b691-fa6b92607595"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:11:13 crc kubenswrapper[4943]: I0307 15:11:13.358553 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/241288de-75c2-4fdf-b691-fa6b92607595-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:13 crc kubenswrapper[4943]: I0307 15:11:13.358592 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/241288de-75c2-4fdf-b691-fa6b92607595-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:13 crc kubenswrapper[4943]: I0307 15:11:13.358606 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g6r4g\" (UniqueName: \"kubernetes.io/projected/241288de-75c2-4fdf-b691-fa6b92607595-kube-api-access-g6r4g\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:13 crc kubenswrapper[4943]: I0307 15:11:13.358620 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/241288de-75c2-4fdf-b691-fa6b92607595-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:13 crc kubenswrapper[4943]: I0307 15:11:13.358633 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/241288de-75c2-4fdf-b691-fa6b92607595-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:13 crc kubenswrapper[4943]: I0307 15:11:13.358644 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/241288de-75c2-4fdf-b691-fa6b92607595-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:13 crc kubenswrapper[4943]: I0307 15:11:13.874508 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d80f5e504b2cc605c20655d516f30f48471c93738c34cc8a2b596ebd56765447" Mar 07 15:11:13 crc kubenswrapper[4943]: I0307 15:11:13.874576 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-86tkz" Mar 07 15:11:14 crc kubenswrapper[4943]: I0307 15:11:14.513228 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-ljqll"] Mar 07 15:11:14 crc kubenswrapper[4943]: E0307 15:11:14.513962 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="241288de-75c2-4fdf-b691-fa6b92607595" containerName="swift-ring-rebalance" Mar 07 15:11:14 crc kubenswrapper[4943]: I0307 15:11:14.513994 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="241288de-75c2-4fdf-b691-fa6b92607595" containerName="swift-ring-rebalance" Mar 07 15:11:14 crc kubenswrapper[4943]: I0307 15:11:14.514283 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="241288de-75c2-4fdf-b691-fa6b92607595" containerName="swift-ring-rebalance" Mar 07 15:11:14 crc kubenswrapper[4943]: I0307 15:11:14.515218 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ljqll" Mar 07 15:11:14 crc kubenswrapper[4943]: I0307 15:11:14.517843 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:11:14 crc kubenswrapper[4943]: I0307 15:11:14.527594 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:11:14 crc kubenswrapper[4943]: I0307 15:11:14.527589 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-ljqll"] Mar 07 15:11:14 crc kubenswrapper[4943]: I0307 15:11:14.581118 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e-etc-swift\") pod \"swift-ring-rebalance-debug-ljqll\" (UID: \"16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ljqll" Mar 07 15:11:14 crc kubenswrapper[4943]: I0307 15:11:14.581182 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2w7x\" (UniqueName: \"kubernetes.io/projected/16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e-kube-api-access-x2w7x\") pod \"swift-ring-rebalance-debug-ljqll\" (UID: \"16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ljqll" Mar 07 15:11:14 crc kubenswrapper[4943]: I0307 15:11:14.581465 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e-dispersionconf\") pod \"swift-ring-rebalance-debug-ljqll\" (UID: \"16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ljqll" Mar 07 15:11:14 crc kubenswrapper[4943]: I0307 15:11:14.581531 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e-scripts\") pod \"swift-ring-rebalance-debug-ljqll\" (UID: \"16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ljqll" Mar 07 15:11:14 crc kubenswrapper[4943]: I0307 15:11:14.581575 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e-ring-data-devices\") pod \"swift-ring-rebalance-debug-ljqll\" (UID: \"16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ljqll" Mar 07 15:11:14 crc kubenswrapper[4943]: I0307 15:11:14.581708 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e-swiftconf\") pod \"swift-ring-rebalance-debug-ljqll\" (UID: \"16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ljqll" Mar 07 15:11:14 crc kubenswrapper[4943]: I0307 15:11:14.683587 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2w7x\" (UniqueName: \"kubernetes.io/projected/16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e-kube-api-access-x2w7x\") pod \"swift-ring-rebalance-debug-ljqll\" (UID: \"16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ljqll" Mar 07 15:11:14 crc kubenswrapper[4943]: I0307 15:11:14.684241 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e-dispersionconf\") pod \"swift-ring-rebalance-debug-ljqll\" (UID: \"16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ljqll" Mar 07 15:11:14 crc kubenswrapper[4943]: I0307 15:11:14.684447 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e-scripts\") pod \"swift-ring-rebalance-debug-ljqll\" (UID: \"16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ljqll" Mar 07 15:11:14 crc kubenswrapper[4943]: I0307 15:11:14.684618 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e-ring-data-devices\") pod \"swift-ring-rebalance-debug-ljqll\" (UID: \"16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ljqll" Mar 07 15:11:14 crc kubenswrapper[4943]: I0307 15:11:14.684780 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e-swiftconf\") pod \"swift-ring-rebalance-debug-ljqll\" (UID: \"16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ljqll" Mar 07 15:11:14 crc kubenswrapper[4943]: I0307 15:11:14.684996 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e-etc-swift\") pod \"swift-ring-rebalance-debug-ljqll\" (UID: \"16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ljqll" Mar 07 15:11:14 crc kubenswrapper[4943]: I0307 15:11:14.685376 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e-scripts\") pod \"swift-ring-rebalance-debug-ljqll\" (UID: \"16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ljqll" Mar 07 15:11:14 crc kubenswrapper[4943]: I0307 15:11:14.686257 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e-ring-data-devices\") pod \"swift-ring-rebalance-debug-ljqll\" (UID: \"16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ljqll" Mar 07 15:11:14 crc kubenswrapper[4943]: I0307 15:11:14.686582 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e-etc-swift\") pod \"swift-ring-rebalance-debug-ljqll\" (UID: \"16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ljqll" Mar 07 15:11:14 crc kubenswrapper[4943]: I0307 15:11:14.689811 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e-swiftconf\") pod \"swift-ring-rebalance-debug-ljqll\" (UID: \"16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ljqll" Mar 07 15:11:14 crc kubenswrapper[4943]: I0307 15:11:14.690231 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e-dispersionconf\") pod \"swift-ring-rebalance-debug-ljqll\" (UID: \"16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ljqll" Mar 07 15:11:14 crc kubenswrapper[4943]: I0307 15:11:14.708155 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2w7x\" (UniqueName: \"kubernetes.io/projected/16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e-kube-api-access-x2w7x\") pod \"swift-ring-rebalance-debug-ljqll\" (UID: \"16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ljqll" Mar 07 15:11:14 crc kubenswrapper[4943]: I0307 15:11:14.768309 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="241288de-75c2-4fdf-b691-fa6b92607595" path="/var/lib/kubelet/pods/241288de-75c2-4fdf-b691-fa6b92607595/volumes" Mar 07 15:11:14 crc kubenswrapper[4943]: I0307 15:11:14.861563 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ljqll" Mar 07 15:11:15 crc kubenswrapper[4943]: I0307 15:11:15.362448 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-ljqll"] Mar 07 15:11:15 crc kubenswrapper[4943]: I0307 15:11:15.895914 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ljqll" event={"ID":"16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e","Type":"ContainerStarted","Data":"c1ae345069473e9b091f876427afe14066b3e06b2c88ea08a4ab5d08935084cb"} Mar 07 15:11:15 crc kubenswrapper[4943]: I0307 15:11:15.896004 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ljqll" event={"ID":"16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e","Type":"ContainerStarted","Data":"497dccc8550907988d093082c16cb5db094afabab9efb161548044f5a690ce6e"} Mar 07 15:11:15 crc kubenswrapper[4943]: I0307 15:11:15.915391 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ljqll" podStartSLOduration=1.915363784 podStartE2EDuration="1.915363784s" podCreationTimestamp="2026-03-07 15:11:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:11:15.914638146 +0000 UTC m=+1917.866774644" watchObservedRunningTime="2026-03-07 15:11:15.915363784 +0000 UTC m=+1917.867500322" Mar 07 15:11:16 crc kubenswrapper[4943]: I0307 15:11:16.915389 4943 generic.go:334] "Generic (PLEG): container finished" podID="16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e" containerID="c1ae345069473e9b091f876427afe14066b3e06b2c88ea08a4ab5d08935084cb" exitCode=0 Mar 07 15:11:16 crc kubenswrapper[4943]: I0307 15:11:16.915599 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ljqll" event={"ID":"16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e","Type":"ContainerDied","Data":"c1ae345069473e9b091f876427afe14066b3e06b2c88ea08a4ab5d08935084cb"} Mar 07 15:11:18 crc kubenswrapper[4943]: I0307 15:11:18.303084 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ljqll" Mar 07 15:11:18 crc kubenswrapper[4943]: I0307 15:11:18.354666 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-ljqll"] Mar 07 15:11:18 crc kubenswrapper[4943]: I0307 15:11:18.354798 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e-scripts\") pod \"16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e\" (UID: \"16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e\") " Mar 07 15:11:18 crc kubenswrapper[4943]: I0307 15:11:18.354953 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e-ring-data-devices\") pod \"16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e\" (UID: \"16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e\") " Mar 07 15:11:18 crc kubenswrapper[4943]: I0307 15:11:18.355039 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e-etc-swift\") pod \"16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e\" (UID: \"16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e\") " Mar 07 15:11:18 crc kubenswrapper[4943]: I0307 15:11:18.355074 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e-swiftconf\") pod \"16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e\" (UID: \"16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e\") " Mar 07 15:11:18 crc kubenswrapper[4943]: I0307 15:11:18.355109 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2w7x\" (UniqueName: \"kubernetes.io/projected/16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e-kube-api-access-x2w7x\") pod \"16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e\" (UID: \"16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e\") " Mar 07 15:11:18 crc kubenswrapper[4943]: I0307 15:11:18.355146 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e-dispersionconf\") pod \"16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e\" (UID: \"16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e\") " Mar 07 15:11:18 crc kubenswrapper[4943]: I0307 15:11:18.355961 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e" (UID: "16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:11:18 crc kubenswrapper[4943]: I0307 15:11:18.356439 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e" (UID: "16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:11:18 crc kubenswrapper[4943]: I0307 15:11:18.361852 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e-kube-api-access-x2w7x" (OuterVolumeSpecName: "kube-api-access-x2w7x") pod "16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e" (UID: "16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e"). InnerVolumeSpecName "kube-api-access-x2w7x". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:11:18 crc kubenswrapper[4943]: I0307 15:11:18.365666 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-ljqll"] Mar 07 15:11:18 crc kubenswrapper[4943]: I0307 15:11:18.379711 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e-scripts" (OuterVolumeSpecName: "scripts") pod "16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e" (UID: "16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:11:18 crc kubenswrapper[4943]: I0307 15:11:18.388567 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e" (UID: "16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:11:18 crc kubenswrapper[4943]: I0307 15:11:18.402776 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e" (UID: "16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:11:18 crc kubenswrapper[4943]: I0307 15:11:18.457288 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:18 crc kubenswrapper[4943]: I0307 15:11:18.457330 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:18 crc kubenswrapper[4943]: I0307 15:11:18.457344 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2w7x\" (UniqueName: \"kubernetes.io/projected/16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e-kube-api-access-x2w7x\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:18 crc kubenswrapper[4943]: I0307 15:11:18.457358 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:18 crc kubenswrapper[4943]: I0307 15:11:18.457370 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:18 crc kubenswrapper[4943]: I0307 15:11:18.457381 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:18 crc kubenswrapper[4943]: I0307 15:11:18.768245 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e" path="/var/lib/kubelet/pods/16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e/volumes" Mar 07 15:11:18 crc kubenswrapper[4943]: I0307 15:11:18.941454 4943 scope.go:117] "RemoveContainer" containerID="c1ae345069473e9b091f876427afe14066b3e06b2c88ea08a4ab5d08935084cb" Mar 07 15:11:18 crc kubenswrapper[4943]: I0307 15:11:18.941536 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ljqll" Mar 07 15:11:19 crc kubenswrapper[4943]: I0307 15:11:19.558199 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-qfgh8"] Mar 07 15:11:19 crc kubenswrapper[4943]: E0307 15:11:19.558645 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e" containerName="swift-ring-rebalance" Mar 07 15:11:19 crc kubenswrapper[4943]: I0307 15:11:19.558666 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e" containerName="swift-ring-rebalance" Mar 07 15:11:19 crc kubenswrapper[4943]: I0307 15:11:19.559009 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="16c9b8b2-b0ca-4a6e-970f-db25ca1aca3e" containerName="swift-ring-rebalance" Mar 07 15:11:19 crc kubenswrapper[4943]: I0307 15:11:19.559766 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qfgh8" Mar 07 15:11:19 crc kubenswrapper[4943]: I0307 15:11:19.562299 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:11:19 crc kubenswrapper[4943]: I0307 15:11:19.564719 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:11:19 crc kubenswrapper[4943]: I0307 15:11:19.571891 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-qfgh8"] Mar 07 15:11:19 crc kubenswrapper[4943]: I0307 15:11:19.674101 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/553ae130-0272-41b6-8a03-40c8141a73eb-dispersionconf\") pod \"swift-ring-rebalance-debug-qfgh8\" (UID: \"553ae130-0272-41b6-8a03-40c8141a73eb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qfgh8" Mar 07 15:11:19 crc kubenswrapper[4943]: I0307 15:11:19.674214 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7g7sz\" (UniqueName: \"kubernetes.io/projected/553ae130-0272-41b6-8a03-40c8141a73eb-kube-api-access-7g7sz\") pod \"swift-ring-rebalance-debug-qfgh8\" (UID: \"553ae130-0272-41b6-8a03-40c8141a73eb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qfgh8" Mar 07 15:11:19 crc kubenswrapper[4943]: I0307 15:11:19.674433 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/553ae130-0272-41b6-8a03-40c8141a73eb-swiftconf\") pod \"swift-ring-rebalance-debug-qfgh8\" (UID: \"553ae130-0272-41b6-8a03-40c8141a73eb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qfgh8" Mar 07 15:11:19 crc kubenswrapper[4943]: I0307 15:11:19.674477 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/553ae130-0272-41b6-8a03-40c8141a73eb-ring-data-devices\") pod \"swift-ring-rebalance-debug-qfgh8\" (UID: \"553ae130-0272-41b6-8a03-40c8141a73eb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qfgh8" Mar 07 15:11:19 crc kubenswrapper[4943]: I0307 15:11:19.674584 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/553ae130-0272-41b6-8a03-40c8141a73eb-scripts\") pod \"swift-ring-rebalance-debug-qfgh8\" (UID: \"553ae130-0272-41b6-8a03-40c8141a73eb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qfgh8" Mar 07 15:11:19 crc kubenswrapper[4943]: I0307 15:11:19.674629 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/553ae130-0272-41b6-8a03-40c8141a73eb-etc-swift\") pod \"swift-ring-rebalance-debug-qfgh8\" (UID: \"553ae130-0272-41b6-8a03-40c8141a73eb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qfgh8" Mar 07 15:11:19 crc kubenswrapper[4943]: I0307 15:11:19.776205 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/553ae130-0272-41b6-8a03-40c8141a73eb-dispersionconf\") pod \"swift-ring-rebalance-debug-qfgh8\" (UID: \"553ae130-0272-41b6-8a03-40c8141a73eb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qfgh8" Mar 07 15:11:19 crc kubenswrapper[4943]: I0307 15:11:19.776321 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7g7sz\" (UniqueName: \"kubernetes.io/projected/553ae130-0272-41b6-8a03-40c8141a73eb-kube-api-access-7g7sz\") pod \"swift-ring-rebalance-debug-qfgh8\" (UID: \"553ae130-0272-41b6-8a03-40c8141a73eb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qfgh8" Mar 07 15:11:19 crc kubenswrapper[4943]: I0307 15:11:19.776387 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/553ae130-0272-41b6-8a03-40c8141a73eb-swiftconf\") pod \"swift-ring-rebalance-debug-qfgh8\" (UID: \"553ae130-0272-41b6-8a03-40c8141a73eb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qfgh8" Mar 07 15:11:19 crc kubenswrapper[4943]: I0307 15:11:19.776433 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/553ae130-0272-41b6-8a03-40c8141a73eb-ring-data-devices\") pod \"swift-ring-rebalance-debug-qfgh8\" (UID: \"553ae130-0272-41b6-8a03-40c8141a73eb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qfgh8" Mar 07 15:11:19 crc kubenswrapper[4943]: I0307 15:11:19.776548 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/553ae130-0272-41b6-8a03-40c8141a73eb-scripts\") pod \"swift-ring-rebalance-debug-qfgh8\" (UID: \"553ae130-0272-41b6-8a03-40c8141a73eb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qfgh8" Mar 07 15:11:19 crc kubenswrapper[4943]: I0307 15:11:19.776604 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/553ae130-0272-41b6-8a03-40c8141a73eb-etc-swift\") pod \"swift-ring-rebalance-debug-qfgh8\" (UID: \"553ae130-0272-41b6-8a03-40c8141a73eb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qfgh8" Mar 07 15:11:19 crc kubenswrapper[4943]: I0307 15:11:19.777495 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/553ae130-0272-41b6-8a03-40c8141a73eb-etc-swift\") pod \"swift-ring-rebalance-debug-qfgh8\" (UID: \"553ae130-0272-41b6-8a03-40c8141a73eb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qfgh8" Mar 07 15:11:19 crc kubenswrapper[4943]: I0307 15:11:19.778070 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/553ae130-0272-41b6-8a03-40c8141a73eb-ring-data-devices\") pod \"swift-ring-rebalance-debug-qfgh8\" (UID: \"553ae130-0272-41b6-8a03-40c8141a73eb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qfgh8" Mar 07 15:11:19 crc kubenswrapper[4943]: I0307 15:11:19.779220 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/553ae130-0272-41b6-8a03-40c8141a73eb-scripts\") pod \"swift-ring-rebalance-debug-qfgh8\" (UID: \"553ae130-0272-41b6-8a03-40c8141a73eb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qfgh8" Mar 07 15:11:19 crc kubenswrapper[4943]: I0307 15:11:19.791819 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/553ae130-0272-41b6-8a03-40c8141a73eb-swiftconf\") pod \"swift-ring-rebalance-debug-qfgh8\" (UID: \"553ae130-0272-41b6-8a03-40c8141a73eb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qfgh8" Mar 07 15:11:19 crc kubenswrapper[4943]: I0307 15:11:19.793433 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/553ae130-0272-41b6-8a03-40c8141a73eb-dispersionconf\") pod \"swift-ring-rebalance-debug-qfgh8\" (UID: \"553ae130-0272-41b6-8a03-40c8141a73eb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qfgh8" Mar 07 15:11:19 crc kubenswrapper[4943]: I0307 15:11:19.808164 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7g7sz\" (UniqueName: \"kubernetes.io/projected/553ae130-0272-41b6-8a03-40c8141a73eb-kube-api-access-7g7sz\") pod \"swift-ring-rebalance-debug-qfgh8\" (UID: \"553ae130-0272-41b6-8a03-40c8141a73eb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qfgh8" Mar 07 15:11:19 crc kubenswrapper[4943]: I0307 15:11:19.914793 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qfgh8" Mar 07 15:11:20 crc kubenswrapper[4943]: W0307 15:11:20.173513 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod553ae130_0272_41b6_8a03_40c8141a73eb.slice/crio-9bc57b92f088cb9c1aa868d62073cb46914929e2c11e1a8888857103d290b95f WatchSource:0}: Error finding container 9bc57b92f088cb9c1aa868d62073cb46914929e2c11e1a8888857103d290b95f: Status 404 returned error can't find the container with id 9bc57b92f088cb9c1aa868d62073cb46914929e2c11e1a8888857103d290b95f Mar 07 15:11:20 crc kubenswrapper[4943]: I0307 15:11:20.175560 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-qfgh8"] Mar 07 15:11:20 crc kubenswrapper[4943]: I0307 15:11:20.972733 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qfgh8" event={"ID":"553ae130-0272-41b6-8a03-40c8141a73eb","Type":"ContainerStarted","Data":"7f636729abd0912c1a4f9afeb6d60b85afa90aede4d7e0dc6f518232bc393483"} Mar 07 15:11:20 crc kubenswrapper[4943]: I0307 15:11:20.973210 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qfgh8" event={"ID":"553ae130-0272-41b6-8a03-40c8141a73eb","Type":"ContainerStarted","Data":"9bc57b92f088cb9c1aa868d62073cb46914929e2c11e1a8888857103d290b95f"} Mar 07 15:11:21 crc kubenswrapper[4943]: I0307 15:11:21.010425 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qfgh8" podStartSLOduration=2.010400261 podStartE2EDuration="2.010400261s" podCreationTimestamp="2026-03-07 15:11:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:11:20.998308025 +0000 UTC m=+1922.950444573" watchObservedRunningTime="2026-03-07 15:11:21.010400261 +0000 UTC m=+1922.962536789" Mar 07 15:11:21 crc kubenswrapper[4943]: I0307 15:11:21.986985 4943 generic.go:334] "Generic (PLEG): container finished" podID="553ae130-0272-41b6-8a03-40c8141a73eb" containerID="7f636729abd0912c1a4f9afeb6d60b85afa90aede4d7e0dc6f518232bc393483" exitCode=0 Mar 07 15:11:21 crc kubenswrapper[4943]: I0307 15:11:21.988312 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qfgh8" event={"ID":"553ae130-0272-41b6-8a03-40c8141a73eb","Type":"ContainerDied","Data":"7f636729abd0912c1a4f9afeb6d60b85afa90aede4d7e0dc6f518232bc393483"} Mar 07 15:11:23 crc kubenswrapper[4943]: I0307 15:11:23.379373 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qfgh8" Mar 07 15:11:23 crc kubenswrapper[4943]: I0307 15:11:23.433501 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-qfgh8"] Mar 07 15:11:23 crc kubenswrapper[4943]: I0307 15:11:23.433911 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/553ae130-0272-41b6-8a03-40c8141a73eb-etc-swift\") pod \"553ae130-0272-41b6-8a03-40c8141a73eb\" (UID: \"553ae130-0272-41b6-8a03-40c8141a73eb\") " Mar 07 15:11:23 crc kubenswrapper[4943]: I0307 15:11:23.434098 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/553ae130-0272-41b6-8a03-40c8141a73eb-ring-data-devices\") pod \"553ae130-0272-41b6-8a03-40c8141a73eb\" (UID: \"553ae130-0272-41b6-8a03-40c8141a73eb\") " Mar 07 15:11:23 crc kubenswrapper[4943]: I0307 15:11:23.434177 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/553ae130-0272-41b6-8a03-40c8141a73eb-dispersionconf\") pod \"553ae130-0272-41b6-8a03-40c8141a73eb\" (UID: \"553ae130-0272-41b6-8a03-40c8141a73eb\") " Mar 07 15:11:23 crc kubenswrapper[4943]: I0307 15:11:23.434237 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/553ae130-0272-41b6-8a03-40c8141a73eb-swiftconf\") pod \"553ae130-0272-41b6-8a03-40c8141a73eb\" (UID: \"553ae130-0272-41b6-8a03-40c8141a73eb\") " Mar 07 15:11:23 crc kubenswrapper[4943]: I0307 15:11:23.434363 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7g7sz\" (UniqueName: \"kubernetes.io/projected/553ae130-0272-41b6-8a03-40c8141a73eb-kube-api-access-7g7sz\") pod \"553ae130-0272-41b6-8a03-40c8141a73eb\" (UID: \"553ae130-0272-41b6-8a03-40c8141a73eb\") " Mar 07 15:11:23 crc kubenswrapper[4943]: I0307 15:11:23.434496 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/553ae130-0272-41b6-8a03-40c8141a73eb-scripts\") pod \"553ae130-0272-41b6-8a03-40c8141a73eb\" (UID: \"553ae130-0272-41b6-8a03-40c8141a73eb\") " Mar 07 15:11:23 crc kubenswrapper[4943]: I0307 15:11:23.435084 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/553ae130-0272-41b6-8a03-40c8141a73eb-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "553ae130-0272-41b6-8a03-40c8141a73eb" (UID: "553ae130-0272-41b6-8a03-40c8141a73eb"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:11:23 crc kubenswrapper[4943]: I0307 15:11:23.435102 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/553ae130-0272-41b6-8a03-40c8141a73eb-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "553ae130-0272-41b6-8a03-40c8141a73eb" (UID: "553ae130-0272-41b6-8a03-40c8141a73eb"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:11:23 crc kubenswrapper[4943]: I0307 15:11:23.435263 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/553ae130-0272-41b6-8a03-40c8141a73eb-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:23 crc kubenswrapper[4943]: I0307 15:11:23.435283 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/553ae130-0272-41b6-8a03-40c8141a73eb-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:23 crc kubenswrapper[4943]: I0307 15:11:23.442604 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/553ae130-0272-41b6-8a03-40c8141a73eb-kube-api-access-7g7sz" (OuterVolumeSpecName: "kube-api-access-7g7sz") pod "553ae130-0272-41b6-8a03-40c8141a73eb" (UID: "553ae130-0272-41b6-8a03-40c8141a73eb"). InnerVolumeSpecName "kube-api-access-7g7sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:11:23 crc kubenswrapper[4943]: I0307 15:11:23.444335 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-qfgh8"] Mar 07 15:11:23 crc kubenswrapper[4943]: I0307 15:11:23.459491 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/553ae130-0272-41b6-8a03-40c8141a73eb-scripts" (OuterVolumeSpecName: "scripts") pod "553ae130-0272-41b6-8a03-40c8141a73eb" (UID: "553ae130-0272-41b6-8a03-40c8141a73eb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:11:23 crc kubenswrapper[4943]: I0307 15:11:23.472963 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/553ae130-0272-41b6-8a03-40c8141a73eb-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "553ae130-0272-41b6-8a03-40c8141a73eb" (UID: "553ae130-0272-41b6-8a03-40c8141a73eb"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:11:23 crc kubenswrapper[4943]: I0307 15:11:23.482525 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/553ae130-0272-41b6-8a03-40c8141a73eb-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "553ae130-0272-41b6-8a03-40c8141a73eb" (UID: "553ae130-0272-41b6-8a03-40c8141a73eb"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:11:23 crc kubenswrapper[4943]: I0307 15:11:23.536613 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/553ae130-0272-41b6-8a03-40c8141a73eb-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:23 crc kubenswrapper[4943]: I0307 15:11:23.536652 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/553ae130-0272-41b6-8a03-40c8141a73eb-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:23 crc kubenswrapper[4943]: I0307 15:11:23.536673 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7g7sz\" (UniqueName: \"kubernetes.io/projected/553ae130-0272-41b6-8a03-40c8141a73eb-kube-api-access-7g7sz\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:23 crc kubenswrapper[4943]: I0307 15:11:23.536694 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/553ae130-0272-41b6-8a03-40c8141a73eb-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:24 crc kubenswrapper[4943]: I0307 15:11:24.009799 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9bc57b92f088cb9c1aa868d62073cb46914929e2c11e1a8888857103d290b95f" Mar 07 15:11:24 crc kubenswrapper[4943]: I0307 15:11:24.009881 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qfgh8" Mar 07 15:11:24 crc kubenswrapper[4943]: I0307 15:11:24.660050 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-ptsx6"] Mar 07 15:11:24 crc kubenswrapper[4943]: E0307 15:11:24.660341 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="553ae130-0272-41b6-8a03-40c8141a73eb" containerName="swift-ring-rebalance" Mar 07 15:11:24 crc kubenswrapper[4943]: I0307 15:11:24.660353 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="553ae130-0272-41b6-8a03-40c8141a73eb" containerName="swift-ring-rebalance" Mar 07 15:11:24 crc kubenswrapper[4943]: I0307 15:11:24.660504 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="553ae130-0272-41b6-8a03-40c8141a73eb" containerName="swift-ring-rebalance" Mar 07 15:11:24 crc kubenswrapper[4943]: I0307 15:11:24.660991 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ptsx6" Mar 07 15:11:24 crc kubenswrapper[4943]: I0307 15:11:24.664167 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:11:24 crc kubenswrapper[4943]: I0307 15:11:24.664514 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:11:24 crc kubenswrapper[4943]: I0307 15:11:24.680011 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-ptsx6"] Mar 07 15:11:24 crc kubenswrapper[4943]: I0307 15:11:24.757476 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3b5d32e3-4841-4eed-8378-ed01678002d8-etc-swift\") pod \"swift-ring-rebalance-debug-ptsx6\" (UID: \"3b5d32e3-4841-4eed-8378-ed01678002d8\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ptsx6" Mar 07 15:11:24 crc kubenswrapper[4943]: I0307 15:11:24.757654 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3b5d32e3-4841-4eed-8378-ed01678002d8-swiftconf\") pod \"swift-ring-rebalance-debug-ptsx6\" (UID: \"3b5d32e3-4841-4eed-8378-ed01678002d8\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ptsx6" Mar 07 15:11:24 crc kubenswrapper[4943]: I0307 15:11:24.757723 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3b5d32e3-4841-4eed-8378-ed01678002d8-dispersionconf\") pod \"swift-ring-rebalance-debug-ptsx6\" (UID: \"3b5d32e3-4841-4eed-8378-ed01678002d8\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ptsx6" Mar 07 15:11:24 crc kubenswrapper[4943]: I0307 15:11:24.757769 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3b5d32e3-4841-4eed-8378-ed01678002d8-scripts\") pod \"swift-ring-rebalance-debug-ptsx6\" (UID: \"3b5d32e3-4841-4eed-8378-ed01678002d8\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ptsx6" Mar 07 15:11:24 crc kubenswrapper[4943]: I0307 15:11:24.757821 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3b5d32e3-4841-4eed-8378-ed01678002d8-ring-data-devices\") pod \"swift-ring-rebalance-debug-ptsx6\" (UID: \"3b5d32e3-4841-4eed-8378-ed01678002d8\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ptsx6" Mar 07 15:11:24 crc kubenswrapper[4943]: I0307 15:11:24.757850 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6srqr\" (UniqueName: \"kubernetes.io/projected/3b5d32e3-4841-4eed-8378-ed01678002d8-kube-api-access-6srqr\") pod \"swift-ring-rebalance-debug-ptsx6\" (UID: \"3b5d32e3-4841-4eed-8378-ed01678002d8\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ptsx6" Mar 07 15:11:24 crc kubenswrapper[4943]: I0307 15:11:24.764655 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="553ae130-0272-41b6-8a03-40c8141a73eb" path="/var/lib/kubelet/pods/553ae130-0272-41b6-8a03-40c8141a73eb/volumes" Mar 07 15:11:24 crc kubenswrapper[4943]: I0307 15:11:24.859885 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3b5d32e3-4841-4eed-8378-ed01678002d8-etc-swift\") pod \"swift-ring-rebalance-debug-ptsx6\" (UID: \"3b5d32e3-4841-4eed-8378-ed01678002d8\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ptsx6" Mar 07 15:11:24 crc kubenswrapper[4943]: I0307 15:11:24.860046 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3b5d32e3-4841-4eed-8378-ed01678002d8-swiftconf\") pod \"swift-ring-rebalance-debug-ptsx6\" (UID: \"3b5d32e3-4841-4eed-8378-ed01678002d8\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ptsx6" Mar 07 15:11:24 crc kubenswrapper[4943]: I0307 15:11:24.860099 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3b5d32e3-4841-4eed-8378-ed01678002d8-dispersionconf\") pod \"swift-ring-rebalance-debug-ptsx6\" (UID: \"3b5d32e3-4841-4eed-8378-ed01678002d8\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ptsx6" Mar 07 15:11:24 crc kubenswrapper[4943]: I0307 15:11:24.860151 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3b5d32e3-4841-4eed-8378-ed01678002d8-scripts\") pod \"swift-ring-rebalance-debug-ptsx6\" (UID: \"3b5d32e3-4841-4eed-8378-ed01678002d8\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ptsx6" Mar 07 15:11:24 crc kubenswrapper[4943]: I0307 15:11:24.860222 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3b5d32e3-4841-4eed-8378-ed01678002d8-ring-data-devices\") pod \"swift-ring-rebalance-debug-ptsx6\" (UID: \"3b5d32e3-4841-4eed-8378-ed01678002d8\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ptsx6" Mar 07 15:11:24 crc kubenswrapper[4943]: I0307 15:11:24.860261 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6srqr\" (UniqueName: \"kubernetes.io/projected/3b5d32e3-4841-4eed-8378-ed01678002d8-kube-api-access-6srqr\") pod \"swift-ring-rebalance-debug-ptsx6\" (UID: \"3b5d32e3-4841-4eed-8378-ed01678002d8\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ptsx6" Mar 07 15:11:24 crc kubenswrapper[4943]: I0307 15:11:24.860681 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3b5d32e3-4841-4eed-8378-ed01678002d8-etc-swift\") pod \"swift-ring-rebalance-debug-ptsx6\" (UID: \"3b5d32e3-4841-4eed-8378-ed01678002d8\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ptsx6" Mar 07 15:11:24 crc kubenswrapper[4943]: I0307 15:11:24.862624 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3b5d32e3-4841-4eed-8378-ed01678002d8-ring-data-devices\") pod \"swift-ring-rebalance-debug-ptsx6\" (UID: \"3b5d32e3-4841-4eed-8378-ed01678002d8\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ptsx6" Mar 07 15:11:24 crc kubenswrapper[4943]: I0307 15:11:24.863582 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3b5d32e3-4841-4eed-8378-ed01678002d8-scripts\") pod \"swift-ring-rebalance-debug-ptsx6\" (UID: \"3b5d32e3-4841-4eed-8378-ed01678002d8\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ptsx6" Mar 07 15:11:24 crc kubenswrapper[4943]: I0307 15:11:24.873402 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3b5d32e3-4841-4eed-8378-ed01678002d8-swiftconf\") pod \"swift-ring-rebalance-debug-ptsx6\" (UID: \"3b5d32e3-4841-4eed-8378-ed01678002d8\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ptsx6" Mar 07 15:11:24 crc kubenswrapper[4943]: I0307 15:11:24.874056 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3b5d32e3-4841-4eed-8378-ed01678002d8-dispersionconf\") pod \"swift-ring-rebalance-debug-ptsx6\" (UID: \"3b5d32e3-4841-4eed-8378-ed01678002d8\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ptsx6" Mar 07 15:11:24 crc kubenswrapper[4943]: I0307 15:11:24.880528 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6srqr\" (UniqueName: \"kubernetes.io/projected/3b5d32e3-4841-4eed-8378-ed01678002d8-kube-api-access-6srqr\") pod \"swift-ring-rebalance-debug-ptsx6\" (UID: \"3b5d32e3-4841-4eed-8378-ed01678002d8\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ptsx6" Mar 07 15:11:25 crc kubenswrapper[4943]: I0307 15:11:25.026452 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ptsx6" Mar 07 15:11:25 crc kubenswrapper[4943]: I0307 15:11:25.548460 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-ptsx6"] Mar 07 15:11:26 crc kubenswrapper[4943]: I0307 15:11:26.033217 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ptsx6" event={"ID":"3b5d32e3-4841-4eed-8378-ed01678002d8","Type":"ContainerStarted","Data":"905315b500be35ba2642d7e0175befa98c18a39b2909e90fbe3decf077957769"} Mar 07 15:11:26 crc kubenswrapper[4943]: I0307 15:11:26.033598 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ptsx6" event={"ID":"3b5d32e3-4841-4eed-8378-ed01678002d8","Type":"ContainerStarted","Data":"01f35ac23e606b84c1f90d6e18605095dc01816284a6a70a31c4b6bc1780951a"} Mar 07 15:11:26 crc kubenswrapper[4943]: I0307 15:11:26.070761 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ptsx6" podStartSLOduration=2.070737398 podStartE2EDuration="2.070737398s" podCreationTimestamp="2026-03-07 15:11:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:11:26.062686531 +0000 UTC m=+1928.014823059" watchObservedRunningTime="2026-03-07 15:11:26.070737398 +0000 UTC m=+1928.022873936" Mar 07 15:11:27 crc kubenswrapper[4943]: E0307 15:11:27.024014 4943 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3b5d32e3_4841_4eed_8378_ed01678002d8.slice/crio-905315b500be35ba2642d7e0175befa98c18a39b2909e90fbe3decf077957769.scope\": RecentStats: unable to find data in memory cache]" Mar 07 15:11:27 crc kubenswrapper[4943]: I0307 15:11:27.054401 4943 generic.go:334] "Generic (PLEG): container finished" podID="3b5d32e3-4841-4eed-8378-ed01678002d8" containerID="905315b500be35ba2642d7e0175befa98c18a39b2909e90fbe3decf077957769" exitCode=0 Mar 07 15:11:27 crc kubenswrapper[4943]: I0307 15:11:27.054469 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ptsx6" event={"ID":"3b5d32e3-4841-4eed-8378-ed01678002d8","Type":"ContainerDied","Data":"905315b500be35ba2642d7e0175befa98c18a39b2909e90fbe3decf077957769"} Mar 07 15:11:28 crc kubenswrapper[4943]: I0307 15:11:28.456012 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ptsx6" Mar 07 15:11:28 crc kubenswrapper[4943]: I0307 15:11:28.497002 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-ptsx6"] Mar 07 15:11:28 crc kubenswrapper[4943]: I0307 15:11:28.508650 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-ptsx6"] Mar 07 15:11:28 crc kubenswrapper[4943]: I0307 15:11:28.624024 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3b5d32e3-4841-4eed-8378-ed01678002d8-etc-swift\") pod \"3b5d32e3-4841-4eed-8378-ed01678002d8\" (UID: \"3b5d32e3-4841-4eed-8378-ed01678002d8\") " Mar 07 15:11:28 crc kubenswrapper[4943]: I0307 15:11:28.624140 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3b5d32e3-4841-4eed-8378-ed01678002d8-ring-data-devices\") pod \"3b5d32e3-4841-4eed-8378-ed01678002d8\" (UID: \"3b5d32e3-4841-4eed-8378-ed01678002d8\") " Mar 07 15:11:28 crc kubenswrapper[4943]: I0307 15:11:28.624190 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3b5d32e3-4841-4eed-8378-ed01678002d8-dispersionconf\") pod \"3b5d32e3-4841-4eed-8378-ed01678002d8\" (UID: \"3b5d32e3-4841-4eed-8378-ed01678002d8\") " Mar 07 15:11:28 crc kubenswrapper[4943]: I0307 15:11:28.624216 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3b5d32e3-4841-4eed-8378-ed01678002d8-scripts\") pod \"3b5d32e3-4841-4eed-8378-ed01678002d8\" (UID: \"3b5d32e3-4841-4eed-8378-ed01678002d8\") " Mar 07 15:11:28 crc kubenswrapper[4943]: I0307 15:11:28.624263 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3b5d32e3-4841-4eed-8378-ed01678002d8-swiftconf\") pod \"3b5d32e3-4841-4eed-8378-ed01678002d8\" (UID: \"3b5d32e3-4841-4eed-8378-ed01678002d8\") " Mar 07 15:11:28 crc kubenswrapper[4943]: I0307 15:11:28.624286 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6srqr\" (UniqueName: \"kubernetes.io/projected/3b5d32e3-4841-4eed-8378-ed01678002d8-kube-api-access-6srqr\") pod \"3b5d32e3-4841-4eed-8378-ed01678002d8\" (UID: \"3b5d32e3-4841-4eed-8378-ed01678002d8\") " Mar 07 15:11:28 crc kubenswrapper[4943]: I0307 15:11:28.624913 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b5d32e3-4841-4eed-8378-ed01678002d8-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "3b5d32e3-4841-4eed-8378-ed01678002d8" (UID: "3b5d32e3-4841-4eed-8378-ed01678002d8"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:11:28 crc kubenswrapper[4943]: I0307 15:11:28.625185 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b5d32e3-4841-4eed-8378-ed01678002d8-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "3b5d32e3-4841-4eed-8378-ed01678002d8" (UID: "3b5d32e3-4841-4eed-8378-ed01678002d8"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:11:28 crc kubenswrapper[4943]: I0307 15:11:28.632995 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b5d32e3-4841-4eed-8378-ed01678002d8-kube-api-access-6srqr" (OuterVolumeSpecName: "kube-api-access-6srqr") pod "3b5d32e3-4841-4eed-8378-ed01678002d8" (UID: "3b5d32e3-4841-4eed-8378-ed01678002d8"). InnerVolumeSpecName "kube-api-access-6srqr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:11:28 crc kubenswrapper[4943]: I0307 15:11:28.648585 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b5d32e3-4841-4eed-8378-ed01678002d8-scripts" (OuterVolumeSpecName: "scripts") pod "3b5d32e3-4841-4eed-8378-ed01678002d8" (UID: "3b5d32e3-4841-4eed-8378-ed01678002d8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:11:28 crc kubenswrapper[4943]: I0307 15:11:28.655051 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b5d32e3-4841-4eed-8378-ed01678002d8-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "3b5d32e3-4841-4eed-8378-ed01678002d8" (UID: "3b5d32e3-4841-4eed-8378-ed01678002d8"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:11:28 crc kubenswrapper[4943]: I0307 15:11:28.667989 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b5d32e3-4841-4eed-8378-ed01678002d8-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "3b5d32e3-4841-4eed-8378-ed01678002d8" (UID: "3b5d32e3-4841-4eed-8378-ed01678002d8"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:11:28 crc kubenswrapper[4943]: I0307 15:11:28.726097 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3b5d32e3-4841-4eed-8378-ed01678002d8-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:28 crc kubenswrapper[4943]: I0307 15:11:28.726149 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3b5d32e3-4841-4eed-8378-ed01678002d8-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:28 crc kubenswrapper[4943]: I0307 15:11:28.726170 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3b5d32e3-4841-4eed-8378-ed01678002d8-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:28 crc kubenswrapper[4943]: I0307 15:11:28.726188 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3b5d32e3-4841-4eed-8378-ed01678002d8-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:28 crc kubenswrapper[4943]: I0307 15:11:28.726205 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6srqr\" (UniqueName: \"kubernetes.io/projected/3b5d32e3-4841-4eed-8378-ed01678002d8-kube-api-access-6srqr\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:28 crc kubenswrapper[4943]: I0307 15:11:28.726222 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3b5d32e3-4841-4eed-8378-ed01678002d8-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:28 crc kubenswrapper[4943]: I0307 15:11:28.784133 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b5d32e3-4841-4eed-8378-ed01678002d8" path="/var/lib/kubelet/pods/3b5d32e3-4841-4eed-8378-ed01678002d8/volumes" Mar 07 15:11:29 crc kubenswrapper[4943]: I0307 15:11:29.076922 4943 scope.go:117] "RemoveContainer" containerID="905315b500be35ba2642d7e0175befa98c18a39b2909e90fbe3decf077957769" Mar 07 15:11:29 crc kubenswrapper[4943]: I0307 15:11:29.076990 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ptsx6" Mar 07 15:11:29 crc kubenswrapper[4943]: I0307 15:11:29.661796 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-qm27r"] Mar 07 15:11:29 crc kubenswrapper[4943]: E0307 15:11:29.662567 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b5d32e3-4841-4eed-8378-ed01678002d8" containerName="swift-ring-rebalance" Mar 07 15:11:29 crc kubenswrapper[4943]: I0307 15:11:29.662588 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b5d32e3-4841-4eed-8378-ed01678002d8" containerName="swift-ring-rebalance" Mar 07 15:11:29 crc kubenswrapper[4943]: I0307 15:11:29.662793 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b5d32e3-4841-4eed-8378-ed01678002d8" containerName="swift-ring-rebalance" Mar 07 15:11:29 crc kubenswrapper[4943]: I0307 15:11:29.663456 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qm27r" Mar 07 15:11:29 crc kubenswrapper[4943]: I0307 15:11:29.667909 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:11:29 crc kubenswrapper[4943]: I0307 15:11:29.667909 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:11:29 crc kubenswrapper[4943]: I0307 15:11:29.685316 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-qm27r"] Mar 07 15:11:29 crc kubenswrapper[4943]: I0307 15:11:29.845427 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a0d422aa-335e-42bd-9c48-e7c9f2f0389d-dispersionconf\") pod \"swift-ring-rebalance-debug-qm27r\" (UID: \"a0d422aa-335e-42bd-9c48-e7c9f2f0389d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qm27r" Mar 07 15:11:29 crc kubenswrapper[4943]: I0307 15:11:29.845512 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a0d422aa-335e-42bd-9c48-e7c9f2f0389d-scripts\") pod \"swift-ring-rebalance-debug-qm27r\" (UID: \"a0d422aa-335e-42bd-9c48-e7c9f2f0389d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qm27r" Mar 07 15:11:29 crc kubenswrapper[4943]: I0307 15:11:29.845543 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a0d422aa-335e-42bd-9c48-e7c9f2f0389d-swiftconf\") pod \"swift-ring-rebalance-debug-qm27r\" (UID: \"a0d422aa-335e-42bd-9c48-e7c9f2f0389d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qm27r" Mar 07 15:11:29 crc kubenswrapper[4943]: I0307 15:11:29.845561 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a0d422aa-335e-42bd-9c48-e7c9f2f0389d-etc-swift\") pod \"swift-ring-rebalance-debug-qm27r\" (UID: \"a0d422aa-335e-42bd-9c48-e7c9f2f0389d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qm27r" Mar 07 15:11:29 crc kubenswrapper[4943]: I0307 15:11:29.845612 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jq9mz\" (UniqueName: \"kubernetes.io/projected/a0d422aa-335e-42bd-9c48-e7c9f2f0389d-kube-api-access-jq9mz\") pod \"swift-ring-rebalance-debug-qm27r\" (UID: \"a0d422aa-335e-42bd-9c48-e7c9f2f0389d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qm27r" Mar 07 15:11:29 crc kubenswrapper[4943]: I0307 15:11:29.845635 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a0d422aa-335e-42bd-9c48-e7c9f2f0389d-ring-data-devices\") pod \"swift-ring-rebalance-debug-qm27r\" (UID: \"a0d422aa-335e-42bd-9c48-e7c9f2f0389d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qm27r" Mar 07 15:11:29 crc kubenswrapper[4943]: I0307 15:11:29.947048 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a0d422aa-335e-42bd-9c48-e7c9f2f0389d-dispersionconf\") pod \"swift-ring-rebalance-debug-qm27r\" (UID: \"a0d422aa-335e-42bd-9c48-e7c9f2f0389d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qm27r" Mar 07 15:11:29 crc kubenswrapper[4943]: I0307 15:11:29.947093 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a0d422aa-335e-42bd-9c48-e7c9f2f0389d-scripts\") pod \"swift-ring-rebalance-debug-qm27r\" (UID: \"a0d422aa-335e-42bd-9c48-e7c9f2f0389d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qm27r" Mar 07 15:11:29 crc kubenswrapper[4943]: I0307 15:11:29.947121 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a0d422aa-335e-42bd-9c48-e7c9f2f0389d-swiftconf\") pod \"swift-ring-rebalance-debug-qm27r\" (UID: \"a0d422aa-335e-42bd-9c48-e7c9f2f0389d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qm27r" Mar 07 15:11:29 crc kubenswrapper[4943]: I0307 15:11:29.947140 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a0d422aa-335e-42bd-9c48-e7c9f2f0389d-etc-swift\") pod \"swift-ring-rebalance-debug-qm27r\" (UID: \"a0d422aa-335e-42bd-9c48-e7c9f2f0389d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qm27r" Mar 07 15:11:29 crc kubenswrapper[4943]: I0307 15:11:29.947196 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jq9mz\" (UniqueName: \"kubernetes.io/projected/a0d422aa-335e-42bd-9c48-e7c9f2f0389d-kube-api-access-jq9mz\") pod \"swift-ring-rebalance-debug-qm27r\" (UID: \"a0d422aa-335e-42bd-9c48-e7c9f2f0389d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qm27r" Mar 07 15:11:29 crc kubenswrapper[4943]: I0307 15:11:29.947218 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a0d422aa-335e-42bd-9c48-e7c9f2f0389d-ring-data-devices\") pod \"swift-ring-rebalance-debug-qm27r\" (UID: \"a0d422aa-335e-42bd-9c48-e7c9f2f0389d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qm27r" Mar 07 15:11:29 crc kubenswrapper[4943]: I0307 15:11:29.947847 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a0d422aa-335e-42bd-9c48-e7c9f2f0389d-ring-data-devices\") pod \"swift-ring-rebalance-debug-qm27r\" (UID: \"a0d422aa-335e-42bd-9c48-e7c9f2f0389d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qm27r" Mar 07 15:11:29 crc kubenswrapper[4943]: I0307 15:11:29.947903 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a0d422aa-335e-42bd-9c48-e7c9f2f0389d-etc-swift\") pod \"swift-ring-rebalance-debug-qm27r\" (UID: \"a0d422aa-335e-42bd-9c48-e7c9f2f0389d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qm27r" Mar 07 15:11:29 crc kubenswrapper[4943]: I0307 15:11:29.948392 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a0d422aa-335e-42bd-9c48-e7c9f2f0389d-scripts\") pod \"swift-ring-rebalance-debug-qm27r\" (UID: \"a0d422aa-335e-42bd-9c48-e7c9f2f0389d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qm27r" Mar 07 15:11:29 crc kubenswrapper[4943]: I0307 15:11:29.952084 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a0d422aa-335e-42bd-9c48-e7c9f2f0389d-swiftconf\") pod \"swift-ring-rebalance-debug-qm27r\" (UID: \"a0d422aa-335e-42bd-9c48-e7c9f2f0389d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qm27r" Mar 07 15:11:29 crc kubenswrapper[4943]: I0307 15:11:29.952469 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a0d422aa-335e-42bd-9c48-e7c9f2f0389d-dispersionconf\") pod \"swift-ring-rebalance-debug-qm27r\" (UID: \"a0d422aa-335e-42bd-9c48-e7c9f2f0389d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qm27r" Mar 07 15:11:29 crc kubenswrapper[4943]: I0307 15:11:29.968701 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jq9mz\" (UniqueName: \"kubernetes.io/projected/a0d422aa-335e-42bd-9c48-e7c9f2f0389d-kube-api-access-jq9mz\") pod \"swift-ring-rebalance-debug-qm27r\" (UID: \"a0d422aa-335e-42bd-9c48-e7c9f2f0389d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qm27r" Mar 07 15:11:30 crc kubenswrapper[4943]: I0307 15:11:30.023411 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qm27r" Mar 07 15:11:30 crc kubenswrapper[4943]: I0307 15:11:30.248888 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-qm27r"] Mar 07 15:11:30 crc kubenswrapper[4943]: W0307 15:11:30.252583 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda0d422aa_335e_42bd_9c48_e7c9f2f0389d.slice/crio-1c17f698f511d975f9a456e4ee5b3787d38c860ac4429065131838f4d28605ea WatchSource:0}: Error finding container 1c17f698f511d975f9a456e4ee5b3787d38c860ac4429065131838f4d28605ea: Status 404 returned error can't find the container with id 1c17f698f511d975f9a456e4ee5b3787d38c860ac4429065131838f4d28605ea Mar 07 15:11:31 crc kubenswrapper[4943]: I0307 15:11:31.096024 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qm27r" event={"ID":"a0d422aa-335e-42bd-9c48-e7c9f2f0389d","Type":"ContainerStarted","Data":"001de98844014410af24877aeeb22718020bffccafe294976e9da66e536d469c"} Mar 07 15:11:31 crc kubenswrapper[4943]: I0307 15:11:31.096370 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qm27r" event={"ID":"a0d422aa-335e-42bd-9c48-e7c9f2f0389d","Type":"ContainerStarted","Data":"1c17f698f511d975f9a456e4ee5b3787d38c860ac4429065131838f4d28605ea"} Mar 07 15:11:31 crc kubenswrapper[4943]: I0307 15:11:31.120209 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qm27r" podStartSLOduration=2.120190698 podStartE2EDuration="2.120190698s" podCreationTimestamp="2026-03-07 15:11:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:11:31.116700352 +0000 UTC m=+1933.068836890" watchObservedRunningTime="2026-03-07 15:11:31.120190698 +0000 UTC m=+1933.072327196" Mar 07 15:11:32 crc kubenswrapper[4943]: I0307 15:11:32.110029 4943 generic.go:334] "Generic (PLEG): container finished" podID="a0d422aa-335e-42bd-9c48-e7c9f2f0389d" containerID="001de98844014410af24877aeeb22718020bffccafe294976e9da66e536d469c" exitCode=0 Mar 07 15:11:32 crc kubenswrapper[4943]: I0307 15:11:32.110109 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qm27r" event={"ID":"a0d422aa-335e-42bd-9c48-e7c9f2f0389d","Type":"ContainerDied","Data":"001de98844014410af24877aeeb22718020bffccafe294976e9da66e536d469c"} Mar 07 15:11:33 crc kubenswrapper[4943]: I0307 15:11:33.536552 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qm27r" Mar 07 15:11:33 crc kubenswrapper[4943]: I0307 15:11:33.586824 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-qm27r"] Mar 07 15:11:33 crc kubenswrapper[4943]: I0307 15:11:33.593843 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-qm27r"] Mar 07 15:11:33 crc kubenswrapper[4943]: I0307 15:11:33.613608 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a0d422aa-335e-42bd-9c48-e7c9f2f0389d-dispersionconf\") pod \"a0d422aa-335e-42bd-9c48-e7c9f2f0389d\" (UID: \"a0d422aa-335e-42bd-9c48-e7c9f2f0389d\") " Mar 07 15:11:33 crc kubenswrapper[4943]: I0307 15:11:33.613720 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a0d422aa-335e-42bd-9c48-e7c9f2f0389d-swiftconf\") pod \"a0d422aa-335e-42bd-9c48-e7c9f2f0389d\" (UID: \"a0d422aa-335e-42bd-9c48-e7c9f2f0389d\") " Mar 07 15:11:33 crc kubenswrapper[4943]: I0307 15:11:33.613772 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a0d422aa-335e-42bd-9c48-e7c9f2f0389d-etc-swift\") pod \"a0d422aa-335e-42bd-9c48-e7c9f2f0389d\" (UID: \"a0d422aa-335e-42bd-9c48-e7c9f2f0389d\") " Mar 07 15:11:33 crc kubenswrapper[4943]: I0307 15:11:33.613847 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a0d422aa-335e-42bd-9c48-e7c9f2f0389d-ring-data-devices\") pod \"a0d422aa-335e-42bd-9c48-e7c9f2f0389d\" (UID: \"a0d422aa-335e-42bd-9c48-e7c9f2f0389d\") " Mar 07 15:11:33 crc kubenswrapper[4943]: I0307 15:11:33.613876 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jq9mz\" (UniqueName: \"kubernetes.io/projected/a0d422aa-335e-42bd-9c48-e7c9f2f0389d-kube-api-access-jq9mz\") pod \"a0d422aa-335e-42bd-9c48-e7c9f2f0389d\" (UID: \"a0d422aa-335e-42bd-9c48-e7c9f2f0389d\") " Mar 07 15:11:33 crc kubenswrapper[4943]: I0307 15:11:33.613961 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a0d422aa-335e-42bd-9c48-e7c9f2f0389d-scripts\") pod \"a0d422aa-335e-42bd-9c48-e7c9f2f0389d\" (UID: \"a0d422aa-335e-42bd-9c48-e7c9f2f0389d\") " Mar 07 15:11:33 crc kubenswrapper[4943]: I0307 15:11:33.615848 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a0d422aa-335e-42bd-9c48-e7c9f2f0389d-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "a0d422aa-335e-42bd-9c48-e7c9f2f0389d" (UID: "a0d422aa-335e-42bd-9c48-e7c9f2f0389d"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:11:33 crc kubenswrapper[4943]: I0307 15:11:33.616224 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0d422aa-335e-42bd-9c48-e7c9f2f0389d-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "a0d422aa-335e-42bd-9c48-e7c9f2f0389d" (UID: "a0d422aa-335e-42bd-9c48-e7c9f2f0389d"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:11:33 crc kubenswrapper[4943]: I0307 15:11:33.625421 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0d422aa-335e-42bd-9c48-e7c9f2f0389d-kube-api-access-jq9mz" (OuterVolumeSpecName: "kube-api-access-jq9mz") pod "a0d422aa-335e-42bd-9c48-e7c9f2f0389d" (UID: "a0d422aa-335e-42bd-9c48-e7c9f2f0389d"). InnerVolumeSpecName "kube-api-access-jq9mz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:11:33 crc kubenswrapper[4943]: I0307 15:11:33.640541 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a0d422aa-335e-42bd-9c48-e7c9f2f0389d-scripts" (OuterVolumeSpecName: "scripts") pod "a0d422aa-335e-42bd-9c48-e7c9f2f0389d" (UID: "a0d422aa-335e-42bd-9c48-e7c9f2f0389d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:11:33 crc kubenswrapper[4943]: I0307 15:11:33.643105 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0d422aa-335e-42bd-9c48-e7c9f2f0389d-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "a0d422aa-335e-42bd-9c48-e7c9f2f0389d" (UID: "a0d422aa-335e-42bd-9c48-e7c9f2f0389d"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:11:33 crc kubenswrapper[4943]: I0307 15:11:33.648082 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0d422aa-335e-42bd-9c48-e7c9f2f0389d-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "a0d422aa-335e-42bd-9c48-e7c9f2f0389d" (UID: "a0d422aa-335e-42bd-9c48-e7c9f2f0389d"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:11:33 crc kubenswrapper[4943]: I0307 15:11:33.715416 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a0d422aa-335e-42bd-9c48-e7c9f2f0389d-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:33 crc kubenswrapper[4943]: I0307 15:11:33.715446 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jq9mz\" (UniqueName: \"kubernetes.io/projected/a0d422aa-335e-42bd-9c48-e7c9f2f0389d-kube-api-access-jq9mz\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:33 crc kubenswrapper[4943]: I0307 15:11:33.715456 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a0d422aa-335e-42bd-9c48-e7c9f2f0389d-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:33 crc kubenswrapper[4943]: I0307 15:11:33.715466 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a0d422aa-335e-42bd-9c48-e7c9f2f0389d-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:33 crc kubenswrapper[4943]: I0307 15:11:33.715474 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a0d422aa-335e-42bd-9c48-e7c9f2f0389d-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:33 crc kubenswrapper[4943]: I0307 15:11:33.715482 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a0d422aa-335e-42bd-9c48-e7c9f2f0389d-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:34 crc kubenswrapper[4943]: I0307 15:11:34.148169 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1c17f698f511d975f9a456e4ee5b3787d38c860ac4429065131838f4d28605ea" Mar 07 15:11:34 crc kubenswrapper[4943]: I0307 15:11:34.148547 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qm27r" Mar 07 15:11:34 crc kubenswrapper[4943]: I0307 15:11:34.772381 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0d422aa-335e-42bd-9c48-e7c9f2f0389d" path="/var/lib/kubelet/pods/a0d422aa-335e-42bd-9c48-e7c9f2f0389d/volumes" Mar 07 15:11:34 crc kubenswrapper[4943]: I0307 15:11:34.774090 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-5zz9h"] Mar 07 15:11:34 crc kubenswrapper[4943]: E0307 15:11:34.774733 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0d422aa-335e-42bd-9c48-e7c9f2f0389d" containerName="swift-ring-rebalance" Mar 07 15:11:34 crc kubenswrapper[4943]: I0307 15:11:34.774777 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0d422aa-335e-42bd-9c48-e7c9f2f0389d" containerName="swift-ring-rebalance" Mar 07 15:11:34 crc kubenswrapper[4943]: I0307 15:11:34.775165 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0d422aa-335e-42bd-9c48-e7c9f2f0389d" containerName="swift-ring-rebalance" Mar 07 15:11:34 crc kubenswrapper[4943]: I0307 15:11:34.776142 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5zz9h" Mar 07 15:11:34 crc kubenswrapper[4943]: I0307 15:11:34.781447 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:11:34 crc kubenswrapper[4943]: I0307 15:11:34.782141 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:11:34 crc kubenswrapper[4943]: I0307 15:11:34.806454 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-5zz9h"] Mar 07 15:11:34 crc kubenswrapper[4943]: I0307 15:11:34.932671 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/043e2d86-b3cb-4b39-9ad0-adf4575a369f-scripts\") pod \"swift-ring-rebalance-debug-5zz9h\" (UID: \"043e2d86-b3cb-4b39-9ad0-adf4575a369f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5zz9h" Mar 07 15:11:34 crc kubenswrapper[4943]: I0307 15:11:34.932764 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/043e2d86-b3cb-4b39-9ad0-adf4575a369f-etc-swift\") pod \"swift-ring-rebalance-debug-5zz9h\" (UID: \"043e2d86-b3cb-4b39-9ad0-adf4575a369f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5zz9h" Mar 07 15:11:34 crc kubenswrapper[4943]: I0307 15:11:34.932829 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6pzw2\" (UniqueName: \"kubernetes.io/projected/043e2d86-b3cb-4b39-9ad0-adf4575a369f-kube-api-access-6pzw2\") pod \"swift-ring-rebalance-debug-5zz9h\" (UID: \"043e2d86-b3cb-4b39-9ad0-adf4575a369f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5zz9h" Mar 07 15:11:34 crc kubenswrapper[4943]: I0307 15:11:34.932865 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/043e2d86-b3cb-4b39-9ad0-adf4575a369f-ring-data-devices\") pod \"swift-ring-rebalance-debug-5zz9h\" (UID: \"043e2d86-b3cb-4b39-9ad0-adf4575a369f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5zz9h" Mar 07 15:11:34 crc kubenswrapper[4943]: I0307 15:11:34.933075 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/043e2d86-b3cb-4b39-9ad0-adf4575a369f-dispersionconf\") pod \"swift-ring-rebalance-debug-5zz9h\" (UID: \"043e2d86-b3cb-4b39-9ad0-adf4575a369f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5zz9h" Mar 07 15:11:34 crc kubenswrapper[4943]: I0307 15:11:34.933131 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/043e2d86-b3cb-4b39-9ad0-adf4575a369f-swiftconf\") pod \"swift-ring-rebalance-debug-5zz9h\" (UID: \"043e2d86-b3cb-4b39-9ad0-adf4575a369f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5zz9h" Mar 07 15:11:35 crc kubenswrapper[4943]: I0307 15:11:35.035149 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/043e2d86-b3cb-4b39-9ad0-adf4575a369f-dispersionconf\") pod \"swift-ring-rebalance-debug-5zz9h\" (UID: \"043e2d86-b3cb-4b39-9ad0-adf4575a369f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5zz9h" Mar 07 15:11:35 crc kubenswrapper[4943]: I0307 15:11:35.035207 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/043e2d86-b3cb-4b39-9ad0-adf4575a369f-swiftconf\") pod \"swift-ring-rebalance-debug-5zz9h\" (UID: \"043e2d86-b3cb-4b39-9ad0-adf4575a369f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5zz9h" Mar 07 15:11:35 crc kubenswrapper[4943]: I0307 15:11:35.035290 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/043e2d86-b3cb-4b39-9ad0-adf4575a369f-scripts\") pod \"swift-ring-rebalance-debug-5zz9h\" (UID: \"043e2d86-b3cb-4b39-9ad0-adf4575a369f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5zz9h" Mar 07 15:11:35 crc kubenswrapper[4943]: I0307 15:11:35.035326 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/043e2d86-b3cb-4b39-9ad0-adf4575a369f-etc-swift\") pod \"swift-ring-rebalance-debug-5zz9h\" (UID: \"043e2d86-b3cb-4b39-9ad0-adf4575a369f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5zz9h" Mar 07 15:11:35 crc kubenswrapper[4943]: I0307 15:11:35.035361 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/043e2d86-b3cb-4b39-9ad0-adf4575a369f-ring-data-devices\") pod \"swift-ring-rebalance-debug-5zz9h\" (UID: \"043e2d86-b3cb-4b39-9ad0-adf4575a369f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5zz9h" Mar 07 15:11:35 crc kubenswrapper[4943]: I0307 15:11:35.035386 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6pzw2\" (UniqueName: \"kubernetes.io/projected/043e2d86-b3cb-4b39-9ad0-adf4575a369f-kube-api-access-6pzw2\") pod \"swift-ring-rebalance-debug-5zz9h\" (UID: \"043e2d86-b3cb-4b39-9ad0-adf4575a369f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5zz9h" Mar 07 15:11:35 crc kubenswrapper[4943]: I0307 15:11:35.036307 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/043e2d86-b3cb-4b39-9ad0-adf4575a369f-etc-swift\") pod \"swift-ring-rebalance-debug-5zz9h\" (UID: \"043e2d86-b3cb-4b39-9ad0-adf4575a369f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5zz9h" Mar 07 15:11:35 crc kubenswrapper[4943]: I0307 15:11:35.036868 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/043e2d86-b3cb-4b39-9ad0-adf4575a369f-scripts\") pod \"swift-ring-rebalance-debug-5zz9h\" (UID: \"043e2d86-b3cb-4b39-9ad0-adf4575a369f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5zz9h" Mar 07 15:11:35 crc kubenswrapper[4943]: I0307 15:11:35.037001 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/043e2d86-b3cb-4b39-9ad0-adf4575a369f-ring-data-devices\") pod \"swift-ring-rebalance-debug-5zz9h\" (UID: \"043e2d86-b3cb-4b39-9ad0-adf4575a369f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5zz9h" Mar 07 15:11:35 crc kubenswrapper[4943]: I0307 15:11:35.045769 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/043e2d86-b3cb-4b39-9ad0-adf4575a369f-dispersionconf\") pod \"swift-ring-rebalance-debug-5zz9h\" (UID: \"043e2d86-b3cb-4b39-9ad0-adf4575a369f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5zz9h" Mar 07 15:11:35 crc kubenswrapper[4943]: I0307 15:11:35.046862 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/043e2d86-b3cb-4b39-9ad0-adf4575a369f-swiftconf\") pod \"swift-ring-rebalance-debug-5zz9h\" (UID: \"043e2d86-b3cb-4b39-9ad0-adf4575a369f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5zz9h" Mar 07 15:11:35 crc kubenswrapper[4943]: I0307 15:11:35.058878 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6pzw2\" (UniqueName: \"kubernetes.io/projected/043e2d86-b3cb-4b39-9ad0-adf4575a369f-kube-api-access-6pzw2\") pod \"swift-ring-rebalance-debug-5zz9h\" (UID: \"043e2d86-b3cb-4b39-9ad0-adf4575a369f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5zz9h" Mar 07 15:11:35 crc kubenswrapper[4943]: I0307 15:11:35.105279 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5zz9h" Mar 07 15:11:35 crc kubenswrapper[4943]: I0307 15:11:35.386364 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-5zz9h"] Mar 07 15:11:35 crc kubenswrapper[4943]: W0307 15:11:35.399191 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod043e2d86_b3cb_4b39_9ad0_adf4575a369f.slice/crio-a66b66518867eec5ba039d00492f03b5697ee39ee0a5939031d3d0eeb7c04e42 WatchSource:0}: Error finding container a66b66518867eec5ba039d00492f03b5697ee39ee0a5939031d3d0eeb7c04e42: Status 404 returned error can't find the container with id a66b66518867eec5ba039d00492f03b5697ee39ee0a5939031d3d0eeb7c04e42 Mar 07 15:11:36 crc kubenswrapper[4943]: I0307 15:11:36.173449 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5zz9h" event={"ID":"043e2d86-b3cb-4b39-9ad0-adf4575a369f","Type":"ContainerStarted","Data":"1d8f6816c1adb3578c6bd1057566658b416c86e04f874459714f33df073d3e26"} Mar 07 15:11:36 crc kubenswrapper[4943]: I0307 15:11:36.173506 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5zz9h" event={"ID":"043e2d86-b3cb-4b39-9ad0-adf4575a369f","Type":"ContainerStarted","Data":"a66b66518867eec5ba039d00492f03b5697ee39ee0a5939031d3d0eeb7c04e42"} Mar 07 15:11:36 crc kubenswrapper[4943]: I0307 15:11:36.205617 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5zz9h" podStartSLOduration=2.2055861 podStartE2EDuration="2.2055861s" podCreationTimestamp="2026-03-07 15:11:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:11:36.2031638 +0000 UTC m=+1938.155300298" watchObservedRunningTime="2026-03-07 15:11:36.2055861 +0000 UTC m=+1938.157722628" Mar 07 15:11:37 crc kubenswrapper[4943]: I0307 15:11:37.185726 4943 generic.go:334] "Generic (PLEG): container finished" podID="043e2d86-b3cb-4b39-9ad0-adf4575a369f" containerID="1d8f6816c1adb3578c6bd1057566658b416c86e04f874459714f33df073d3e26" exitCode=0 Mar 07 15:11:37 crc kubenswrapper[4943]: I0307 15:11:37.186075 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5zz9h" event={"ID":"043e2d86-b3cb-4b39-9ad0-adf4575a369f","Type":"ContainerDied","Data":"1d8f6816c1adb3578c6bd1057566658b416c86e04f874459714f33df073d3e26"} Mar 07 15:11:38 crc kubenswrapper[4943]: I0307 15:11:38.620404 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5zz9h" Mar 07 15:11:38 crc kubenswrapper[4943]: I0307 15:11:38.659363 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-5zz9h"] Mar 07 15:11:38 crc kubenswrapper[4943]: I0307 15:11:38.664171 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-5zz9h"] Mar 07 15:11:38 crc kubenswrapper[4943]: I0307 15:11:38.692092 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/043e2d86-b3cb-4b39-9ad0-adf4575a369f-dispersionconf\") pod \"043e2d86-b3cb-4b39-9ad0-adf4575a369f\" (UID: \"043e2d86-b3cb-4b39-9ad0-adf4575a369f\") " Mar 07 15:11:38 crc kubenswrapper[4943]: I0307 15:11:38.692422 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6pzw2\" (UniqueName: \"kubernetes.io/projected/043e2d86-b3cb-4b39-9ad0-adf4575a369f-kube-api-access-6pzw2\") pod \"043e2d86-b3cb-4b39-9ad0-adf4575a369f\" (UID: \"043e2d86-b3cb-4b39-9ad0-adf4575a369f\") " Mar 07 15:11:38 crc kubenswrapper[4943]: I0307 15:11:38.692478 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/043e2d86-b3cb-4b39-9ad0-adf4575a369f-etc-swift\") pod \"043e2d86-b3cb-4b39-9ad0-adf4575a369f\" (UID: \"043e2d86-b3cb-4b39-9ad0-adf4575a369f\") " Mar 07 15:11:38 crc kubenswrapper[4943]: I0307 15:11:38.692512 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/043e2d86-b3cb-4b39-9ad0-adf4575a369f-scripts\") pod \"043e2d86-b3cb-4b39-9ad0-adf4575a369f\" (UID: \"043e2d86-b3cb-4b39-9ad0-adf4575a369f\") " Mar 07 15:11:38 crc kubenswrapper[4943]: I0307 15:11:38.692543 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/043e2d86-b3cb-4b39-9ad0-adf4575a369f-ring-data-devices\") pod \"043e2d86-b3cb-4b39-9ad0-adf4575a369f\" (UID: \"043e2d86-b3cb-4b39-9ad0-adf4575a369f\") " Mar 07 15:11:38 crc kubenswrapper[4943]: I0307 15:11:38.692564 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/043e2d86-b3cb-4b39-9ad0-adf4575a369f-swiftconf\") pod \"043e2d86-b3cb-4b39-9ad0-adf4575a369f\" (UID: \"043e2d86-b3cb-4b39-9ad0-adf4575a369f\") " Mar 07 15:11:38 crc kubenswrapper[4943]: I0307 15:11:38.693779 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/043e2d86-b3cb-4b39-9ad0-adf4575a369f-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "043e2d86-b3cb-4b39-9ad0-adf4575a369f" (UID: "043e2d86-b3cb-4b39-9ad0-adf4575a369f"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:11:38 crc kubenswrapper[4943]: I0307 15:11:38.694034 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/043e2d86-b3cb-4b39-9ad0-adf4575a369f-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "043e2d86-b3cb-4b39-9ad0-adf4575a369f" (UID: "043e2d86-b3cb-4b39-9ad0-adf4575a369f"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:11:38 crc kubenswrapper[4943]: I0307 15:11:38.697011 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/043e2d86-b3cb-4b39-9ad0-adf4575a369f-kube-api-access-6pzw2" (OuterVolumeSpecName: "kube-api-access-6pzw2") pod "043e2d86-b3cb-4b39-9ad0-adf4575a369f" (UID: "043e2d86-b3cb-4b39-9ad0-adf4575a369f"). InnerVolumeSpecName "kube-api-access-6pzw2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:11:38 crc kubenswrapper[4943]: I0307 15:11:38.711144 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/043e2d86-b3cb-4b39-9ad0-adf4575a369f-scripts" (OuterVolumeSpecName: "scripts") pod "043e2d86-b3cb-4b39-9ad0-adf4575a369f" (UID: "043e2d86-b3cb-4b39-9ad0-adf4575a369f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:11:38 crc kubenswrapper[4943]: I0307 15:11:38.721175 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/043e2d86-b3cb-4b39-9ad0-adf4575a369f-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "043e2d86-b3cb-4b39-9ad0-adf4575a369f" (UID: "043e2d86-b3cb-4b39-9ad0-adf4575a369f"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:11:38 crc kubenswrapper[4943]: I0307 15:11:38.732140 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/043e2d86-b3cb-4b39-9ad0-adf4575a369f-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "043e2d86-b3cb-4b39-9ad0-adf4575a369f" (UID: "043e2d86-b3cb-4b39-9ad0-adf4575a369f"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:11:38 crc kubenswrapper[4943]: I0307 15:11:38.774997 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="043e2d86-b3cb-4b39-9ad0-adf4575a369f" path="/var/lib/kubelet/pods/043e2d86-b3cb-4b39-9ad0-adf4575a369f/volumes" Mar 07 15:11:38 crc kubenswrapper[4943]: I0307 15:11:38.794274 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6pzw2\" (UniqueName: \"kubernetes.io/projected/043e2d86-b3cb-4b39-9ad0-adf4575a369f-kube-api-access-6pzw2\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:38 crc kubenswrapper[4943]: I0307 15:11:38.794333 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/043e2d86-b3cb-4b39-9ad0-adf4575a369f-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:38 crc kubenswrapper[4943]: I0307 15:11:38.794359 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/043e2d86-b3cb-4b39-9ad0-adf4575a369f-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:38 crc kubenswrapper[4943]: I0307 15:11:38.794386 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/043e2d86-b3cb-4b39-9ad0-adf4575a369f-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:38 crc kubenswrapper[4943]: I0307 15:11:38.794408 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/043e2d86-b3cb-4b39-9ad0-adf4575a369f-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:38 crc kubenswrapper[4943]: I0307 15:11:38.794433 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/043e2d86-b3cb-4b39-9ad0-adf4575a369f-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:39 crc kubenswrapper[4943]: I0307 15:11:39.210293 4943 scope.go:117] "RemoveContainer" containerID="1d8f6816c1adb3578c6bd1057566658b416c86e04f874459714f33df073d3e26" Mar 07 15:11:39 crc kubenswrapper[4943]: I0307 15:11:39.210433 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5zz9h" Mar 07 15:11:39 crc kubenswrapper[4943]: I0307 15:11:39.863064 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-xvmbf"] Mar 07 15:11:39 crc kubenswrapper[4943]: E0307 15:11:39.864327 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="043e2d86-b3cb-4b39-9ad0-adf4575a369f" containerName="swift-ring-rebalance" Mar 07 15:11:39 crc kubenswrapper[4943]: I0307 15:11:39.864500 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="043e2d86-b3cb-4b39-9ad0-adf4575a369f" containerName="swift-ring-rebalance" Mar 07 15:11:39 crc kubenswrapper[4943]: I0307 15:11:39.864871 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="043e2d86-b3cb-4b39-9ad0-adf4575a369f" containerName="swift-ring-rebalance" Mar 07 15:11:39 crc kubenswrapper[4943]: I0307 15:11:39.865556 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-xvmbf" Mar 07 15:11:39 crc kubenswrapper[4943]: I0307 15:11:39.867798 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:11:39 crc kubenswrapper[4943]: I0307 15:11:39.868069 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:11:39 crc kubenswrapper[4943]: I0307 15:11:39.875365 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-xvmbf"] Mar 07 15:11:39 crc kubenswrapper[4943]: I0307 15:11:39.909745 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1410559c-dd49-4a70-b66b-667bff6f1122-scripts\") pod \"swift-ring-rebalance-debug-xvmbf\" (UID: \"1410559c-dd49-4a70-b66b-667bff6f1122\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xvmbf" Mar 07 15:11:39 crc kubenswrapper[4943]: I0307 15:11:39.909852 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1410559c-dd49-4a70-b66b-667bff6f1122-etc-swift\") pod \"swift-ring-rebalance-debug-xvmbf\" (UID: \"1410559c-dd49-4a70-b66b-667bff6f1122\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xvmbf" Mar 07 15:11:39 crc kubenswrapper[4943]: I0307 15:11:39.909909 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hsc9\" (UniqueName: \"kubernetes.io/projected/1410559c-dd49-4a70-b66b-667bff6f1122-kube-api-access-4hsc9\") pod \"swift-ring-rebalance-debug-xvmbf\" (UID: \"1410559c-dd49-4a70-b66b-667bff6f1122\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xvmbf" Mar 07 15:11:39 crc kubenswrapper[4943]: I0307 15:11:39.910129 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1410559c-dd49-4a70-b66b-667bff6f1122-swiftconf\") pod \"swift-ring-rebalance-debug-xvmbf\" (UID: \"1410559c-dd49-4a70-b66b-667bff6f1122\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xvmbf" Mar 07 15:11:39 crc kubenswrapper[4943]: I0307 15:11:39.910303 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1410559c-dd49-4a70-b66b-667bff6f1122-ring-data-devices\") pod \"swift-ring-rebalance-debug-xvmbf\" (UID: \"1410559c-dd49-4a70-b66b-667bff6f1122\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xvmbf" Mar 07 15:11:39 crc kubenswrapper[4943]: I0307 15:11:39.910384 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1410559c-dd49-4a70-b66b-667bff6f1122-dispersionconf\") pod \"swift-ring-rebalance-debug-xvmbf\" (UID: \"1410559c-dd49-4a70-b66b-667bff6f1122\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xvmbf" Mar 07 15:11:40 crc kubenswrapper[4943]: I0307 15:11:40.011917 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1410559c-dd49-4a70-b66b-667bff6f1122-dispersionconf\") pod \"swift-ring-rebalance-debug-xvmbf\" (UID: \"1410559c-dd49-4a70-b66b-667bff6f1122\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xvmbf" Mar 07 15:11:40 crc kubenswrapper[4943]: I0307 15:11:40.012011 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1410559c-dd49-4a70-b66b-667bff6f1122-scripts\") pod \"swift-ring-rebalance-debug-xvmbf\" (UID: \"1410559c-dd49-4a70-b66b-667bff6f1122\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xvmbf" Mar 07 15:11:40 crc kubenswrapper[4943]: I0307 15:11:40.012071 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1410559c-dd49-4a70-b66b-667bff6f1122-etc-swift\") pod \"swift-ring-rebalance-debug-xvmbf\" (UID: \"1410559c-dd49-4a70-b66b-667bff6f1122\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xvmbf" Mar 07 15:11:40 crc kubenswrapper[4943]: I0307 15:11:40.012130 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hsc9\" (UniqueName: \"kubernetes.io/projected/1410559c-dd49-4a70-b66b-667bff6f1122-kube-api-access-4hsc9\") pod \"swift-ring-rebalance-debug-xvmbf\" (UID: \"1410559c-dd49-4a70-b66b-667bff6f1122\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xvmbf" Mar 07 15:11:40 crc kubenswrapper[4943]: I0307 15:11:40.012210 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1410559c-dd49-4a70-b66b-667bff6f1122-swiftconf\") pod \"swift-ring-rebalance-debug-xvmbf\" (UID: \"1410559c-dd49-4a70-b66b-667bff6f1122\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xvmbf" Mar 07 15:11:40 crc kubenswrapper[4943]: I0307 15:11:40.012266 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1410559c-dd49-4a70-b66b-667bff6f1122-ring-data-devices\") pod \"swift-ring-rebalance-debug-xvmbf\" (UID: \"1410559c-dd49-4a70-b66b-667bff6f1122\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xvmbf" Mar 07 15:11:40 crc kubenswrapper[4943]: I0307 15:11:40.012611 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1410559c-dd49-4a70-b66b-667bff6f1122-etc-swift\") pod \"swift-ring-rebalance-debug-xvmbf\" (UID: \"1410559c-dd49-4a70-b66b-667bff6f1122\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xvmbf" Mar 07 15:11:40 crc kubenswrapper[4943]: I0307 15:11:40.015521 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1410559c-dd49-4a70-b66b-667bff6f1122-ring-data-devices\") pod \"swift-ring-rebalance-debug-xvmbf\" (UID: \"1410559c-dd49-4a70-b66b-667bff6f1122\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xvmbf" Mar 07 15:11:40 crc kubenswrapper[4943]: I0307 15:11:40.016064 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1410559c-dd49-4a70-b66b-667bff6f1122-scripts\") pod \"swift-ring-rebalance-debug-xvmbf\" (UID: \"1410559c-dd49-4a70-b66b-667bff6f1122\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xvmbf" Mar 07 15:11:40 crc kubenswrapper[4943]: I0307 15:11:40.017147 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1410559c-dd49-4a70-b66b-667bff6f1122-dispersionconf\") pod \"swift-ring-rebalance-debug-xvmbf\" (UID: \"1410559c-dd49-4a70-b66b-667bff6f1122\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xvmbf" Mar 07 15:11:40 crc kubenswrapper[4943]: I0307 15:11:40.018829 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1410559c-dd49-4a70-b66b-667bff6f1122-swiftconf\") pod \"swift-ring-rebalance-debug-xvmbf\" (UID: \"1410559c-dd49-4a70-b66b-667bff6f1122\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xvmbf" Mar 07 15:11:40 crc kubenswrapper[4943]: I0307 15:11:40.042028 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hsc9\" (UniqueName: \"kubernetes.io/projected/1410559c-dd49-4a70-b66b-667bff6f1122-kube-api-access-4hsc9\") pod \"swift-ring-rebalance-debug-xvmbf\" (UID: \"1410559c-dd49-4a70-b66b-667bff6f1122\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xvmbf" Mar 07 15:11:40 crc kubenswrapper[4943]: I0307 15:11:40.193690 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-xvmbf" Mar 07 15:11:40 crc kubenswrapper[4943]: I0307 15:11:40.681108 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-xvmbf"] Mar 07 15:11:40 crc kubenswrapper[4943]: W0307 15:11:40.691359 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1410559c_dd49_4a70_b66b_667bff6f1122.slice/crio-cb2dcb58f6cd2c76d8b03d08eabb14ce9a2b24e48a389db1a3ddd168d7eb051e WatchSource:0}: Error finding container cb2dcb58f6cd2c76d8b03d08eabb14ce9a2b24e48a389db1a3ddd168d7eb051e: Status 404 returned error can't find the container with id cb2dcb58f6cd2c76d8b03d08eabb14ce9a2b24e48a389db1a3ddd168d7eb051e Mar 07 15:11:41 crc kubenswrapper[4943]: I0307 15:11:41.236231 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-xvmbf" event={"ID":"1410559c-dd49-4a70-b66b-667bff6f1122","Type":"ContainerStarted","Data":"0ef227a2835e6837b378f6818163dba489e1445ccb0365b2e4f3711ce07e264d"} Mar 07 15:11:41 crc kubenswrapper[4943]: I0307 15:11:41.236702 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-xvmbf" event={"ID":"1410559c-dd49-4a70-b66b-667bff6f1122","Type":"ContainerStarted","Data":"cb2dcb58f6cd2c76d8b03d08eabb14ce9a2b24e48a389db1a3ddd168d7eb051e"} Mar 07 15:11:41 crc kubenswrapper[4943]: I0307 15:11:41.271711 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-xvmbf" podStartSLOduration=2.271685767 podStartE2EDuration="2.271685767s" podCreationTimestamp="2026-03-07 15:11:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:11:41.263676121 +0000 UTC m=+1943.215812629" watchObservedRunningTime="2026-03-07 15:11:41.271685767 +0000 UTC m=+1943.223822305" Mar 07 15:11:42 crc kubenswrapper[4943]: I0307 15:11:42.253256 4943 generic.go:334] "Generic (PLEG): container finished" podID="1410559c-dd49-4a70-b66b-667bff6f1122" containerID="0ef227a2835e6837b378f6818163dba489e1445ccb0365b2e4f3711ce07e264d" exitCode=0 Mar 07 15:11:42 crc kubenswrapper[4943]: I0307 15:11:42.253327 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-xvmbf" event={"ID":"1410559c-dd49-4a70-b66b-667bff6f1122","Type":"ContainerDied","Data":"0ef227a2835e6837b378f6818163dba489e1445ccb0365b2e4f3711ce07e264d"} Mar 07 15:11:43 crc kubenswrapper[4943]: I0307 15:11:43.644792 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-xvmbf" Mar 07 15:11:43 crc kubenswrapper[4943]: I0307 15:11:43.687251 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-xvmbf"] Mar 07 15:11:43 crc kubenswrapper[4943]: I0307 15:11:43.694478 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-xvmbf"] Mar 07 15:11:43 crc kubenswrapper[4943]: I0307 15:11:43.781920 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1410559c-dd49-4a70-b66b-667bff6f1122-etc-swift\") pod \"1410559c-dd49-4a70-b66b-667bff6f1122\" (UID: \"1410559c-dd49-4a70-b66b-667bff6f1122\") " Mar 07 15:11:43 crc kubenswrapper[4943]: I0307 15:11:43.782003 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1410559c-dd49-4a70-b66b-667bff6f1122-ring-data-devices\") pod \"1410559c-dd49-4a70-b66b-667bff6f1122\" (UID: \"1410559c-dd49-4a70-b66b-667bff6f1122\") " Mar 07 15:11:43 crc kubenswrapper[4943]: I0307 15:11:43.782105 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1410559c-dd49-4a70-b66b-667bff6f1122-swiftconf\") pod \"1410559c-dd49-4a70-b66b-667bff6f1122\" (UID: \"1410559c-dd49-4a70-b66b-667bff6f1122\") " Mar 07 15:11:43 crc kubenswrapper[4943]: I0307 15:11:43.782227 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hsc9\" (UniqueName: \"kubernetes.io/projected/1410559c-dd49-4a70-b66b-667bff6f1122-kube-api-access-4hsc9\") pod \"1410559c-dd49-4a70-b66b-667bff6f1122\" (UID: \"1410559c-dd49-4a70-b66b-667bff6f1122\") " Mar 07 15:11:43 crc kubenswrapper[4943]: I0307 15:11:43.782264 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1410559c-dd49-4a70-b66b-667bff6f1122-scripts\") pod \"1410559c-dd49-4a70-b66b-667bff6f1122\" (UID: \"1410559c-dd49-4a70-b66b-667bff6f1122\") " Mar 07 15:11:43 crc kubenswrapper[4943]: I0307 15:11:43.782318 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1410559c-dd49-4a70-b66b-667bff6f1122-dispersionconf\") pod \"1410559c-dd49-4a70-b66b-667bff6f1122\" (UID: \"1410559c-dd49-4a70-b66b-667bff6f1122\") " Mar 07 15:11:43 crc kubenswrapper[4943]: I0307 15:11:43.783282 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1410559c-dd49-4a70-b66b-667bff6f1122-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "1410559c-dd49-4a70-b66b-667bff6f1122" (UID: "1410559c-dd49-4a70-b66b-667bff6f1122"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:11:43 crc kubenswrapper[4943]: I0307 15:11:43.783759 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1410559c-dd49-4a70-b66b-667bff6f1122-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "1410559c-dd49-4a70-b66b-667bff6f1122" (UID: "1410559c-dd49-4a70-b66b-667bff6f1122"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:11:43 crc kubenswrapper[4943]: I0307 15:11:43.788913 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1410559c-dd49-4a70-b66b-667bff6f1122-kube-api-access-4hsc9" (OuterVolumeSpecName: "kube-api-access-4hsc9") pod "1410559c-dd49-4a70-b66b-667bff6f1122" (UID: "1410559c-dd49-4a70-b66b-667bff6f1122"). InnerVolumeSpecName "kube-api-access-4hsc9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:11:43 crc kubenswrapper[4943]: I0307 15:11:43.808439 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1410559c-dd49-4a70-b66b-667bff6f1122-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "1410559c-dd49-4a70-b66b-667bff6f1122" (UID: "1410559c-dd49-4a70-b66b-667bff6f1122"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:11:43 crc kubenswrapper[4943]: I0307 15:11:43.820825 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1410559c-dd49-4a70-b66b-667bff6f1122-scripts" (OuterVolumeSpecName: "scripts") pod "1410559c-dd49-4a70-b66b-667bff6f1122" (UID: "1410559c-dd49-4a70-b66b-667bff6f1122"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:11:43 crc kubenswrapper[4943]: I0307 15:11:43.820882 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1410559c-dd49-4a70-b66b-667bff6f1122-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "1410559c-dd49-4a70-b66b-667bff6f1122" (UID: "1410559c-dd49-4a70-b66b-667bff6f1122"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:11:43 crc kubenswrapper[4943]: I0307 15:11:43.883959 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1410559c-dd49-4a70-b66b-667bff6f1122-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:43 crc kubenswrapper[4943]: I0307 15:11:43.883996 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1410559c-dd49-4a70-b66b-667bff6f1122-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:43 crc kubenswrapper[4943]: I0307 15:11:43.884012 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1410559c-dd49-4a70-b66b-667bff6f1122-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:43 crc kubenswrapper[4943]: I0307 15:11:43.884023 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hsc9\" (UniqueName: \"kubernetes.io/projected/1410559c-dd49-4a70-b66b-667bff6f1122-kube-api-access-4hsc9\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:43 crc kubenswrapper[4943]: I0307 15:11:43.884044 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1410559c-dd49-4a70-b66b-667bff6f1122-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:43 crc kubenswrapper[4943]: I0307 15:11:43.884055 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1410559c-dd49-4a70-b66b-667bff6f1122-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:44 crc kubenswrapper[4943]: I0307 15:11:44.274250 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb2dcb58f6cd2c76d8b03d08eabb14ce9a2b24e48a389db1a3ddd168d7eb051e" Mar 07 15:11:44 crc kubenswrapper[4943]: I0307 15:11:44.274369 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-xvmbf" Mar 07 15:11:44 crc kubenswrapper[4943]: I0307 15:11:44.769893 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1410559c-dd49-4a70-b66b-667bff6f1122" path="/var/lib/kubelet/pods/1410559c-dd49-4a70-b66b-667bff6f1122/volumes" Mar 07 15:11:44 crc kubenswrapper[4943]: I0307 15:11:44.923214 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-lgbbh"] Mar 07 15:11:44 crc kubenswrapper[4943]: E0307 15:11:44.924443 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1410559c-dd49-4a70-b66b-667bff6f1122" containerName="swift-ring-rebalance" Mar 07 15:11:44 crc kubenswrapper[4943]: I0307 15:11:44.924481 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="1410559c-dd49-4a70-b66b-667bff6f1122" containerName="swift-ring-rebalance" Mar 07 15:11:44 crc kubenswrapper[4943]: I0307 15:11:44.924966 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="1410559c-dd49-4a70-b66b-667bff6f1122" containerName="swift-ring-rebalance" Mar 07 15:11:44 crc kubenswrapper[4943]: I0307 15:11:44.925799 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-lgbbh" Mar 07 15:11:44 crc kubenswrapper[4943]: I0307 15:11:44.929871 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:11:44 crc kubenswrapper[4943]: I0307 15:11:44.932632 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:11:44 crc kubenswrapper[4943]: I0307 15:11:44.941512 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-lgbbh"] Mar 07 15:11:45 crc kubenswrapper[4943]: I0307 15:11:45.006560 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ac0cbf27-35f8-4646-b18b-a3416757472c-etc-swift\") pod \"swift-ring-rebalance-debug-lgbbh\" (UID: \"ac0cbf27-35f8-4646-b18b-a3416757472c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-lgbbh" Mar 07 15:11:45 crc kubenswrapper[4943]: I0307 15:11:45.006635 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcp6x\" (UniqueName: \"kubernetes.io/projected/ac0cbf27-35f8-4646-b18b-a3416757472c-kube-api-access-rcp6x\") pod \"swift-ring-rebalance-debug-lgbbh\" (UID: \"ac0cbf27-35f8-4646-b18b-a3416757472c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-lgbbh" Mar 07 15:11:45 crc kubenswrapper[4943]: I0307 15:11:45.006712 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ac0cbf27-35f8-4646-b18b-a3416757472c-ring-data-devices\") pod \"swift-ring-rebalance-debug-lgbbh\" (UID: \"ac0cbf27-35f8-4646-b18b-a3416757472c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-lgbbh" Mar 07 15:11:45 crc kubenswrapper[4943]: I0307 15:11:45.006745 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ac0cbf27-35f8-4646-b18b-a3416757472c-swiftconf\") pod \"swift-ring-rebalance-debug-lgbbh\" (UID: \"ac0cbf27-35f8-4646-b18b-a3416757472c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-lgbbh" Mar 07 15:11:45 crc kubenswrapper[4943]: I0307 15:11:45.006790 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ac0cbf27-35f8-4646-b18b-a3416757472c-scripts\") pod \"swift-ring-rebalance-debug-lgbbh\" (UID: \"ac0cbf27-35f8-4646-b18b-a3416757472c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-lgbbh" Mar 07 15:11:45 crc kubenswrapper[4943]: I0307 15:11:45.006876 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ac0cbf27-35f8-4646-b18b-a3416757472c-dispersionconf\") pod \"swift-ring-rebalance-debug-lgbbh\" (UID: \"ac0cbf27-35f8-4646-b18b-a3416757472c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-lgbbh" Mar 07 15:11:45 crc kubenswrapper[4943]: I0307 15:11:45.108423 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ac0cbf27-35f8-4646-b18b-a3416757472c-scripts\") pod \"swift-ring-rebalance-debug-lgbbh\" (UID: \"ac0cbf27-35f8-4646-b18b-a3416757472c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-lgbbh" Mar 07 15:11:45 crc kubenswrapper[4943]: I0307 15:11:45.109089 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ac0cbf27-35f8-4646-b18b-a3416757472c-dispersionconf\") pod \"swift-ring-rebalance-debug-lgbbh\" (UID: \"ac0cbf27-35f8-4646-b18b-a3416757472c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-lgbbh" Mar 07 15:11:45 crc kubenswrapper[4943]: I0307 15:11:45.109327 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ac0cbf27-35f8-4646-b18b-a3416757472c-etc-swift\") pod \"swift-ring-rebalance-debug-lgbbh\" (UID: \"ac0cbf27-35f8-4646-b18b-a3416757472c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-lgbbh" Mar 07 15:11:45 crc kubenswrapper[4943]: I0307 15:11:45.109512 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcp6x\" (UniqueName: \"kubernetes.io/projected/ac0cbf27-35f8-4646-b18b-a3416757472c-kube-api-access-rcp6x\") pod \"swift-ring-rebalance-debug-lgbbh\" (UID: \"ac0cbf27-35f8-4646-b18b-a3416757472c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-lgbbh" Mar 07 15:11:45 crc kubenswrapper[4943]: I0307 15:11:45.109714 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ac0cbf27-35f8-4646-b18b-a3416757472c-ring-data-devices\") pod \"swift-ring-rebalance-debug-lgbbh\" (UID: \"ac0cbf27-35f8-4646-b18b-a3416757472c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-lgbbh" Mar 07 15:11:45 crc kubenswrapper[4943]: I0307 15:11:45.109913 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ac0cbf27-35f8-4646-b18b-a3416757472c-swiftconf\") pod \"swift-ring-rebalance-debug-lgbbh\" (UID: \"ac0cbf27-35f8-4646-b18b-a3416757472c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-lgbbh" Mar 07 15:11:45 crc kubenswrapper[4943]: I0307 15:11:45.109843 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ac0cbf27-35f8-4646-b18b-a3416757472c-etc-swift\") pod \"swift-ring-rebalance-debug-lgbbh\" (UID: \"ac0cbf27-35f8-4646-b18b-a3416757472c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-lgbbh" Mar 07 15:11:45 crc kubenswrapper[4943]: I0307 15:11:45.109528 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ac0cbf27-35f8-4646-b18b-a3416757472c-scripts\") pod \"swift-ring-rebalance-debug-lgbbh\" (UID: \"ac0cbf27-35f8-4646-b18b-a3416757472c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-lgbbh" Mar 07 15:11:45 crc kubenswrapper[4943]: I0307 15:11:45.110512 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ac0cbf27-35f8-4646-b18b-a3416757472c-ring-data-devices\") pod \"swift-ring-rebalance-debug-lgbbh\" (UID: \"ac0cbf27-35f8-4646-b18b-a3416757472c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-lgbbh" Mar 07 15:11:45 crc kubenswrapper[4943]: I0307 15:11:45.114138 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ac0cbf27-35f8-4646-b18b-a3416757472c-dispersionconf\") pod \"swift-ring-rebalance-debug-lgbbh\" (UID: \"ac0cbf27-35f8-4646-b18b-a3416757472c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-lgbbh" Mar 07 15:11:45 crc kubenswrapper[4943]: I0307 15:11:45.123321 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ac0cbf27-35f8-4646-b18b-a3416757472c-swiftconf\") pod \"swift-ring-rebalance-debug-lgbbh\" (UID: \"ac0cbf27-35f8-4646-b18b-a3416757472c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-lgbbh" Mar 07 15:11:45 crc kubenswrapper[4943]: I0307 15:11:45.143669 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rcp6x\" (UniqueName: \"kubernetes.io/projected/ac0cbf27-35f8-4646-b18b-a3416757472c-kube-api-access-rcp6x\") pod \"swift-ring-rebalance-debug-lgbbh\" (UID: \"ac0cbf27-35f8-4646-b18b-a3416757472c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-lgbbh" Mar 07 15:11:45 crc kubenswrapper[4943]: I0307 15:11:45.255047 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-lgbbh" Mar 07 15:11:45 crc kubenswrapper[4943]: I0307 15:11:45.511836 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-lgbbh"] Mar 07 15:11:46 crc kubenswrapper[4943]: I0307 15:11:46.295172 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-lgbbh" event={"ID":"ac0cbf27-35f8-4646-b18b-a3416757472c","Type":"ContainerStarted","Data":"ead602cb4d8d3d03b05cfcbf8fcb86519261f201698044c7f52a83582c4c3d7e"} Mar 07 15:11:46 crc kubenswrapper[4943]: I0307 15:11:46.295533 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-lgbbh" event={"ID":"ac0cbf27-35f8-4646-b18b-a3416757472c","Type":"ContainerStarted","Data":"30bccb710c1c74dbdd0c3b2ee194271b3763725378ff3bf72a6dccd5b054b316"} Mar 07 15:11:46 crc kubenswrapper[4943]: I0307 15:11:46.329986 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-lgbbh" podStartSLOduration=2.329963313 podStartE2EDuration="2.329963313s" podCreationTimestamp="2026-03-07 15:11:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:11:46.323994176 +0000 UTC m=+1948.276130684" watchObservedRunningTime="2026-03-07 15:11:46.329963313 +0000 UTC m=+1948.282099851" Mar 07 15:11:47 crc kubenswrapper[4943]: I0307 15:11:47.310237 4943 generic.go:334] "Generic (PLEG): container finished" podID="ac0cbf27-35f8-4646-b18b-a3416757472c" containerID="ead602cb4d8d3d03b05cfcbf8fcb86519261f201698044c7f52a83582c4c3d7e" exitCode=0 Mar 07 15:11:47 crc kubenswrapper[4943]: I0307 15:11:47.310305 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-lgbbh" event={"ID":"ac0cbf27-35f8-4646-b18b-a3416757472c","Type":"ContainerDied","Data":"ead602cb4d8d3d03b05cfcbf8fcb86519261f201698044c7f52a83582c4c3d7e"} Mar 07 15:11:48 crc kubenswrapper[4943]: I0307 15:11:48.703693 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-lgbbh" Mar 07 15:11:48 crc kubenswrapper[4943]: I0307 15:11:48.765208 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ac0cbf27-35f8-4646-b18b-a3416757472c-scripts\") pod \"ac0cbf27-35f8-4646-b18b-a3416757472c\" (UID: \"ac0cbf27-35f8-4646-b18b-a3416757472c\") " Mar 07 15:11:48 crc kubenswrapper[4943]: I0307 15:11:48.765302 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ac0cbf27-35f8-4646-b18b-a3416757472c-ring-data-devices\") pod \"ac0cbf27-35f8-4646-b18b-a3416757472c\" (UID: \"ac0cbf27-35f8-4646-b18b-a3416757472c\") " Mar 07 15:11:48 crc kubenswrapper[4943]: I0307 15:11:48.765643 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ac0cbf27-35f8-4646-b18b-a3416757472c-swiftconf\") pod \"ac0cbf27-35f8-4646-b18b-a3416757472c\" (UID: \"ac0cbf27-35f8-4646-b18b-a3416757472c\") " Mar 07 15:11:48 crc kubenswrapper[4943]: I0307 15:11:48.765801 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ac0cbf27-35f8-4646-b18b-a3416757472c-etc-swift\") pod \"ac0cbf27-35f8-4646-b18b-a3416757472c\" (UID: \"ac0cbf27-35f8-4646-b18b-a3416757472c\") " Mar 07 15:11:48 crc kubenswrapper[4943]: I0307 15:11:48.766016 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rcp6x\" (UniqueName: \"kubernetes.io/projected/ac0cbf27-35f8-4646-b18b-a3416757472c-kube-api-access-rcp6x\") pod \"ac0cbf27-35f8-4646-b18b-a3416757472c\" (UID: \"ac0cbf27-35f8-4646-b18b-a3416757472c\") " Mar 07 15:11:48 crc kubenswrapper[4943]: I0307 15:11:48.766461 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ac0cbf27-35f8-4646-b18b-a3416757472c-dispersionconf\") pod \"ac0cbf27-35f8-4646-b18b-a3416757472c\" (UID: \"ac0cbf27-35f8-4646-b18b-a3416757472c\") " Mar 07 15:11:48 crc kubenswrapper[4943]: I0307 15:11:48.766983 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac0cbf27-35f8-4646-b18b-a3416757472c-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "ac0cbf27-35f8-4646-b18b-a3416757472c" (UID: "ac0cbf27-35f8-4646-b18b-a3416757472c"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:11:48 crc kubenswrapper[4943]: I0307 15:11:48.767236 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac0cbf27-35f8-4646-b18b-a3416757472c-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "ac0cbf27-35f8-4646-b18b-a3416757472c" (UID: "ac0cbf27-35f8-4646-b18b-a3416757472c"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:11:48 crc kubenswrapper[4943]: I0307 15:11:48.769382 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ac0cbf27-35f8-4646-b18b-a3416757472c-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:48 crc kubenswrapper[4943]: I0307 15:11:48.769416 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ac0cbf27-35f8-4646-b18b-a3416757472c-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:48 crc kubenswrapper[4943]: I0307 15:11:48.780142 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac0cbf27-35f8-4646-b18b-a3416757472c-kube-api-access-rcp6x" (OuterVolumeSpecName: "kube-api-access-rcp6x") pod "ac0cbf27-35f8-4646-b18b-a3416757472c" (UID: "ac0cbf27-35f8-4646-b18b-a3416757472c"). InnerVolumeSpecName "kube-api-access-rcp6x". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:11:48 crc kubenswrapper[4943]: I0307 15:11:48.797480 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac0cbf27-35f8-4646-b18b-a3416757472c-scripts" (OuterVolumeSpecName: "scripts") pod "ac0cbf27-35f8-4646-b18b-a3416757472c" (UID: "ac0cbf27-35f8-4646-b18b-a3416757472c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:11:48 crc kubenswrapper[4943]: I0307 15:11:48.798431 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac0cbf27-35f8-4646-b18b-a3416757472c-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "ac0cbf27-35f8-4646-b18b-a3416757472c" (UID: "ac0cbf27-35f8-4646-b18b-a3416757472c"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:11:48 crc kubenswrapper[4943]: I0307 15:11:48.807026 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac0cbf27-35f8-4646-b18b-a3416757472c-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "ac0cbf27-35f8-4646-b18b-a3416757472c" (UID: "ac0cbf27-35f8-4646-b18b-a3416757472c"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:11:48 crc kubenswrapper[4943]: I0307 15:11:48.852613 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-lgbbh"] Mar 07 15:11:48 crc kubenswrapper[4943]: I0307 15:11:48.852691 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-lgbbh"] Mar 07 15:11:48 crc kubenswrapper[4943]: I0307 15:11:48.871076 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rcp6x\" (UniqueName: \"kubernetes.io/projected/ac0cbf27-35f8-4646-b18b-a3416757472c-kube-api-access-rcp6x\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:48 crc kubenswrapper[4943]: I0307 15:11:48.871295 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ac0cbf27-35f8-4646-b18b-a3416757472c-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:48 crc kubenswrapper[4943]: I0307 15:11:48.871422 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ac0cbf27-35f8-4646-b18b-a3416757472c-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:48 crc kubenswrapper[4943]: I0307 15:11:48.871540 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ac0cbf27-35f8-4646-b18b-a3416757472c-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:49 crc kubenswrapper[4943]: I0307 15:11:49.337409 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="30bccb710c1c74dbdd0c3b2ee194271b3763725378ff3bf72a6dccd5b054b316" Mar 07 15:11:49 crc kubenswrapper[4943]: I0307 15:11:49.337526 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-lgbbh" Mar 07 15:11:49 crc kubenswrapper[4943]: I0307 15:11:49.882141 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-qqpw5"] Mar 07 15:11:49 crc kubenswrapper[4943]: E0307 15:11:49.882720 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac0cbf27-35f8-4646-b18b-a3416757472c" containerName="swift-ring-rebalance" Mar 07 15:11:49 crc kubenswrapper[4943]: I0307 15:11:49.882749 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac0cbf27-35f8-4646-b18b-a3416757472c" containerName="swift-ring-rebalance" Mar 07 15:11:49 crc kubenswrapper[4943]: I0307 15:11:49.883124 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac0cbf27-35f8-4646-b18b-a3416757472c" containerName="swift-ring-rebalance" Mar 07 15:11:49 crc kubenswrapper[4943]: I0307 15:11:49.884076 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qqpw5" Mar 07 15:11:49 crc kubenswrapper[4943]: I0307 15:11:49.887430 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-qqpw5"] Mar 07 15:11:49 crc kubenswrapper[4943]: I0307 15:11:49.888040 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:11:49 crc kubenswrapper[4943]: I0307 15:11:49.889269 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:11:49 crc kubenswrapper[4943]: I0307 15:11:49.989419 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9566f0f2-6987-44fa-8cd6-e08ea34379fc-etc-swift\") pod \"swift-ring-rebalance-debug-qqpw5\" (UID: \"9566f0f2-6987-44fa-8cd6-e08ea34379fc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qqpw5" Mar 07 15:11:49 crc kubenswrapper[4943]: I0307 15:11:49.989491 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9566f0f2-6987-44fa-8cd6-e08ea34379fc-swiftconf\") pod \"swift-ring-rebalance-debug-qqpw5\" (UID: \"9566f0f2-6987-44fa-8cd6-e08ea34379fc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qqpw5" Mar 07 15:11:49 crc kubenswrapper[4943]: I0307 15:11:49.989507 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9566f0f2-6987-44fa-8cd6-e08ea34379fc-dispersionconf\") pod \"swift-ring-rebalance-debug-qqpw5\" (UID: \"9566f0f2-6987-44fa-8cd6-e08ea34379fc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qqpw5" Mar 07 15:11:49 crc kubenswrapper[4943]: I0307 15:11:49.989687 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9566f0f2-6987-44fa-8cd6-e08ea34379fc-ring-data-devices\") pod \"swift-ring-rebalance-debug-qqpw5\" (UID: \"9566f0f2-6987-44fa-8cd6-e08ea34379fc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qqpw5" Mar 07 15:11:49 crc kubenswrapper[4943]: I0307 15:11:49.989805 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qg962\" (UniqueName: \"kubernetes.io/projected/9566f0f2-6987-44fa-8cd6-e08ea34379fc-kube-api-access-qg962\") pod \"swift-ring-rebalance-debug-qqpw5\" (UID: \"9566f0f2-6987-44fa-8cd6-e08ea34379fc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qqpw5" Mar 07 15:11:49 crc kubenswrapper[4943]: I0307 15:11:49.989839 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9566f0f2-6987-44fa-8cd6-e08ea34379fc-scripts\") pod \"swift-ring-rebalance-debug-qqpw5\" (UID: \"9566f0f2-6987-44fa-8cd6-e08ea34379fc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qqpw5" Mar 07 15:11:50 crc kubenswrapper[4943]: I0307 15:11:50.090644 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9566f0f2-6987-44fa-8cd6-e08ea34379fc-ring-data-devices\") pod \"swift-ring-rebalance-debug-qqpw5\" (UID: \"9566f0f2-6987-44fa-8cd6-e08ea34379fc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qqpw5" Mar 07 15:11:50 crc kubenswrapper[4943]: I0307 15:11:50.091309 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qg962\" (UniqueName: \"kubernetes.io/projected/9566f0f2-6987-44fa-8cd6-e08ea34379fc-kube-api-access-qg962\") pod \"swift-ring-rebalance-debug-qqpw5\" (UID: \"9566f0f2-6987-44fa-8cd6-e08ea34379fc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qqpw5" Mar 07 15:11:50 crc kubenswrapper[4943]: I0307 15:11:50.091545 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9566f0f2-6987-44fa-8cd6-e08ea34379fc-scripts\") pod \"swift-ring-rebalance-debug-qqpw5\" (UID: \"9566f0f2-6987-44fa-8cd6-e08ea34379fc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qqpw5" Mar 07 15:11:50 crc kubenswrapper[4943]: I0307 15:11:50.091549 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9566f0f2-6987-44fa-8cd6-e08ea34379fc-ring-data-devices\") pod \"swift-ring-rebalance-debug-qqpw5\" (UID: \"9566f0f2-6987-44fa-8cd6-e08ea34379fc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qqpw5" Mar 07 15:11:50 crc kubenswrapper[4943]: I0307 15:11:50.091833 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9566f0f2-6987-44fa-8cd6-e08ea34379fc-etc-swift\") pod \"swift-ring-rebalance-debug-qqpw5\" (UID: \"9566f0f2-6987-44fa-8cd6-e08ea34379fc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qqpw5" Mar 07 15:11:50 crc kubenswrapper[4943]: I0307 15:11:50.092128 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9566f0f2-6987-44fa-8cd6-e08ea34379fc-swiftconf\") pod \"swift-ring-rebalance-debug-qqpw5\" (UID: \"9566f0f2-6987-44fa-8cd6-e08ea34379fc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qqpw5" Mar 07 15:11:50 crc kubenswrapper[4943]: I0307 15:11:50.092179 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9566f0f2-6987-44fa-8cd6-e08ea34379fc-dispersionconf\") pod \"swift-ring-rebalance-debug-qqpw5\" (UID: \"9566f0f2-6987-44fa-8cd6-e08ea34379fc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qqpw5" Mar 07 15:11:50 crc kubenswrapper[4943]: I0307 15:11:50.092783 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9566f0f2-6987-44fa-8cd6-e08ea34379fc-scripts\") pod \"swift-ring-rebalance-debug-qqpw5\" (UID: \"9566f0f2-6987-44fa-8cd6-e08ea34379fc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qqpw5" Mar 07 15:11:50 crc kubenswrapper[4943]: I0307 15:11:50.093295 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9566f0f2-6987-44fa-8cd6-e08ea34379fc-etc-swift\") pod \"swift-ring-rebalance-debug-qqpw5\" (UID: \"9566f0f2-6987-44fa-8cd6-e08ea34379fc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qqpw5" Mar 07 15:11:50 crc kubenswrapper[4943]: I0307 15:11:50.095662 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9566f0f2-6987-44fa-8cd6-e08ea34379fc-dispersionconf\") pod \"swift-ring-rebalance-debug-qqpw5\" (UID: \"9566f0f2-6987-44fa-8cd6-e08ea34379fc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qqpw5" Mar 07 15:11:50 crc kubenswrapper[4943]: I0307 15:11:50.105018 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9566f0f2-6987-44fa-8cd6-e08ea34379fc-swiftconf\") pod \"swift-ring-rebalance-debug-qqpw5\" (UID: \"9566f0f2-6987-44fa-8cd6-e08ea34379fc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qqpw5" Mar 07 15:11:50 crc kubenswrapper[4943]: I0307 15:11:50.109797 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qg962\" (UniqueName: \"kubernetes.io/projected/9566f0f2-6987-44fa-8cd6-e08ea34379fc-kube-api-access-qg962\") pod \"swift-ring-rebalance-debug-qqpw5\" (UID: \"9566f0f2-6987-44fa-8cd6-e08ea34379fc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qqpw5" Mar 07 15:11:50 crc kubenswrapper[4943]: I0307 15:11:50.204277 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qqpw5" Mar 07 15:11:50 crc kubenswrapper[4943]: I0307 15:11:50.683348 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-qqpw5"] Mar 07 15:11:50 crc kubenswrapper[4943]: I0307 15:11:50.771820 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac0cbf27-35f8-4646-b18b-a3416757472c" path="/var/lib/kubelet/pods/ac0cbf27-35f8-4646-b18b-a3416757472c/volumes" Mar 07 15:11:51 crc kubenswrapper[4943]: I0307 15:11:51.352914 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qqpw5" event={"ID":"9566f0f2-6987-44fa-8cd6-e08ea34379fc","Type":"ContainerStarted","Data":"40334ac85a352f802ad7e5593220d3e8811b459855f74de6c90b7bb16b7077e3"} Mar 07 15:11:51 crc kubenswrapper[4943]: I0307 15:11:51.354412 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qqpw5" event={"ID":"9566f0f2-6987-44fa-8cd6-e08ea34379fc","Type":"ContainerStarted","Data":"6366604814a140e90a0b191afdea74eca822766c0e1277257dfa743cd2492080"} Mar 07 15:11:51 crc kubenswrapper[4943]: I0307 15:11:51.372573 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qqpw5" podStartSLOduration=2.372556554 podStartE2EDuration="2.372556554s" podCreationTimestamp="2026-03-07 15:11:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:11:51.367458179 +0000 UTC m=+1953.319594687" watchObservedRunningTime="2026-03-07 15:11:51.372556554 +0000 UTC m=+1953.324693062" Mar 07 15:11:52 crc kubenswrapper[4943]: I0307 15:11:52.366338 4943 generic.go:334] "Generic (PLEG): container finished" podID="9566f0f2-6987-44fa-8cd6-e08ea34379fc" containerID="40334ac85a352f802ad7e5593220d3e8811b459855f74de6c90b7bb16b7077e3" exitCode=0 Mar 07 15:11:52 crc kubenswrapper[4943]: I0307 15:11:52.366415 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qqpw5" event={"ID":"9566f0f2-6987-44fa-8cd6-e08ea34379fc","Type":"ContainerDied","Data":"40334ac85a352f802ad7e5593220d3e8811b459855f74de6c90b7bb16b7077e3"} Mar 07 15:11:53 crc kubenswrapper[4943]: I0307 15:11:53.779068 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qqpw5" Mar 07 15:11:53 crc kubenswrapper[4943]: I0307 15:11:53.813368 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-qqpw5"] Mar 07 15:11:53 crc kubenswrapper[4943]: I0307 15:11:53.820135 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-qqpw5"] Mar 07 15:11:53 crc kubenswrapper[4943]: I0307 15:11:53.855083 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9566f0f2-6987-44fa-8cd6-e08ea34379fc-ring-data-devices\") pod \"9566f0f2-6987-44fa-8cd6-e08ea34379fc\" (UID: \"9566f0f2-6987-44fa-8cd6-e08ea34379fc\") " Mar 07 15:11:53 crc kubenswrapper[4943]: I0307 15:11:53.855138 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9566f0f2-6987-44fa-8cd6-e08ea34379fc-dispersionconf\") pod \"9566f0f2-6987-44fa-8cd6-e08ea34379fc\" (UID: \"9566f0f2-6987-44fa-8cd6-e08ea34379fc\") " Mar 07 15:11:53 crc kubenswrapper[4943]: I0307 15:11:53.855165 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg962\" (UniqueName: \"kubernetes.io/projected/9566f0f2-6987-44fa-8cd6-e08ea34379fc-kube-api-access-qg962\") pod \"9566f0f2-6987-44fa-8cd6-e08ea34379fc\" (UID: \"9566f0f2-6987-44fa-8cd6-e08ea34379fc\") " Mar 07 15:11:53 crc kubenswrapper[4943]: I0307 15:11:53.855205 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9566f0f2-6987-44fa-8cd6-e08ea34379fc-etc-swift\") pod \"9566f0f2-6987-44fa-8cd6-e08ea34379fc\" (UID: \"9566f0f2-6987-44fa-8cd6-e08ea34379fc\") " Mar 07 15:11:53 crc kubenswrapper[4943]: I0307 15:11:53.855252 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9566f0f2-6987-44fa-8cd6-e08ea34379fc-scripts\") pod \"9566f0f2-6987-44fa-8cd6-e08ea34379fc\" (UID: \"9566f0f2-6987-44fa-8cd6-e08ea34379fc\") " Mar 07 15:11:53 crc kubenswrapper[4943]: I0307 15:11:53.855346 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9566f0f2-6987-44fa-8cd6-e08ea34379fc-swiftconf\") pod \"9566f0f2-6987-44fa-8cd6-e08ea34379fc\" (UID: \"9566f0f2-6987-44fa-8cd6-e08ea34379fc\") " Mar 07 15:11:53 crc kubenswrapper[4943]: I0307 15:11:53.856693 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9566f0f2-6987-44fa-8cd6-e08ea34379fc-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "9566f0f2-6987-44fa-8cd6-e08ea34379fc" (UID: "9566f0f2-6987-44fa-8cd6-e08ea34379fc"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:11:53 crc kubenswrapper[4943]: I0307 15:11:53.857163 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9566f0f2-6987-44fa-8cd6-e08ea34379fc-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "9566f0f2-6987-44fa-8cd6-e08ea34379fc" (UID: "9566f0f2-6987-44fa-8cd6-e08ea34379fc"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:11:53 crc kubenswrapper[4943]: I0307 15:11:53.861924 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9566f0f2-6987-44fa-8cd6-e08ea34379fc-kube-api-access-qg962" (OuterVolumeSpecName: "kube-api-access-qg962") pod "9566f0f2-6987-44fa-8cd6-e08ea34379fc" (UID: "9566f0f2-6987-44fa-8cd6-e08ea34379fc"). InnerVolumeSpecName "kube-api-access-qg962". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:11:53 crc kubenswrapper[4943]: I0307 15:11:53.876040 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9566f0f2-6987-44fa-8cd6-e08ea34379fc-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "9566f0f2-6987-44fa-8cd6-e08ea34379fc" (UID: "9566f0f2-6987-44fa-8cd6-e08ea34379fc"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:11:53 crc kubenswrapper[4943]: I0307 15:11:53.884841 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9566f0f2-6987-44fa-8cd6-e08ea34379fc-scripts" (OuterVolumeSpecName: "scripts") pod "9566f0f2-6987-44fa-8cd6-e08ea34379fc" (UID: "9566f0f2-6987-44fa-8cd6-e08ea34379fc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:11:53 crc kubenswrapper[4943]: I0307 15:11:53.887531 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9566f0f2-6987-44fa-8cd6-e08ea34379fc-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "9566f0f2-6987-44fa-8cd6-e08ea34379fc" (UID: "9566f0f2-6987-44fa-8cd6-e08ea34379fc"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:11:53 crc kubenswrapper[4943]: I0307 15:11:53.956420 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9566f0f2-6987-44fa-8cd6-e08ea34379fc-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:53 crc kubenswrapper[4943]: I0307 15:11:53.956650 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9566f0f2-6987-44fa-8cd6-e08ea34379fc-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:53 crc kubenswrapper[4943]: I0307 15:11:53.956659 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg962\" (UniqueName: \"kubernetes.io/projected/9566f0f2-6987-44fa-8cd6-e08ea34379fc-kube-api-access-qg962\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:53 crc kubenswrapper[4943]: I0307 15:11:53.956671 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9566f0f2-6987-44fa-8cd6-e08ea34379fc-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:53 crc kubenswrapper[4943]: I0307 15:11:53.956680 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9566f0f2-6987-44fa-8cd6-e08ea34379fc-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:53 crc kubenswrapper[4943]: I0307 15:11:53.956688 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9566f0f2-6987-44fa-8cd6-e08ea34379fc-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:54 crc kubenswrapper[4943]: I0307 15:11:54.386339 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6366604814a140e90a0b191afdea74eca822766c0e1277257dfa743cd2492080" Mar 07 15:11:54 crc kubenswrapper[4943]: I0307 15:11:54.386425 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qqpw5" Mar 07 15:11:54 crc kubenswrapper[4943]: I0307 15:11:54.769985 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9566f0f2-6987-44fa-8cd6-e08ea34379fc" path="/var/lib/kubelet/pods/9566f0f2-6987-44fa-8cd6-e08ea34379fc/volumes" Mar 07 15:11:54 crc kubenswrapper[4943]: I0307 15:11:54.988808 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-27b5n"] Mar 07 15:11:54 crc kubenswrapper[4943]: E0307 15:11:54.989269 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9566f0f2-6987-44fa-8cd6-e08ea34379fc" containerName="swift-ring-rebalance" Mar 07 15:11:54 crc kubenswrapper[4943]: I0307 15:11:54.989291 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="9566f0f2-6987-44fa-8cd6-e08ea34379fc" containerName="swift-ring-rebalance" Mar 07 15:11:54 crc kubenswrapper[4943]: I0307 15:11:54.989582 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="9566f0f2-6987-44fa-8cd6-e08ea34379fc" containerName="swift-ring-rebalance" Mar 07 15:11:54 crc kubenswrapper[4943]: I0307 15:11:54.990420 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-27b5n" Mar 07 15:11:54 crc kubenswrapper[4943]: I0307 15:11:54.994160 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:11:54 crc kubenswrapper[4943]: I0307 15:11:54.995224 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:11:55 crc kubenswrapper[4943]: I0307 15:11:55.008168 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-27b5n"] Mar 07 15:11:55 crc kubenswrapper[4943]: I0307 15:11:55.074049 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/94f21e40-abc9-4003-b3ec-c0e8dc801582-ring-data-devices\") pod \"swift-ring-rebalance-debug-27b5n\" (UID: \"94f21e40-abc9-4003-b3ec-c0e8dc801582\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-27b5n" Mar 07 15:11:55 crc kubenswrapper[4943]: I0307 15:11:55.074124 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j74tk\" (UniqueName: \"kubernetes.io/projected/94f21e40-abc9-4003-b3ec-c0e8dc801582-kube-api-access-j74tk\") pod \"swift-ring-rebalance-debug-27b5n\" (UID: \"94f21e40-abc9-4003-b3ec-c0e8dc801582\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-27b5n" Mar 07 15:11:55 crc kubenswrapper[4943]: I0307 15:11:55.074279 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/94f21e40-abc9-4003-b3ec-c0e8dc801582-scripts\") pod \"swift-ring-rebalance-debug-27b5n\" (UID: \"94f21e40-abc9-4003-b3ec-c0e8dc801582\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-27b5n" Mar 07 15:11:55 crc kubenswrapper[4943]: I0307 15:11:55.074402 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/94f21e40-abc9-4003-b3ec-c0e8dc801582-dispersionconf\") pod \"swift-ring-rebalance-debug-27b5n\" (UID: \"94f21e40-abc9-4003-b3ec-c0e8dc801582\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-27b5n" Mar 07 15:11:55 crc kubenswrapper[4943]: I0307 15:11:55.074549 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/94f21e40-abc9-4003-b3ec-c0e8dc801582-etc-swift\") pod \"swift-ring-rebalance-debug-27b5n\" (UID: \"94f21e40-abc9-4003-b3ec-c0e8dc801582\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-27b5n" Mar 07 15:11:55 crc kubenswrapper[4943]: I0307 15:11:55.074741 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/94f21e40-abc9-4003-b3ec-c0e8dc801582-swiftconf\") pod \"swift-ring-rebalance-debug-27b5n\" (UID: \"94f21e40-abc9-4003-b3ec-c0e8dc801582\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-27b5n" Mar 07 15:11:55 crc kubenswrapper[4943]: I0307 15:11:55.176600 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/94f21e40-abc9-4003-b3ec-c0e8dc801582-swiftconf\") pod \"swift-ring-rebalance-debug-27b5n\" (UID: \"94f21e40-abc9-4003-b3ec-c0e8dc801582\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-27b5n" Mar 07 15:11:55 crc kubenswrapper[4943]: I0307 15:11:55.176688 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/94f21e40-abc9-4003-b3ec-c0e8dc801582-ring-data-devices\") pod \"swift-ring-rebalance-debug-27b5n\" (UID: \"94f21e40-abc9-4003-b3ec-c0e8dc801582\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-27b5n" Mar 07 15:11:55 crc kubenswrapper[4943]: I0307 15:11:55.177715 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/94f21e40-abc9-4003-b3ec-c0e8dc801582-ring-data-devices\") pod \"swift-ring-rebalance-debug-27b5n\" (UID: \"94f21e40-abc9-4003-b3ec-c0e8dc801582\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-27b5n" Mar 07 15:11:55 crc kubenswrapper[4943]: I0307 15:11:55.176719 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j74tk\" (UniqueName: \"kubernetes.io/projected/94f21e40-abc9-4003-b3ec-c0e8dc801582-kube-api-access-j74tk\") pod \"swift-ring-rebalance-debug-27b5n\" (UID: \"94f21e40-abc9-4003-b3ec-c0e8dc801582\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-27b5n" Mar 07 15:11:55 crc kubenswrapper[4943]: I0307 15:11:55.177815 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/94f21e40-abc9-4003-b3ec-c0e8dc801582-scripts\") pod \"swift-ring-rebalance-debug-27b5n\" (UID: \"94f21e40-abc9-4003-b3ec-c0e8dc801582\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-27b5n" Mar 07 15:11:55 crc kubenswrapper[4943]: I0307 15:11:55.177913 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/94f21e40-abc9-4003-b3ec-c0e8dc801582-dispersionconf\") pod \"swift-ring-rebalance-debug-27b5n\" (UID: \"94f21e40-abc9-4003-b3ec-c0e8dc801582\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-27b5n" Mar 07 15:11:55 crc kubenswrapper[4943]: I0307 15:11:55.178423 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/94f21e40-abc9-4003-b3ec-c0e8dc801582-etc-swift\") pod \"swift-ring-rebalance-debug-27b5n\" (UID: \"94f21e40-abc9-4003-b3ec-c0e8dc801582\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-27b5n" Mar 07 15:11:55 crc kubenswrapper[4943]: I0307 15:11:55.178527 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/94f21e40-abc9-4003-b3ec-c0e8dc801582-scripts\") pod \"swift-ring-rebalance-debug-27b5n\" (UID: \"94f21e40-abc9-4003-b3ec-c0e8dc801582\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-27b5n" Mar 07 15:11:55 crc kubenswrapper[4943]: I0307 15:11:55.178855 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/94f21e40-abc9-4003-b3ec-c0e8dc801582-etc-swift\") pod \"swift-ring-rebalance-debug-27b5n\" (UID: \"94f21e40-abc9-4003-b3ec-c0e8dc801582\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-27b5n" Mar 07 15:11:55 crc kubenswrapper[4943]: I0307 15:11:55.182279 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/94f21e40-abc9-4003-b3ec-c0e8dc801582-dispersionconf\") pod \"swift-ring-rebalance-debug-27b5n\" (UID: \"94f21e40-abc9-4003-b3ec-c0e8dc801582\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-27b5n" Mar 07 15:11:55 crc kubenswrapper[4943]: I0307 15:11:55.182355 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/94f21e40-abc9-4003-b3ec-c0e8dc801582-swiftconf\") pod \"swift-ring-rebalance-debug-27b5n\" (UID: \"94f21e40-abc9-4003-b3ec-c0e8dc801582\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-27b5n" Mar 07 15:11:55 crc kubenswrapper[4943]: I0307 15:11:55.198333 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j74tk\" (UniqueName: \"kubernetes.io/projected/94f21e40-abc9-4003-b3ec-c0e8dc801582-kube-api-access-j74tk\") pod \"swift-ring-rebalance-debug-27b5n\" (UID: \"94f21e40-abc9-4003-b3ec-c0e8dc801582\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-27b5n" Mar 07 15:11:55 crc kubenswrapper[4943]: I0307 15:11:55.318625 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-27b5n" Mar 07 15:11:55 crc kubenswrapper[4943]: I0307 15:11:55.780616 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-27b5n"] Mar 07 15:11:55 crc kubenswrapper[4943]: W0307 15:11:55.789546 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod94f21e40_abc9_4003_b3ec_c0e8dc801582.slice/crio-db31cf9b3fe5b4cb87454c3b33c542a11fad4b38ef557aae511d946fe7f335d1 WatchSource:0}: Error finding container db31cf9b3fe5b4cb87454c3b33c542a11fad4b38ef557aae511d946fe7f335d1: Status 404 returned error can't find the container with id db31cf9b3fe5b4cb87454c3b33c542a11fad4b38ef557aae511d946fe7f335d1 Mar 07 15:11:56 crc kubenswrapper[4943]: I0307 15:11:56.411891 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-27b5n" event={"ID":"94f21e40-abc9-4003-b3ec-c0e8dc801582","Type":"ContainerStarted","Data":"dd1d4566544d3718923d8b461f0a50398b77407a748b0fab8c31f40fa2ebe405"} Mar 07 15:11:56 crc kubenswrapper[4943]: I0307 15:11:56.412207 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-27b5n" event={"ID":"94f21e40-abc9-4003-b3ec-c0e8dc801582","Type":"ContainerStarted","Data":"db31cf9b3fe5b4cb87454c3b33c542a11fad4b38ef557aae511d946fe7f335d1"} Mar 07 15:11:56 crc kubenswrapper[4943]: I0307 15:11:56.454005 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-27b5n" podStartSLOduration=2.4539759070000002 podStartE2EDuration="2.453975907s" podCreationTimestamp="2026-03-07 15:11:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:11:56.440771554 +0000 UTC m=+1958.392908122" watchObservedRunningTime="2026-03-07 15:11:56.453975907 +0000 UTC m=+1958.406112445" Mar 07 15:11:57 crc kubenswrapper[4943]: I0307 15:11:57.422733 4943 generic.go:334] "Generic (PLEG): container finished" podID="94f21e40-abc9-4003-b3ec-c0e8dc801582" containerID="dd1d4566544d3718923d8b461f0a50398b77407a748b0fab8c31f40fa2ebe405" exitCode=0 Mar 07 15:11:57 crc kubenswrapper[4943]: I0307 15:11:57.423126 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-27b5n" event={"ID":"94f21e40-abc9-4003-b3ec-c0e8dc801582","Type":"ContainerDied","Data":"dd1d4566544d3718923d8b461f0a50398b77407a748b0fab8c31f40fa2ebe405"} Mar 07 15:11:59 crc kubenswrapper[4943]: I0307 15:11:59.667601 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-27b5n" Mar 07 15:11:59 crc kubenswrapper[4943]: I0307 15:11:59.712748 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-27b5n"] Mar 07 15:11:59 crc kubenswrapper[4943]: I0307 15:11:59.719892 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-27b5n"] Mar 07 15:11:59 crc kubenswrapper[4943]: I0307 15:11:59.774131 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/94f21e40-abc9-4003-b3ec-c0e8dc801582-dispersionconf\") pod \"94f21e40-abc9-4003-b3ec-c0e8dc801582\" (UID: \"94f21e40-abc9-4003-b3ec-c0e8dc801582\") " Mar 07 15:11:59 crc kubenswrapper[4943]: I0307 15:11:59.774213 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/94f21e40-abc9-4003-b3ec-c0e8dc801582-etc-swift\") pod \"94f21e40-abc9-4003-b3ec-c0e8dc801582\" (UID: \"94f21e40-abc9-4003-b3ec-c0e8dc801582\") " Mar 07 15:11:59 crc kubenswrapper[4943]: I0307 15:11:59.774364 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/94f21e40-abc9-4003-b3ec-c0e8dc801582-scripts\") pod \"94f21e40-abc9-4003-b3ec-c0e8dc801582\" (UID: \"94f21e40-abc9-4003-b3ec-c0e8dc801582\") " Mar 07 15:11:59 crc kubenswrapper[4943]: I0307 15:11:59.774402 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j74tk\" (UniqueName: \"kubernetes.io/projected/94f21e40-abc9-4003-b3ec-c0e8dc801582-kube-api-access-j74tk\") pod \"94f21e40-abc9-4003-b3ec-c0e8dc801582\" (UID: \"94f21e40-abc9-4003-b3ec-c0e8dc801582\") " Mar 07 15:11:59 crc kubenswrapper[4943]: I0307 15:11:59.774446 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/94f21e40-abc9-4003-b3ec-c0e8dc801582-swiftconf\") pod \"94f21e40-abc9-4003-b3ec-c0e8dc801582\" (UID: \"94f21e40-abc9-4003-b3ec-c0e8dc801582\") " Mar 07 15:11:59 crc kubenswrapper[4943]: I0307 15:11:59.774468 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/94f21e40-abc9-4003-b3ec-c0e8dc801582-ring-data-devices\") pod \"94f21e40-abc9-4003-b3ec-c0e8dc801582\" (UID: \"94f21e40-abc9-4003-b3ec-c0e8dc801582\") " Mar 07 15:11:59 crc kubenswrapper[4943]: I0307 15:11:59.774826 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94f21e40-abc9-4003-b3ec-c0e8dc801582-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "94f21e40-abc9-4003-b3ec-c0e8dc801582" (UID: "94f21e40-abc9-4003-b3ec-c0e8dc801582"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:11:59 crc kubenswrapper[4943]: I0307 15:11:59.775308 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94f21e40-abc9-4003-b3ec-c0e8dc801582-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "94f21e40-abc9-4003-b3ec-c0e8dc801582" (UID: "94f21e40-abc9-4003-b3ec-c0e8dc801582"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:11:59 crc kubenswrapper[4943]: I0307 15:11:59.781269 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94f21e40-abc9-4003-b3ec-c0e8dc801582-kube-api-access-j74tk" (OuterVolumeSpecName: "kube-api-access-j74tk") pod "94f21e40-abc9-4003-b3ec-c0e8dc801582" (UID: "94f21e40-abc9-4003-b3ec-c0e8dc801582"). InnerVolumeSpecName "kube-api-access-j74tk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:11:59 crc kubenswrapper[4943]: I0307 15:11:59.810973 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94f21e40-abc9-4003-b3ec-c0e8dc801582-scripts" (OuterVolumeSpecName: "scripts") pod "94f21e40-abc9-4003-b3ec-c0e8dc801582" (UID: "94f21e40-abc9-4003-b3ec-c0e8dc801582"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:11:59 crc kubenswrapper[4943]: I0307 15:11:59.813545 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94f21e40-abc9-4003-b3ec-c0e8dc801582-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "94f21e40-abc9-4003-b3ec-c0e8dc801582" (UID: "94f21e40-abc9-4003-b3ec-c0e8dc801582"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:11:59 crc kubenswrapper[4943]: I0307 15:11:59.815268 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94f21e40-abc9-4003-b3ec-c0e8dc801582-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "94f21e40-abc9-4003-b3ec-c0e8dc801582" (UID: "94f21e40-abc9-4003-b3ec-c0e8dc801582"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:11:59 crc kubenswrapper[4943]: I0307 15:11:59.875898 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/94f21e40-abc9-4003-b3ec-c0e8dc801582-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:59 crc kubenswrapper[4943]: I0307 15:11:59.876025 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/94f21e40-abc9-4003-b3ec-c0e8dc801582-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:59 crc kubenswrapper[4943]: I0307 15:11:59.876039 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/94f21e40-abc9-4003-b3ec-c0e8dc801582-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:59 crc kubenswrapper[4943]: I0307 15:11:59.876049 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/94f21e40-abc9-4003-b3ec-c0e8dc801582-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:59 crc kubenswrapper[4943]: I0307 15:11:59.876059 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/94f21e40-abc9-4003-b3ec-c0e8dc801582-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:11:59 crc kubenswrapper[4943]: I0307 15:11:59.876071 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j74tk\" (UniqueName: \"kubernetes.io/projected/94f21e40-abc9-4003-b3ec-c0e8dc801582-kube-api-access-j74tk\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:00 crc kubenswrapper[4943]: I0307 15:12:00.154889 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29548272-qdhrf"] Mar 07 15:12:00 crc kubenswrapper[4943]: E0307 15:12:00.155353 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94f21e40-abc9-4003-b3ec-c0e8dc801582" containerName="swift-ring-rebalance" Mar 07 15:12:00 crc kubenswrapper[4943]: I0307 15:12:00.155376 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="94f21e40-abc9-4003-b3ec-c0e8dc801582" containerName="swift-ring-rebalance" Mar 07 15:12:00 crc kubenswrapper[4943]: I0307 15:12:00.155639 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="94f21e40-abc9-4003-b3ec-c0e8dc801582" containerName="swift-ring-rebalance" Mar 07 15:12:00 crc kubenswrapper[4943]: I0307 15:12:00.156451 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548272-qdhrf" Mar 07 15:12:00 crc kubenswrapper[4943]: I0307 15:12:00.163556 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 07 15:12:00 crc kubenswrapper[4943]: I0307 15:12:00.164790 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 07 15:12:00 crc kubenswrapper[4943]: I0307 15:12:00.169305 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-qqlk5" Mar 07 15:12:00 crc kubenswrapper[4943]: I0307 15:12:00.187673 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29548272-qdhrf"] Mar 07 15:12:00 crc kubenswrapper[4943]: I0307 15:12:00.282795 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94d22\" (UniqueName: \"kubernetes.io/projected/f3920f3e-037a-4678-b3dd-9da204486979-kube-api-access-94d22\") pod \"auto-csr-approver-29548272-qdhrf\" (UID: \"f3920f3e-037a-4678-b3dd-9da204486979\") " pod="openshift-infra/auto-csr-approver-29548272-qdhrf" Mar 07 15:12:00 crc kubenswrapper[4943]: I0307 15:12:00.300210 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="db31cf9b3fe5b4cb87454c3b33c542a11fad4b38ef557aae511d946fe7f335d1" Mar 07 15:12:00 crc kubenswrapper[4943]: I0307 15:12:00.300397 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-27b5n" Mar 07 15:12:00 crc kubenswrapper[4943]: I0307 15:12:00.384786 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94d22\" (UniqueName: \"kubernetes.io/projected/f3920f3e-037a-4678-b3dd-9da204486979-kube-api-access-94d22\") pod \"auto-csr-approver-29548272-qdhrf\" (UID: \"f3920f3e-037a-4678-b3dd-9da204486979\") " pod="openshift-infra/auto-csr-approver-29548272-qdhrf" Mar 07 15:12:00 crc kubenswrapper[4943]: I0307 15:12:00.406819 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94d22\" (UniqueName: \"kubernetes.io/projected/f3920f3e-037a-4678-b3dd-9da204486979-kube-api-access-94d22\") pod \"auto-csr-approver-29548272-qdhrf\" (UID: \"f3920f3e-037a-4678-b3dd-9da204486979\") " pod="openshift-infra/auto-csr-approver-29548272-qdhrf" Mar 07 15:12:00 crc kubenswrapper[4943]: I0307 15:12:00.497679 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548272-qdhrf" Mar 07 15:12:00 crc kubenswrapper[4943]: I0307 15:12:00.783482 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94f21e40-abc9-4003-b3ec-c0e8dc801582" path="/var/lib/kubelet/pods/94f21e40-abc9-4003-b3ec-c0e8dc801582/volumes" Mar 07 15:12:00 crc kubenswrapper[4943]: I0307 15:12:00.845811 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29548272-qdhrf"] Mar 07 15:12:00 crc kubenswrapper[4943]: I0307 15:12:00.848473 4943 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 07 15:12:00 crc kubenswrapper[4943]: I0307 15:12:00.901652 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-8clvz"] Mar 07 15:12:00 crc kubenswrapper[4943]: I0307 15:12:00.903778 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8clvz" Mar 07 15:12:00 crc kubenswrapper[4943]: I0307 15:12:00.906235 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:12:00 crc kubenswrapper[4943]: I0307 15:12:00.906268 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:12:00 crc kubenswrapper[4943]: I0307 15:12:00.912750 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-8clvz"] Mar 07 15:12:01 crc kubenswrapper[4943]: I0307 15:12:01.098181 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twz6b\" (UniqueName: \"kubernetes.io/projected/36b636da-727c-4f52-a24e-ea8abda67948-kube-api-access-twz6b\") pod \"swift-ring-rebalance-debug-8clvz\" (UID: \"36b636da-727c-4f52-a24e-ea8abda67948\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8clvz" Mar 07 15:12:01 crc kubenswrapper[4943]: I0307 15:12:01.098256 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/36b636da-727c-4f52-a24e-ea8abda67948-ring-data-devices\") pod \"swift-ring-rebalance-debug-8clvz\" (UID: \"36b636da-727c-4f52-a24e-ea8abda67948\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8clvz" Mar 07 15:12:01 crc kubenswrapper[4943]: I0307 15:12:01.098327 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/36b636da-727c-4f52-a24e-ea8abda67948-dispersionconf\") pod \"swift-ring-rebalance-debug-8clvz\" (UID: \"36b636da-727c-4f52-a24e-ea8abda67948\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8clvz" Mar 07 15:12:01 crc kubenswrapper[4943]: I0307 15:12:01.098423 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/36b636da-727c-4f52-a24e-ea8abda67948-scripts\") pod \"swift-ring-rebalance-debug-8clvz\" (UID: \"36b636da-727c-4f52-a24e-ea8abda67948\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8clvz" Mar 07 15:12:01 crc kubenswrapper[4943]: I0307 15:12:01.098523 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/36b636da-727c-4f52-a24e-ea8abda67948-etc-swift\") pod \"swift-ring-rebalance-debug-8clvz\" (UID: \"36b636da-727c-4f52-a24e-ea8abda67948\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8clvz" Mar 07 15:12:01 crc kubenswrapper[4943]: I0307 15:12:01.098594 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/36b636da-727c-4f52-a24e-ea8abda67948-swiftconf\") pod \"swift-ring-rebalance-debug-8clvz\" (UID: \"36b636da-727c-4f52-a24e-ea8abda67948\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8clvz" Mar 07 15:12:01 crc kubenswrapper[4943]: I0307 15:12:01.200625 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/36b636da-727c-4f52-a24e-ea8abda67948-scripts\") pod \"swift-ring-rebalance-debug-8clvz\" (UID: \"36b636da-727c-4f52-a24e-ea8abda67948\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8clvz" Mar 07 15:12:01 crc kubenswrapper[4943]: I0307 15:12:01.200682 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/36b636da-727c-4f52-a24e-ea8abda67948-etc-swift\") pod \"swift-ring-rebalance-debug-8clvz\" (UID: \"36b636da-727c-4f52-a24e-ea8abda67948\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8clvz" Mar 07 15:12:01 crc kubenswrapper[4943]: I0307 15:12:01.200725 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/36b636da-727c-4f52-a24e-ea8abda67948-swiftconf\") pod \"swift-ring-rebalance-debug-8clvz\" (UID: \"36b636da-727c-4f52-a24e-ea8abda67948\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8clvz" Mar 07 15:12:01 crc kubenswrapper[4943]: I0307 15:12:01.200876 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twz6b\" (UniqueName: \"kubernetes.io/projected/36b636da-727c-4f52-a24e-ea8abda67948-kube-api-access-twz6b\") pod \"swift-ring-rebalance-debug-8clvz\" (UID: \"36b636da-727c-4f52-a24e-ea8abda67948\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8clvz" Mar 07 15:12:01 crc kubenswrapper[4943]: I0307 15:12:01.200955 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/36b636da-727c-4f52-a24e-ea8abda67948-ring-data-devices\") pod \"swift-ring-rebalance-debug-8clvz\" (UID: \"36b636da-727c-4f52-a24e-ea8abda67948\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8clvz" Mar 07 15:12:01 crc kubenswrapper[4943]: I0307 15:12:01.201016 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/36b636da-727c-4f52-a24e-ea8abda67948-dispersionconf\") pod \"swift-ring-rebalance-debug-8clvz\" (UID: \"36b636da-727c-4f52-a24e-ea8abda67948\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8clvz" Mar 07 15:12:01 crc kubenswrapper[4943]: I0307 15:12:01.201434 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/36b636da-727c-4f52-a24e-ea8abda67948-etc-swift\") pod \"swift-ring-rebalance-debug-8clvz\" (UID: \"36b636da-727c-4f52-a24e-ea8abda67948\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8clvz" Mar 07 15:12:01 crc kubenswrapper[4943]: I0307 15:12:01.201548 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/36b636da-727c-4f52-a24e-ea8abda67948-scripts\") pod \"swift-ring-rebalance-debug-8clvz\" (UID: \"36b636da-727c-4f52-a24e-ea8abda67948\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8clvz" Mar 07 15:12:01 crc kubenswrapper[4943]: I0307 15:12:01.202034 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/36b636da-727c-4f52-a24e-ea8abda67948-ring-data-devices\") pod \"swift-ring-rebalance-debug-8clvz\" (UID: \"36b636da-727c-4f52-a24e-ea8abda67948\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8clvz" Mar 07 15:12:01 crc kubenswrapper[4943]: I0307 15:12:01.206123 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/36b636da-727c-4f52-a24e-ea8abda67948-swiftconf\") pod \"swift-ring-rebalance-debug-8clvz\" (UID: \"36b636da-727c-4f52-a24e-ea8abda67948\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8clvz" Mar 07 15:12:01 crc kubenswrapper[4943]: I0307 15:12:01.212824 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/36b636da-727c-4f52-a24e-ea8abda67948-dispersionconf\") pod \"swift-ring-rebalance-debug-8clvz\" (UID: \"36b636da-727c-4f52-a24e-ea8abda67948\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8clvz" Mar 07 15:12:01 crc kubenswrapper[4943]: I0307 15:12:01.225811 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twz6b\" (UniqueName: \"kubernetes.io/projected/36b636da-727c-4f52-a24e-ea8abda67948-kube-api-access-twz6b\") pod \"swift-ring-rebalance-debug-8clvz\" (UID: \"36b636da-727c-4f52-a24e-ea8abda67948\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8clvz" Mar 07 15:12:01 crc kubenswrapper[4943]: I0307 15:12:01.227027 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8clvz" Mar 07 15:12:01 crc kubenswrapper[4943]: I0307 15:12:01.312863 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548272-qdhrf" event={"ID":"f3920f3e-037a-4678-b3dd-9da204486979","Type":"ContainerStarted","Data":"f4d86c018fee81b1f8b6407e71b4a0bc08827a7f5e40d1a3886ff158fd0c41bd"} Mar 07 15:12:01 crc kubenswrapper[4943]: W0307 15:12:01.509114 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod36b636da_727c_4f52_a24e_ea8abda67948.slice/crio-70c396478c3231f83692cb91b6dcfcb77e2ab55861f95e7febd58ace57a0593a WatchSource:0}: Error finding container 70c396478c3231f83692cb91b6dcfcb77e2ab55861f95e7febd58ace57a0593a: Status 404 returned error can't find the container with id 70c396478c3231f83692cb91b6dcfcb77e2ab55861f95e7febd58ace57a0593a Mar 07 15:12:01 crc kubenswrapper[4943]: I0307 15:12:01.512032 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-8clvz"] Mar 07 15:12:02 crc kubenswrapper[4943]: I0307 15:12:02.327473 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548272-qdhrf" event={"ID":"f3920f3e-037a-4678-b3dd-9da204486979","Type":"ContainerStarted","Data":"c6a6c74af6e301b6fccd01a45947763df009f2d805ee884c52a266e7fe4cdfc0"} Mar 07 15:12:02 crc kubenswrapper[4943]: I0307 15:12:02.334206 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8clvz" event={"ID":"36b636da-727c-4f52-a24e-ea8abda67948","Type":"ContainerStarted","Data":"8bdb2c800d876fccdead65d0ca37c85a454cd13b12e2418466d5f9d9667fa317"} Mar 07 15:12:02 crc kubenswrapper[4943]: I0307 15:12:02.334310 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8clvz" event={"ID":"36b636da-727c-4f52-a24e-ea8abda67948","Type":"ContainerStarted","Data":"70c396478c3231f83692cb91b6dcfcb77e2ab55861f95e7febd58ace57a0593a"} Mar 07 15:12:02 crc kubenswrapper[4943]: I0307 15:12:02.355761 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29548272-qdhrf" podStartSLOduration=1.5377591590000002 podStartE2EDuration="2.355738824s" podCreationTimestamp="2026-03-07 15:12:00 +0000 UTC" firstStartedPulling="2026-03-07 15:12:00.848031081 +0000 UTC m=+1962.800167609" lastFinishedPulling="2026-03-07 15:12:01.666010766 +0000 UTC m=+1963.618147274" observedRunningTime="2026-03-07 15:12:02.348752973 +0000 UTC m=+1964.300889511" watchObservedRunningTime="2026-03-07 15:12:02.355738824 +0000 UTC m=+1964.307875342" Mar 07 15:12:02 crc kubenswrapper[4943]: I0307 15:12:02.380059 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8clvz" podStartSLOduration=2.38003498 podStartE2EDuration="2.38003498s" podCreationTimestamp="2026-03-07 15:12:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:12:02.372243319 +0000 UTC m=+1964.324379827" watchObservedRunningTime="2026-03-07 15:12:02.38003498 +0000 UTC m=+1964.332171488" Mar 07 15:12:03 crc kubenswrapper[4943]: I0307 15:12:03.346862 4943 generic.go:334] "Generic (PLEG): container finished" podID="f3920f3e-037a-4678-b3dd-9da204486979" containerID="c6a6c74af6e301b6fccd01a45947763df009f2d805ee884c52a266e7fe4cdfc0" exitCode=0 Mar 07 15:12:03 crc kubenswrapper[4943]: I0307 15:12:03.346958 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548272-qdhrf" event={"ID":"f3920f3e-037a-4678-b3dd-9da204486979","Type":"ContainerDied","Data":"c6a6c74af6e301b6fccd01a45947763df009f2d805ee884c52a266e7fe4cdfc0"} Mar 07 15:12:03 crc kubenswrapper[4943]: I0307 15:12:03.350212 4943 generic.go:334] "Generic (PLEG): container finished" podID="36b636da-727c-4f52-a24e-ea8abda67948" containerID="8bdb2c800d876fccdead65d0ca37c85a454cd13b12e2418466d5f9d9667fa317" exitCode=0 Mar 07 15:12:03 crc kubenswrapper[4943]: I0307 15:12:03.350287 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8clvz" event={"ID":"36b636da-727c-4f52-a24e-ea8abda67948","Type":"ContainerDied","Data":"8bdb2c800d876fccdead65d0ca37c85a454cd13b12e2418466d5f9d9667fa317"} Mar 07 15:12:04 crc kubenswrapper[4943]: I0307 15:12:04.747168 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8clvz" Mar 07 15:12:04 crc kubenswrapper[4943]: I0307 15:12:04.752396 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548272-qdhrf" Mar 07 15:12:04 crc kubenswrapper[4943]: I0307 15:12:04.806094 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-8clvz"] Mar 07 15:12:04 crc kubenswrapper[4943]: I0307 15:12:04.818973 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-8clvz"] Mar 07 15:12:04 crc kubenswrapper[4943]: I0307 15:12:04.865261 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/36b636da-727c-4f52-a24e-ea8abda67948-dispersionconf\") pod \"36b636da-727c-4f52-a24e-ea8abda67948\" (UID: \"36b636da-727c-4f52-a24e-ea8abda67948\") " Mar 07 15:12:04 crc kubenswrapper[4943]: I0307 15:12:04.865320 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/36b636da-727c-4f52-a24e-ea8abda67948-ring-data-devices\") pod \"36b636da-727c-4f52-a24e-ea8abda67948\" (UID: \"36b636da-727c-4f52-a24e-ea8abda67948\") " Mar 07 15:12:04 crc kubenswrapper[4943]: I0307 15:12:04.865343 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/36b636da-727c-4f52-a24e-ea8abda67948-etc-swift\") pod \"36b636da-727c-4f52-a24e-ea8abda67948\" (UID: \"36b636da-727c-4f52-a24e-ea8abda67948\") " Mar 07 15:12:04 crc kubenswrapper[4943]: I0307 15:12:04.865376 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/36b636da-727c-4f52-a24e-ea8abda67948-swiftconf\") pod \"36b636da-727c-4f52-a24e-ea8abda67948\" (UID: \"36b636da-727c-4f52-a24e-ea8abda67948\") " Mar 07 15:12:04 crc kubenswrapper[4943]: I0307 15:12:04.865412 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-94d22\" (UniqueName: \"kubernetes.io/projected/f3920f3e-037a-4678-b3dd-9da204486979-kube-api-access-94d22\") pod \"f3920f3e-037a-4678-b3dd-9da204486979\" (UID: \"f3920f3e-037a-4678-b3dd-9da204486979\") " Mar 07 15:12:04 crc kubenswrapper[4943]: I0307 15:12:04.865495 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-twz6b\" (UniqueName: \"kubernetes.io/projected/36b636da-727c-4f52-a24e-ea8abda67948-kube-api-access-twz6b\") pod \"36b636da-727c-4f52-a24e-ea8abda67948\" (UID: \"36b636da-727c-4f52-a24e-ea8abda67948\") " Mar 07 15:12:04 crc kubenswrapper[4943]: I0307 15:12:04.865568 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/36b636da-727c-4f52-a24e-ea8abda67948-scripts\") pod \"36b636da-727c-4f52-a24e-ea8abda67948\" (UID: \"36b636da-727c-4f52-a24e-ea8abda67948\") " Mar 07 15:12:04 crc kubenswrapper[4943]: I0307 15:12:04.866723 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36b636da-727c-4f52-a24e-ea8abda67948-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "36b636da-727c-4f52-a24e-ea8abda67948" (UID: "36b636da-727c-4f52-a24e-ea8abda67948"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:12:04 crc kubenswrapper[4943]: I0307 15:12:04.867153 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36b636da-727c-4f52-a24e-ea8abda67948-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "36b636da-727c-4f52-a24e-ea8abda67948" (UID: "36b636da-727c-4f52-a24e-ea8abda67948"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:12:04 crc kubenswrapper[4943]: I0307 15:12:04.870938 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36b636da-727c-4f52-a24e-ea8abda67948-kube-api-access-twz6b" (OuterVolumeSpecName: "kube-api-access-twz6b") pod "36b636da-727c-4f52-a24e-ea8abda67948" (UID: "36b636da-727c-4f52-a24e-ea8abda67948"). InnerVolumeSpecName "kube-api-access-twz6b". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:12:04 crc kubenswrapper[4943]: I0307 15:12:04.873416 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3920f3e-037a-4678-b3dd-9da204486979-kube-api-access-94d22" (OuterVolumeSpecName: "kube-api-access-94d22") pod "f3920f3e-037a-4678-b3dd-9da204486979" (UID: "f3920f3e-037a-4678-b3dd-9da204486979"). InnerVolumeSpecName "kube-api-access-94d22". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:12:04 crc kubenswrapper[4943]: I0307 15:12:04.891284 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36b636da-727c-4f52-a24e-ea8abda67948-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "36b636da-727c-4f52-a24e-ea8abda67948" (UID: "36b636da-727c-4f52-a24e-ea8abda67948"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:12:04 crc kubenswrapper[4943]: I0307 15:12:04.902249 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36b636da-727c-4f52-a24e-ea8abda67948-scripts" (OuterVolumeSpecName: "scripts") pod "36b636da-727c-4f52-a24e-ea8abda67948" (UID: "36b636da-727c-4f52-a24e-ea8abda67948"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:12:04 crc kubenswrapper[4943]: I0307 15:12:04.902316 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36b636da-727c-4f52-a24e-ea8abda67948-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "36b636da-727c-4f52-a24e-ea8abda67948" (UID: "36b636da-727c-4f52-a24e-ea8abda67948"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:12:04 crc kubenswrapper[4943]: I0307 15:12:04.967367 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/36b636da-727c-4f52-a24e-ea8abda67948-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:04 crc kubenswrapper[4943]: I0307 15:12:04.967405 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/36b636da-727c-4f52-a24e-ea8abda67948-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:04 crc kubenswrapper[4943]: I0307 15:12:04.967426 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/36b636da-727c-4f52-a24e-ea8abda67948-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:04 crc kubenswrapper[4943]: I0307 15:12:04.967443 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/36b636da-727c-4f52-a24e-ea8abda67948-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:04 crc kubenswrapper[4943]: I0307 15:12:04.967459 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/36b636da-727c-4f52-a24e-ea8abda67948-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:04 crc kubenswrapper[4943]: I0307 15:12:04.967627 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-94d22\" (UniqueName: \"kubernetes.io/projected/f3920f3e-037a-4678-b3dd-9da204486979-kube-api-access-94d22\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:04 crc kubenswrapper[4943]: I0307 15:12:04.967649 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-twz6b\" (UniqueName: \"kubernetes.io/projected/36b636da-727c-4f52-a24e-ea8abda67948-kube-api-access-twz6b\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:05 crc kubenswrapper[4943]: I0307 15:12:05.369424 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="70c396478c3231f83692cb91b6dcfcb77e2ab55861f95e7febd58ace57a0593a" Mar 07 15:12:05 crc kubenswrapper[4943]: I0307 15:12:05.369470 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8clvz" Mar 07 15:12:05 crc kubenswrapper[4943]: I0307 15:12:05.372916 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548272-qdhrf" event={"ID":"f3920f3e-037a-4678-b3dd-9da204486979","Type":"ContainerDied","Data":"f4d86c018fee81b1f8b6407e71b4a0bc08827a7f5e40d1a3886ff158fd0c41bd"} Mar 07 15:12:05 crc kubenswrapper[4943]: I0307 15:12:05.373025 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f4d86c018fee81b1f8b6407e71b4a0bc08827a7f5e40d1a3886ff158fd0c41bd" Mar 07 15:12:05 crc kubenswrapper[4943]: I0307 15:12:05.373131 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548272-qdhrf" Mar 07 15:12:05 crc kubenswrapper[4943]: I0307 15:12:05.442720 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29548266-tw972"] Mar 07 15:12:05 crc kubenswrapper[4943]: I0307 15:12:05.449149 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29548266-tw972"] Mar 07 15:12:05 crc kubenswrapper[4943]: I0307 15:12:05.982468 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-2m4qv"] Mar 07 15:12:05 crc kubenswrapper[4943]: E0307 15:12:05.982890 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3920f3e-037a-4678-b3dd-9da204486979" containerName="oc" Mar 07 15:12:05 crc kubenswrapper[4943]: I0307 15:12:05.982910 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3920f3e-037a-4678-b3dd-9da204486979" containerName="oc" Mar 07 15:12:05 crc kubenswrapper[4943]: E0307 15:12:05.982969 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36b636da-727c-4f52-a24e-ea8abda67948" containerName="swift-ring-rebalance" Mar 07 15:12:05 crc kubenswrapper[4943]: I0307 15:12:05.982983 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="36b636da-727c-4f52-a24e-ea8abda67948" containerName="swift-ring-rebalance" Mar 07 15:12:05 crc kubenswrapper[4943]: I0307 15:12:05.983244 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3920f3e-037a-4678-b3dd-9da204486979" containerName="oc" Mar 07 15:12:05 crc kubenswrapper[4943]: I0307 15:12:05.983278 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="36b636da-727c-4f52-a24e-ea8abda67948" containerName="swift-ring-rebalance" Mar 07 15:12:05 crc kubenswrapper[4943]: I0307 15:12:05.984038 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-2m4qv" Mar 07 15:12:05 crc kubenswrapper[4943]: I0307 15:12:05.986086 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:12:05 crc kubenswrapper[4943]: I0307 15:12:05.987265 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:12:05 crc kubenswrapper[4943]: I0307 15:12:05.996297 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-2m4qv"] Mar 07 15:12:06 crc kubenswrapper[4943]: I0307 15:12:06.084245 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/26efb559-dc00-4f7f-ba68-49357ab4d8a6-scripts\") pod \"swift-ring-rebalance-debug-2m4qv\" (UID: \"26efb559-dc00-4f7f-ba68-49357ab4d8a6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2m4qv" Mar 07 15:12:06 crc kubenswrapper[4943]: I0307 15:12:06.084385 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/26efb559-dc00-4f7f-ba68-49357ab4d8a6-dispersionconf\") pod \"swift-ring-rebalance-debug-2m4qv\" (UID: \"26efb559-dc00-4f7f-ba68-49357ab4d8a6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2m4qv" Mar 07 15:12:06 crc kubenswrapper[4943]: I0307 15:12:06.084641 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/26efb559-dc00-4f7f-ba68-49357ab4d8a6-swiftconf\") pod \"swift-ring-rebalance-debug-2m4qv\" (UID: \"26efb559-dc00-4f7f-ba68-49357ab4d8a6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2m4qv" Mar 07 15:12:06 crc kubenswrapper[4943]: I0307 15:12:06.084728 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/26efb559-dc00-4f7f-ba68-49357ab4d8a6-ring-data-devices\") pod \"swift-ring-rebalance-debug-2m4qv\" (UID: \"26efb559-dc00-4f7f-ba68-49357ab4d8a6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2m4qv" Mar 07 15:12:06 crc kubenswrapper[4943]: I0307 15:12:06.084869 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6k5d\" (UniqueName: \"kubernetes.io/projected/26efb559-dc00-4f7f-ba68-49357ab4d8a6-kube-api-access-x6k5d\") pod \"swift-ring-rebalance-debug-2m4qv\" (UID: \"26efb559-dc00-4f7f-ba68-49357ab4d8a6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2m4qv" Mar 07 15:12:06 crc kubenswrapper[4943]: I0307 15:12:06.084961 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/26efb559-dc00-4f7f-ba68-49357ab4d8a6-etc-swift\") pod \"swift-ring-rebalance-debug-2m4qv\" (UID: \"26efb559-dc00-4f7f-ba68-49357ab4d8a6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2m4qv" Mar 07 15:12:06 crc kubenswrapper[4943]: I0307 15:12:06.186805 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/26efb559-dc00-4f7f-ba68-49357ab4d8a6-dispersionconf\") pod \"swift-ring-rebalance-debug-2m4qv\" (UID: \"26efb559-dc00-4f7f-ba68-49357ab4d8a6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2m4qv" Mar 07 15:12:06 crc kubenswrapper[4943]: I0307 15:12:06.186895 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/26efb559-dc00-4f7f-ba68-49357ab4d8a6-swiftconf\") pod \"swift-ring-rebalance-debug-2m4qv\" (UID: \"26efb559-dc00-4f7f-ba68-49357ab4d8a6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2m4qv" Mar 07 15:12:06 crc kubenswrapper[4943]: I0307 15:12:06.186921 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/26efb559-dc00-4f7f-ba68-49357ab4d8a6-ring-data-devices\") pod \"swift-ring-rebalance-debug-2m4qv\" (UID: \"26efb559-dc00-4f7f-ba68-49357ab4d8a6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2m4qv" Mar 07 15:12:06 crc kubenswrapper[4943]: I0307 15:12:06.186974 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6k5d\" (UniqueName: \"kubernetes.io/projected/26efb559-dc00-4f7f-ba68-49357ab4d8a6-kube-api-access-x6k5d\") pod \"swift-ring-rebalance-debug-2m4qv\" (UID: \"26efb559-dc00-4f7f-ba68-49357ab4d8a6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2m4qv" Mar 07 15:12:06 crc kubenswrapper[4943]: I0307 15:12:06.187001 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/26efb559-dc00-4f7f-ba68-49357ab4d8a6-etc-swift\") pod \"swift-ring-rebalance-debug-2m4qv\" (UID: \"26efb559-dc00-4f7f-ba68-49357ab4d8a6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2m4qv" Mar 07 15:12:06 crc kubenswrapper[4943]: I0307 15:12:06.187027 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/26efb559-dc00-4f7f-ba68-49357ab4d8a6-scripts\") pod \"swift-ring-rebalance-debug-2m4qv\" (UID: \"26efb559-dc00-4f7f-ba68-49357ab4d8a6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2m4qv" Mar 07 15:12:06 crc kubenswrapper[4943]: I0307 15:12:06.187982 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/26efb559-dc00-4f7f-ba68-49357ab4d8a6-etc-swift\") pod \"swift-ring-rebalance-debug-2m4qv\" (UID: \"26efb559-dc00-4f7f-ba68-49357ab4d8a6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2m4qv" Mar 07 15:12:06 crc kubenswrapper[4943]: I0307 15:12:06.188307 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/26efb559-dc00-4f7f-ba68-49357ab4d8a6-scripts\") pod \"swift-ring-rebalance-debug-2m4qv\" (UID: \"26efb559-dc00-4f7f-ba68-49357ab4d8a6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2m4qv" Mar 07 15:12:06 crc kubenswrapper[4943]: I0307 15:12:06.188741 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/26efb559-dc00-4f7f-ba68-49357ab4d8a6-ring-data-devices\") pod \"swift-ring-rebalance-debug-2m4qv\" (UID: \"26efb559-dc00-4f7f-ba68-49357ab4d8a6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2m4qv" Mar 07 15:12:06 crc kubenswrapper[4943]: I0307 15:12:06.193043 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/26efb559-dc00-4f7f-ba68-49357ab4d8a6-swiftconf\") pod \"swift-ring-rebalance-debug-2m4qv\" (UID: \"26efb559-dc00-4f7f-ba68-49357ab4d8a6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2m4qv" Mar 07 15:12:06 crc kubenswrapper[4943]: I0307 15:12:06.201292 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/26efb559-dc00-4f7f-ba68-49357ab4d8a6-dispersionconf\") pod \"swift-ring-rebalance-debug-2m4qv\" (UID: \"26efb559-dc00-4f7f-ba68-49357ab4d8a6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2m4qv" Mar 07 15:12:06 crc kubenswrapper[4943]: I0307 15:12:06.218641 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6k5d\" (UniqueName: \"kubernetes.io/projected/26efb559-dc00-4f7f-ba68-49357ab4d8a6-kube-api-access-x6k5d\") pod \"swift-ring-rebalance-debug-2m4qv\" (UID: \"26efb559-dc00-4f7f-ba68-49357ab4d8a6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2m4qv" Mar 07 15:12:06 crc kubenswrapper[4943]: I0307 15:12:06.300448 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-2m4qv" Mar 07 15:12:06 crc kubenswrapper[4943]: I0307 15:12:06.578600 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-2m4qv"] Mar 07 15:12:06 crc kubenswrapper[4943]: I0307 15:12:06.768758 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36b636da-727c-4f52-a24e-ea8abda67948" path="/var/lib/kubelet/pods/36b636da-727c-4f52-a24e-ea8abda67948/volumes" Mar 07 15:12:06 crc kubenswrapper[4943]: I0307 15:12:06.770460 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed54b4ee-fddf-44f5-928a-326405fac246" path="/var/lib/kubelet/pods/ed54b4ee-fddf-44f5-928a-326405fac246/volumes" Mar 07 15:12:07 crc kubenswrapper[4943]: I0307 15:12:07.396468 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-2m4qv" event={"ID":"26efb559-dc00-4f7f-ba68-49357ab4d8a6","Type":"ContainerStarted","Data":"c1f1429cb6c9119934cd1f73e5430200a689739be373733fd9c4eac0c91fcd3c"} Mar 07 15:12:07 crc kubenswrapper[4943]: I0307 15:12:07.396523 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-2m4qv" event={"ID":"26efb559-dc00-4f7f-ba68-49357ab4d8a6","Type":"ContainerStarted","Data":"a316ae815a02ca591c81625e5c47136045396e167edf25767ff3d61509f41e4c"} Mar 07 15:12:07 crc kubenswrapper[4943]: I0307 15:12:07.433581 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-2m4qv" podStartSLOduration=2.433552149 podStartE2EDuration="2.433552149s" podCreationTimestamp="2026-03-07 15:12:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:12:07.427019709 +0000 UTC m=+1969.379156247" watchObservedRunningTime="2026-03-07 15:12:07.433552149 +0000 UTC m=+1969.385688687" Mar 07 15:12:08 crc kubenswrapper[4943]: I0307 15:12:08.408691 4943 generic.go:334] "Generic (PLEG): container finished" podID="26efb559-dc00-4f7f-ba68-49357ab4d8a6" containerID="c1f1429cb6c9119934cd1f73e5430200a689739be373733fd9c4eac0c91fcd3c" exitCode=0 Mar 07 15:12:08 crc kubenswrapper[4943]: I0307 15:12:08.408861 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-2m4qv" event={"ID":"26efb559-dc00-4f7f-ba68-49357ab4d8a6","Type":"ContainerDied","Data":"c1f1429cb6c9119934cd1f73e5430200a689739be373733fd9c4eac0c91fcd3c"} Mar 07 15:12:09 crc kubenswrapper[4943]: I0307 15:12:09.803418 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-2m4qv" Mar 07 15:12:09 crc kubenswrapper[4943]: I0307 15:12:09.865998 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-2m4qv"] Mar 07 15:12:09 crc kubenswrapper[4943]: I0307 15:12:09.873177 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-2m4qv"] Mar 07 15:12:09 crc kubenswrapper[4943]: I0307 15:12:09.946512 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6k5d\" (UniqueName: \"kubernetes.io/projected/26efb559-dc00-4f7f-ba68-49357ab4d8a6-kube-api-access-x6k5d\") pod \"26efb559-dc00-4f7f-ba68-49357ab4d8a6\" (UID: \"26efb559-dc00-4f7f-ba68-49357ab4d8a6\") " Mar 07 15:12:09 crc kubenswrapper[4943]: I0307 15:12:09.946627 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/26efb559-dc00-4f7f-ba68-49357ab4d8a6-ring-data-devices\") pod \"26efb559-dc00-4f7f-ba68-49357ab4d8a6\" (UID: \"26efb559-dc00-4f7f-ba68-49357ab4d8a6\") " Mar 07 15:12:09 crc kubenswrapper[4943]: I0307 15:12:09.946705 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/26efb559-dc00-4f7f-ba68-49357ab4d8a6-swiftconf\") pod \"26efb559-dc00-4f7f-ba68-49357ab4d8a6\" (UID: \"26efb559-dc00-4f7f-ba68-49357ab4d8a6\") " Mar 07 15:12:09 crc kubenswrapper[4943]: I0307 15:12:09.946737 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/26efb559-dc00-4f7f-ba68-49357ab4d8a6-scripts\") pod \"26efb559-dc00-4f7f-ba68-49357ab4d8a6\" (UID: \"26efb559-dc00-4f7f-ba68-49357ab4d8a6\") " Mar 07 15:12:09 crc kubenswrapper[4943]: I0307 15:12:09.946797 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/26efb559-dc00-4f7f-ba68-49357ab4d8a6-dispersionconf\") pod \"26efb559-dc00-4f7f-ba68-49357ab4d8a6\" (UID: \"26efb559-dc00-4f7f-ba68-49357ab4d8a6\") " Mar 07 15:12:09 crc kubenswrapper[4943]: I0307 15:12:09.946863 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/26efb559-dc00-4f7f-ba68-49357ab4d8a6-etc-swift\") pod \"26efb559-dc00-4f7f-ba68-49357ab4d8a6\" (UID: \"26efb559-dc00-4f7f-ba68-49357ab4d8a6\") " Mar 07 15:12:09 crc kubenswrapper[4943]: I0307 15:12:09.947718 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26efb559-dc00-4f7f-ba68-49357ab4d8a6-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "26efb559-dc00-4f7f-ba68-49357ab4d8a6" (UID: "26efb559-dc00-4f7f-ba68-49357ab4d8a6"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:12:09 crc kubenswrapper[4943]: I0307 15:12:09.947742 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26efb559-dc00-4f7f-ba68-49357ab4d8a6-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "26efb559-dc00-4f7f-ba68-49357ab4d8a6" (UID: "26efb559-dc00-4f7f-ba68-49357ab4d8a6"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:12:09 crc kubenswrapper[4943]: I0307 15:12:09.958276 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26efb559-dc00-4f7f-ba68-49357ab4d8a6-kube-api-access-x6k5d" (OuterVolumeSpecName: "kube-api-access-x6k5d") pod "26efb559-dc00-4f7f-ba68-49357ab4d8a6" (UID: "26efb559-dc00-4f7f-ba68-49357ab4d8a6"). InnerVolumeSpecName "kube-api-access-x6k5d". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:12:09 crc kubenswrapper[4943]: I0307 15:12:09.967446 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26efb559-dc00-4f7f-ba68-49357ab4d8a6-scripts" (OuterVolumeSpecName: "scripts") pod "26efb559-dc00-4f7f-ba68-49357ab4d8a6" (UID: "26efb559-dc00-4f7f-ba68-49357ab4d8a6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:12:09 crc kubenswrapper[4943]: I0307 15:12:09.977092 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26efb559-dc00-4f7f-ba68-49357ab4d8a6-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "26efb559-dc00-4f7f-ba68-49357ab4d8a6" (UID: "26efb559-dc00-4f7f-ba68-49357ab4d8a6"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:12:09 crc kubenswrapper[4943]: I0307 15:12:09.986558 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26efb559-dc00-4f7f-ba68-49357ab4d8a6-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "26efb559-dc00-4f7f-ba68-49357ab4d8a6" (UID: "26efb559-dc00-4f7f-ba68-49357ab4d8a6"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:12:10 crc kubenswrapper[4943]: I0307 15:12:10.049136 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/26efb559-dc00-4f7f-ba68-49357ab4d8a6-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:10 crc kubenswrapper[4943]: I0307 15:12:10.049194 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/26efb559-dc00-4f7f-ba68-49357ab4d8a6-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:10 crc kubenswrapper[4943]: I0307 15:12:10.049212 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/26efb559-dc00-4f7f-ba68-49357ab4d8a6-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:10 crc kubenswrapper[4943]: I0307 15:12:10.049229 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/26efb559-dc00-4f7f-ba68-49357ab4d8a6-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:10 crc kubenswrapper[4943]: I0307 15:12:10.049247 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/26efb559-dc00-4f7f-ba68-49357ab4d8a6-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:10 crc kubenswrapper[4943]: I0307 15:12:10.049263 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6k5d\" (UniqueName: \"kubernetes.io/projected/26efb559-dc00-4f7f-ba68-49357ab4d8a6-kube-api-access-x6k5d\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:10 crc kubenswrapper[4943]: I0307 15:12:10.441041 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a316ae815a02ca591c81625e5c47136045396e167edf25767ff3d61509f41e4c" Mar 07 15:12:10 crc kubenswrapper[4943]: I0307 15:12:10.441208 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-2m4qv" Mar 07 15:12:10 crc kubenswrapper[4943]: I0307 15:12:10.769721 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26efb559-dc00-4f7f-ba68-49357ab4d8a6" path="/var/lib/kubelet/pods/26efb559-dc00-4f7f-ba68-49357ab4d8a6/volumes" Mar 07 15:12:11 crc kubenswrapper[4943]: I0307 15:12:11.046277 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-5mln5"] Mar 07 15:12:11 crc kubenswrapper[4943]: E0307 15:12:11.046994 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26efb559-dc00-4f7f-ba68-49357ab4d8a6" containerName="swift-ring-rebalance" Mar 07 15:12:11 crc kubenswrapper[4943]: I0307 15:12:11.047016 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="26efb559-dc00-4f7f-ba68-49357ab4d8a6" containerName="swift-ring-rebalance" Mar 07 15:12:11 crc kubenswrapper[4943]: I0307 15:12:11.047332 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="26efb559-dc00-4f7f-ba68-49357ab4d8a6" containerName="swift-ring-rebalance" Mar 07 15:12:11 crc kubenswrapper[4943]: I0307 15:12:11.048089 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mln5" Mar 07 15:12:11 crc kubenswrapper[4943]: I0307 15:12:11.050387 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:12:11 crc kubenswrapper[4943]: I0307 15:12:11.050500 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:12:11 crc kubenswrapper[4943]: I0307 15:12:11.071702 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-5mln5"] Mar 07 15:12:11 crc kubenswrapper[4943]: I0307 15:12:11.164053 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/0814782b-ab61-4cfe-8088-2036b434f236-etc-swift\") pod \"swift-ring-rebalance-debug-5mln5\" (UID: \"0814782b-ab61-4cfe-8088-2036b434f236\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mln5" Mar 07 15:12:11 crc kubenswrapper[4943]: I0307 15:12:11.164144 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/0814782b-ab61-4cfe-8088-2036b434f236-swiftconf\") pod \"swift-ring-rebalance-debug-5mln5\" (UID: \"0814782b-ab61-4cfe-8088-2036b434f236\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mln5" Mar 07 15:12:11 crc kubenswrapper[4943]: I0307 15:12:11.164248 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvcvx\" (UniqueName: \"kubernetes.io/projected/0814782b-ab61-4cfe-8088-2036b434f236-kube-api-access-bvcvx\") pod \"swift-ring-rebalance-debug-5mln5\" (UID: \"0814782b-ab61-4cfe-8088-2036b434f236\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mln5" Mar 07 15:12:11 crc kubenswrapper[4943]: I0307 15:12:11.164390 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/0814782b-ab61-4cfe-8088-2036b434f236-ring-data-devices\") pod \"swift-ring-rebalance-debug-5mln5\" (UID: \"0814782b-ab61-4cfe-8088-2036b434f236\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mln5" Mar 07 15:12:11 crc kubenswrapper[4943]: I0307 15:12:11.164536 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/0814782b-ab61-4cfe-8088-2036b434f236-dispersionconf\") pod \"swift-ring-rebalance-debug-5mln5\" (UID: \"0814782b-ab61-4cfe-8088-2036b434f236\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mln5" Mar 07 15:12:11 crc kubenswrapper[4943]: I0307 15:12:11.164661 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0814782b-ab61-4cfe-8088-2036b434f236-scripts\") pod \"swift-ring-rebalance-debug-5mln5\" (UID: \"0814782b-ab61-4cfe-8088-2036b434f236\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mln5" Mar 07 15:12:11 crc kubenswrapper[4943]: I0307 15:12:11.265753 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvcvx\" (UniqueName: \"kubernetes.io/projected/0814782b-ab61-4cfe-8088-2036b434f236-kube-api-access-bvcvx\") pod \"swift-ring-rebalance-debug-5mln5\" (UID: \"0814782b-ab61-4cfe-8088-2036b434f236\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mln5" Mar 07 15:12:11 crc kubenswrapper[4943]: I0307 15:12:11.265863 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/0814782b-ab61-4cfe-8088-2036b434f236-ring-data-devices\") pod \"swift-ring-rebalance-debug-5mln5\" (UID: \"0814782b-ab61-4cfe-8088-2036b434f236\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mln5" Mar 07 15:12:11 crc kubenswrapper[4943]: I0307 15:12:11.265971 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/0814782b-ab61-4cfe-8088-2036b434f236-dispersionconf\") pod \"swift-ring-rebalance-debug-5mln5\" (UID: \"0814782b-ab61-4cfe-8088-2036b434f236\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mln5" Mar 07 15:12:11 crc kubenswrapper[4943]: I0307 15:12:11.266035 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0814782b-ab61-4cfe-8088-2036b434f236-scripts\") pod \"swift-ring-rebalance-debug-5mln5\" (UID: \"0814782b-ab61-4cfe-8088-2036b434f236\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mln5" Mar 07 15:12:11 crc kubenswrapper[4943]: I0307 15:12:11.266114 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/0814782b-ab61-4cfe-8088-2036b434f236-etc-swift\") pod \"swift-ring-rebalance-debug-5mln5\" (UID: \"0814782b-ab61-4cfe-8088-2036b434f236\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mln5" Mar 07 15:12:11 crc kubenswrapper[4943]: I0307 15:12:11.266147 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/0814782b-ab61-4cfe-8088-2036b434f236-swiftconf\") pod \"swift-ring-rebalance-debug-5mln5\" (UID: \"0814782b-ab61-4cfe-8088-2036b434f236\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mln5" Mar 07 15:12:11 crc kubenswrapper[4943]: I0307 15:12:11.266669 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/0814782b-ab61-4cfe-8088-2036b434f236-etc-swift\") pod \"swift-ring-rebalance-debug-5mln5\" (UID: \"0814782b-ab61-4cfe-8088-2036b434f236\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mln5" Mar 07 15:12:11 crc kubenswrapper[4943]: I0307 15:12:11.266885 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/0814782b-ab61-4cfe-8088-2036b434f236-ring-data-devices\") pod \"swift-ring-rebalance-debug-5mln5\" (UID: \"0814782b-ab61-4cfe-8088-2036b434f236\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mln5" Mar 07 15:12:11 crc kubenswrapper[4943]: I0307 15:12:11.267234 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0814782b-ab61-4cfe-8088-2036b434f236-scripts\") pod \"swift-ring-rebalance-debug-5mln5\" (UID: \"0814782b-ab61-4cfe-8088-2036b434f236\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mln5" Mar 07 15:12:11 crc kubenswrapper[4943]: I0307 15:12:11.270506 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/0814782b-ab61-4cfe-8088-2036b434f236-dispersionconf\") pod \"swift-ring-rebalance-debug-5mln5\" (UID: \"0814782b-ab61-4cfe-8088-2036b434f236\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mln5" Mar 07 15:12:11 crc kubenswrapper[4943]: I0307 15:12:11.272437 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/0814782b-ab61-4cfe-8088-2036b434f236-swiftconf\") pod \"swift-ring-rebalance-debug-5mln5\" (UID: \"0814782b-ab61-4cfe-8088-2036b434f236\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mln5" Mar 07 15:12:11 crc kubenswrapper[4943]: I0307 15:12:11.283379 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvcvx\" (UniqueName: \"kubernetes.io/projected/0814782b-ab61-4cfe-8088-2036b434f236-kube-api-access-bvcvx\") pod \"swift-ring-rebalance-debug-5mln5\" (UID: \"0814782b-ab61-4cfe-8088-2036b434f236\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mln5" Mar 07 15:12:11 crc kubenswrapper[4943]: I0307 15:12:11.371321 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mln5" Mar 07 15:12:11 crc kubenswrapper[4943]: I0307 15:12:11.888661 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-5mln5"] Mar 07 15:12:12 crc kubenswrapper[4943]: I0307 15:12:12.462866 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mln5" event={"ID":"0814782b-ab61-4cfe-8088-2036b434f236","Type":"ContainerStarted","Data":"d92a84d136eafdb41db47adc5e8b097377d0e0eb5b3d35e93f3d3b6af516f440"} Mar 07 15:12:12 crc kubenswrapper[4943]: I0307 15:12:12.463290 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mln5" event={"ID":"0814782b-ab61-4cfe-8088-2036b434f236","Type":"ContainerStarted","Data":"306ab5e5ed6682a83b529b8a10800b3d2029ac299c0bc3b0ed7bd7491197bbac"} Mar 07 15:12:12 crc kubenswrapper[4943]: I0307 15:12:12.493262 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mln5" podStartSLOduration=1.4932283 podStartE2EDuration="1.4932283s" podCreationTimestamp="2026-03-07 15:12:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:12:12.490060263 +0000 UTC m=+1974.442196791" watchObservedRunningTime="2026-03-07 15:12:12.4932283 +0000 UTC m=+1974.445364838" Mar 07 15:12:13 crc kubenswrapper[4943]: I0307 15:12:13.476283 4943 generic.go:334] "Generic (PLEG): container finished" podID="0814782b-ab61-4cfe-8088-2036b434f236" containerID="d92a84d136eafdb41db47adc5e8b097377d0e0eb5b3d35e93f3d3b6af516f440" exitCode=0 Mar 07 15:12:13 crc kubenswrapper[4943]: I0307 15:12:13.476357 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mln5" event={"ID":"0814782b-ab61-4cfe-8088-2036b434f236","Type":"ContainerDied","Data":"d92a84d136eafdb41db47adc5e8b097377d0e0eb5b3d35e93f3d3b6af516f440"} Mar 07 15:12:14 crc kubenswrapper[4943]: I0307 15:12:14.851392 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mln5" Mar 07 15:12:14 crc kubenswrapper[4943]: I0307 15:12:14.903814 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-5mln5"] Mar 07 15:12:14 crc kubenswrapper[4943]: I0307 15:12:14.908679 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-5mln5"] Mar 07 15:12:14 crc kubenswrapper[4943]: I0307 15:12:14.930577 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/0814782b-ab61-4cfe-8088-2036b434f236-dispersionconf\") pod \"0814782b-ab61-4cfe-8088-2036b434f236\" (UID: \"0814782b-ab61-4cfe-8088-2036b434f236\") " Mar 07 15:12:14 crc kubenswrapper[4943]: I0307 15:12:14.930697 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/0814782b-ab61-4cfe-8088-2036b434f236-swiftconf\") pod \"0814782b-ab61-4cfe-8088-2036b434f236\" (UID: \"0814782b-ab61-4cfe-8088-2036b434f236\") " Mar 07 15:12:14 crc kubenswrapper[4943]: I0307 15:12:14.930831 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0814782b-ab61-4cfe-8088-2036b434f236-scripts\") pod \"0814782b-ab61-4cfe-8088-2036b434f236\" (UID: \"0814782b-ab61-4cfe-8088-2036b434f236\") " Mar 07 15:12:14 crc kubenswrapper[4943]: I0307 15:12:14.930944 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/0814782b-ab61-4cfe-8088-2036b434f236-etc-swift\") pod \"0814782b-ab61-4cfe-8088-2036b434f236\" (UID: \"0814782b-ab61-4cfe-8088-2036b434f236\") " Mar 07 15:12:14 crc kubenswrapper[4943]: I0307 15:12:14.931110 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bvcvx\" (UniqueName: \"kubernetes.io/projected/0814782b-ab61-4cfe-8088-2036b434f236-kube-api-access-bvcvx\") pod \"0814782b-ab61-4cfe-8088-2036b434f236\" (UID: \"0814782b-ab61-4cfe-8088-2036b434f236\") " Mar 07 15:12:14 crc kubenswrapper[4943]: I0307 15:12:14.931208 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/0814782b-ab61-4cfe-8088-2036b434f236-ring-data-devices\") pod \"0814782b-ab61-4cfe-8088-2036b434f236\" (UID: \"0814782b-ab61-4cfe-8088-2036b434f236\") " Mar 07 15:12:14 crc kubenswrapper[4943]: I0307 15:12:14.931971 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0814782b-ab61-4cfe-8088-2036b434f236-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "0814782b-ab61-4cfe-8088-2036b434f236" (UID: "0814782b-ab61-4cfe-8088-2036b434f236"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:12:14 crc kubenswrapper[4943]: I0307 15:12:14.932396 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0814782b-ab61-4cfe-8088-2036b434f236-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "0814782b-ab61-4cfe-8088-2036b434f236" (UID: "0814782b-ab61-4cfe-8088-2036b434f236"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:12:14 crc kubenswrapper[4943]: I0307 15:12:14.947437 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0814782b-ab61-4cfe-8088-2036b434f236-kube-api-access-bvcvx" (OuterVolumeSpecName: "kube-api-access-bvcvx") pod "0814782b-ab61-4cfe-8088-2036b434f236" (UID: "0814782b-ab61-4cfe-8088-2036b434f236"). InnerVolumeSpecName "kube-api-access-bvcvx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:12:14 crc kubenswrapper[4943]: I0307 15:12:14.954540 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0814782b-ab61-4cfe-8088-2036b434f236-scripts" (OuterVolumeSpecName: "scripts") pod "0814782b-ab61-4cfe-8088-2036b434f236" (UID: "0814782b-ab61-4cfe-8088-2036b434f236"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:12:14 crc kubenswrapper[4943]: I0307 15:12:14.959853 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0814782b-ab61-4cfe-8088-2036b434f236-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "0814782b-ab61-4cfe-8088-2036b434f236" (UID: "0814782b-ab61-4cfe-8088-2036b434f236"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:12:14 crc kubenswrapper[4943]: I0307 15:12:14.974197 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0814782b-ab61-4cfe-8088-2036b434f236-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "0814782b-ab61-4cfe-8088-2036b434f236" (UID: "0814782b-ab61-4cfe-8088-2036b434f236"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:12:15 crc kubenswrapper[4943]: I0307 15:12:15.035030 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0814782b-ab61-4cfe-8088-2036b434f236-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:15 crc kubenswrapper[4943]: I0307 15:12:15.035063 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/0814782b-ab61-4cfe-8088-2036b434f236-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:15 crc kubenswrapper[4943]: I0307 15:12:15.035099 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bvcvx\" (UniqueName: \"kubernetes.io/projected/0814782b-ab61-4cfe-8088-2036b434f236-kube-api-access-bvcvx\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:15 crc kubenswrapper[4943]: I0307 15:12:15.035118 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/0814782b-ab61-4cfe-8088-2036b434f236-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:15 crc kubenswrapper[4943]: I0307 15:12:15.035127 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/0814782b-ab61-4cfe-8088-2036b434f236-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:15 crc kubenswrapper[4943]: I0307 15:12:15.035135 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/0814782b-ab61-4cfe-8088-2036b434f236-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:15 crc kubenswrapper[4943]: I0307 15:12:15.503661 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="306ab5e5ed6682a83b529b8a10800b3d2029ac299c0bc3b0ed7bd7491197bbac" Mar 07 15:12:15 crc kubenswrapper[4943]: I0307 15:12:15.503775 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mln5" Mar 07 15:12:16 crc kubenswrapper[4943]: I0307 15:12:16.067526 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-6n97z"] Mar 07 15:12:16 crc kubenswrapper[4943]: E0307 15:12:16.068378 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0814782b-ab61-4cfe-8088-2036b434f236" containerName="swift-ring-rebalance" Mar 07 15:12:16 crc kubenswrapper[4943]: I0307 15:12:16.068401 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="0814782b-ab61-4cfe-8088-2036b434f236" containerName="swift-ring-rebalance" Mar 07 15:12:16 crc kubenswrapper[4943]: I0307 15:12:16.068647 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="0814782b-ab61-4cfe-8088-2036b434f236" containerName="swift-ring-rebalance" Mar 07 15:12:16 crc kubenswrapper[4943]: I0307 15:12:16.069426 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6n97z" Mar 07 15:12:16 crc kubenswrapper[4943]: I0307 15:12:16.072008 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:12:16 crc kubenswrapper[4943]: I0307 15:12:16.072115 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:12:16 crc kubenswrapper[4943]: I0307 15:12:16.082312 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-6n97z"] Mar 07 15:12:16 crc kubenswrapper[4943]: I0307 15:12:16.197197 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c0016496-2b61-4875-8a2f-f262e2669c4b-dispersionconf\") pod \"swift-ring-rebalance-debug-6n97z\" (UID: \"c0016496-2b61-4875-8a2f-f262e2669c4b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6n97z" Mar 07 15:12:16 crc kubenswrapper[4943]: I0307 15:12:16.197250 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c0016496-2b61-4875-8a2f-f262e2669c4b-etc-swift\") pod \"swift-ring-rebalance-debug-6n97z\" (UID: \"c0016496-2b61-4875-8a2f-f262e2669c4b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6n97z" Mar 07 15:12:16 crc kubenswrapper[4943]: I0307 15:12:16.197283 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c0016496-2b61-4875-8a2f-f262e2669c4b-ring-data-devices\") pod \"swift-ring-rebalance-debug-6n97z\" (UID: \"c0016496-2b61-4875-8a2f-f262e2669c4b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6n97z" Mar 07 15:12:16 crc kubenswrapper[4943]: I0307 15:12:16.197917 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9v7j\" (UniqueName: \"kubernetes.io/projected/c0016496-2b61-4875-8a2f-f262e2669c4b-kube-api-access-m9v7j\") pod \"swift-ring-rebalance-debug-6n97z\" (UID: \"c0016496-2b61-4875-8a2f-f262e2669c4b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6n97z" Mar 07 15:12:16 crc kubenswrapper[4943]: I0307 15:12:16.197982 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c0016496-2b61-4875-8a2f-f262e2669c4b-swiftconf\") pod \"swift-ring-rebalance-debug-6n97z\" (UID: \"c0016496-2b61-4875-8a2f-f262e2669c4b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6n97z" Mar 07 15:12:16 crc kubenswrapper[4943]: I0307 15:12:16.198071 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c0016496-2b61-4875-8a2f-f262e2669c4b-scripts\") pod \"swift-ring-rebalance-debug-6n97z\" (UID: \"c0016496-2b61-4875-8a2f-f262e2669c4b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6n97z" Mar 07 15:12:16 crc kubenswrapper[4943]: I0307 15:12:16.299827 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c0016496-2b61-4875-8a2f-f262e2669c4b-scripts\") pod \"swift-ring-rebalance-debug-6n97z\" (UID: \"c0016496-2b61-4875-8a2f-f262e2669c4b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6n97z" Mar 07 15:12:16 crc kubenswrapper[4943]: I0307 15:12:16.299918 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c0016496-2b61-4875-8a2f-f262e2669c4b-dispersionconf\") pod \"swift-ring-rebalance-debug-6n97z\" (UID: \"c0016496-2b61-4875-8a2f-f262e2669c4b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6n97z" Mar 07 15:12:16 crc kubenswrapper[4943]: I0307 15:12:16.300036 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c0016496-2b61-4875-8a2f-f262e2669c4b-etc-swift\") pod \"swift-ring-rebalance-debug-6n97z\" (UID: \"c0016496-2b61-4875-8a2f-f262e2669c4b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6n97z" Mar 07 15:12:16 crc kubenswrapper[4943]: I0307 15:12:16.300079 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c0016496-2b61-4875-8a2f-f262e2669c4b-ring-data-devices\") pod \"swift-ring-rebalance-debug-6n97z\" (UID: \"c0016496-2b61-4875-8a2f-f262e2669c4b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6n97z" Mar 07 15:12:16 crc kubenswrapper[4943]: I0307 15:12:16.300152 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9v7j\" (UniqueName: \"kubernetes.io/projected/c0016496-2b61-4875-8a2f-f262e2669c4b-kube-api-access-m9v7j\") pod \"swift-ring-rebalance-debug-6n97z\" (UID: \"c0016496-2b61-4875-8a2f-f262e2669c4b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6n97z" Mar 07 15:12:16 crc kubenswrapper[4943]: I0307 15:12:16.300205 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c0016496-2b61-4875-8a2f-f262e2669c4b-swiftconf\") pod \"swift-ring-rebalance-debug-6n97z\" (UID: \"c0016496-2b61-4875-8a2f-f262e2669c4b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6n97z" Mar 07 15:12:16 crc kubenswrapper[4943]: I0307 15:12:16.301303 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c0016496-2b61-4875-8a2f-f262e2669c4b-etc-swift\") pod \"swift-ring-rebalance-debug-6n97z\" (UID: \"c0016496-2b61-4875-8a2f-f262e2669c4b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6n97z" Mar 07 15:12:16 crc kubenswrapper[4943]: I0307 15:12:16.301534 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c0016496-2b61-4875-8a2f-f262e2669c4b-scripts\") pod \"swift-ring-rebalance-debug-6n97z\" (UID: \"c0016496-2b61-4875-8a2f-f262e2669c4b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6n97z" Mar 07 15:12:16 crc kubenswrapper[4943]: I0307 15:12:16.302144 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c0016496-2b61-4875-8a2f-f262e2669c4b-ring-data-devices\") pod \"swift-ring-rebalance-debug-6n97z\" (UID: \"c0016496-2b61-4875-8a2f-f262e2669c4b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6n97z" Mar 07 15:12:16 crc kubenswrapper[4943]: I0307 15:12:16.305824 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c0016496-2b61-4875-8a2f-f262e2669c4b-dispersionconf\") pod \"swift-ring-rebalance-debug-6n97z\" (UID: \"c0016496-2b61-4875-8a2f-f262e2669c4b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6n97z" Mar 07 15:12:16 crc kubenswrapper[4943]: I0307 15:12:16.310531 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c0016496-2b61-4875-8a2f-f262e2669c4b-swiftconf\") pod \"swift-ring-rebalance-debug-6n97z\" (UID: \"c0016496-2b61-4875-8a2f-f262e2669c4b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6n97z" Mar 07 15:12:16 crc kubenswrapper[4943]: I0307 15:12:16.327177 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9v7j\" (UniqueName: \"kubernetes.io/projected/c0016496-2b61-4875-8a2f-f262e2669c4b-kube-api-access-m9v7j\") pod \"swift-ring-rebalance-debug-6n97z\" (UID: \"c0016496-2b61-4875-8a2f-f262e2669c4b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6n97z" Mar 07 15:12:16 crc kubenswrapper[4943]: I0307 15:12:16.396864 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6n97z" Mar 07 15:12:16 crc kubenswrapper[4943]: I0307 15:12:16.731246 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-6n97z"] Mar 07 15:12:16 crc kubenswrapper[4943]: I0307 15:12:16.767659 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0814782b-ab61-4cfe-8088-2036b434f236" path="/var/lib/kubelet/pods/0814782b-ab61-4cfe-8088-2036b434f236/volumes" Mar 07 15:12:17 crc kubenswrapper[4943]: I0307 15:12:17.536651 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6n97z" event={"ID":"c0016496-2b61-4875-8a2f-f262e2669c4b","Type":"ContainerStarted","Data":"cd807c94f54160ad472a6127781bd7d1138328a5f52b97d1acfb48a8bfecf050"} Mar 07 15:12:17 crc kubenswrapper[4943]: I0307 15:12:17.537135 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6n97z" event={"ID":"c0016496-2b61-4875-8a2f-f262e2669c4b","Type":"ContainerStarted","Data":"8db88eb049985533ace156883ce57ef659f928bf6e2da8af029df05104ec894b"} Mar 07 15:12:17 crc kubenswrapper[4943]: I0307 15:12:17.561091 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6n97z" podStartSLOduration=1.5610647100000001 podStartE2EDuration="1.56106471s" podCreationTimestamp="2026-03-07 15:12:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:12:17.551543237 +0000 UTC m=+1979.503679745" watchObservedRunningTime="2026-03-07 15:12:17.56106471 +0000 UTC m=+1979.513201218" Mar 07 15:12:18 crc kubenswrapper[4943]: I0307 15:12:18.547873 4943 generic.go:334] "Generic (PLEG): container finished" podID="c0016496-2b61-4875-8a2f-f262e2669c4b" containerID="cd807c94f54160ad472a6127781bd7d1138328a5f52b97d1acfb48a8bfecf050" exitCode=0 Mar 07 15:12:18 crc kubenswrapper[4943]: I0307 15:12:18.549155 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6n97z" event={"ID":"c0016496-2b61-4875-8a2f-f262e2669c4b","Type":"ContainerDied","Data":"cd807c94f54160ad472a6127781bd7d1138328a5f52b97d1acfb48a8bfecf050"} Mar 07 15:12:19 crc kubenswrapper[4943]: I0307 15:12:19.933097 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6n97z" Mar 07 15:12:19 crc kubenswrapper[4943]: I0307 15:12:19.961924 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c0016496-2b61-4875-8a2f-f262e2669c4b-dispersionconf\") pod \"c0016496-2b61-4875-8a2f-f262e2669c4b\" (UID: \"c0016496-2b61-4875-8a2f-f262e2669c4b\") " Mar 07 15:12:19 crc kubenswrapper[4943]: I0307 15:12:19.962042 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c0016496-2b61-4875-8a2f-f262e2669c4b-ring-data-devices\") pod \"c0016496-2b61-4875-8a2f-f262e2669c4b\" (UID: \"c0016496-2b61-4875-8a2f-f262e2669c4b\") " Mar 07 15:12:19 crc kubenswrapper[4943]: I0307 15:12:19.962104 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c0016496-2b61-4875-8a2f-f262e2669c4b-etc-swift\") pod \"c0016496-2b61-4875-8a2f-f262e2669c4b\" (UID: \"c0016496-2b61-4875-8a2f-f262e2669c4b\") " Mar 07 15:12:19 crc kubenswrapper[4943]: I0307 15:12:19.962142 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c0016496-2b61-4875-8a2f-f262e2669c4b-scripts\") pod \"c0016496-2b61-4875-8a2f-f262e2669c4b\" (UID: \"c0016496-2b61-4875-8a2f-f262e2669c4b\") " Mar 07 15:12:19 crc kubenswrapper[4943]: I0307 15:12:19.962179 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c0016496-2b61-4875-8a2f-f262e2669c4b-swiftconf\") pod \"c0016496-2b61-4875-8a2f-f262e2669c4b\" (UID: \"c0016496-2b61-4875-8a2f-f262e2669c4b\") " Mar 07 15:12:19 crc kubenswrapper[4943]: I0307 15:12:19.962223 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m9v7j\" (UniqueName: \"kubernetes.io/projected/c0016496-2b61-4875-8a2f-f262e2669c4b-kube-api-access-m9v7j\") pod \"c0016496-2b61-4875-8a2f-f262e2669c4b\" (UID: \"c0016496-2b61-4875-8a2f-f262e2669c4b\") " Mar 07 15:12:19 crc kubenswrapper[4943]: I0307 15:12:19.965577 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c0016496-2b61-4875-8a2f-f262e2669c4b-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "c0016496-2b61-4875-8a2f-f262e2669c4b" (UID: "c0016496-2b61-4875-8a2f-f262e2669c4b"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:12:19 crc kubenswrapper[4943]: I0307 15:12:19.966700 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c0016496-2b61-4875-8a2f-f262e2669c4b-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "c0016496-2b61-4875-8a2f-f262e2669c4b" (UID: "c0016496-2b61-4875-8a2f-f262e2669c4b"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:12:19 crc kubenswrapper[4943]: I0307 15:12:19.972678 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0016496-2b61-4875-8a2f-f262e2669c4b-kube-api-access-m9v7j" (OuterVolumeSpecName: "kube-api-access-m9v7j") pod "c0016496-2b61-4875-8a2f-f262e2669c4b" (UID: "c0016496-2b61-4875-8a2f-f262e2669c4b"). InnerVolumeSpecName "kube-api-access-m9v7j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:12:19 crc kubenswrapper[4943]: I0307 15:12:19.994887 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0016496-2b61-4875-8a2f-f262e2669c4b-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "c0016496-2b61-4875-8a2f-f262e2669c4b" (UID: "c0016496-2b61-4875-8a2f-f262e2669c4b"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:12:20 crc kubenswrapper[4943]: I0307 15:12:20.000337 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0016496-2b61-4875-8a2f-f262e2669c4b-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "c0016496-2b61-4875-8a2f-f262e2669c4b" (UID: "c0016496-2b61-4875-8a2f-f262e2669c4b"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:12:20 crc kubenswrapper[4943]: I0307 15:12:20.000734 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c0016496-2b61-4875-8a2f-f262e2669c4b-scripts" (OuterVolumeSpecName: "scripts") pod "c0016496-2b61-4875-8a2f-f262e2669c4b" (UID: "c0016496-2b61-4875-8a2f-f262e2669c4b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:12:20 crc kubenswrapper[4943]: I0307 15:12:20.002297 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-6n97z"] Mar 07 15:12:20 crc kubenswrapper[4943]: I0307 15:12:20.017279 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-6n97z"] Mar 07 15:12:20 crc kubenswrapper[4943]: I0307 15:12:20.064012 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c0016496-2b61-4875-8a2f-f262e2669c4b-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:20 crc kubenswrapper[4943]: I0307 15:12:20.064044 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c0016496-2b61-4875-8a2f-f262e2669c4b-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:20 crc kubenswrapper[4943]: I0307 15:12:20.064056 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c0016496-2b61-4875-8a2f-f262e2669c4b-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:20 crc kubenswrapper[4943]: I0307 15:12:20.064065 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c0016496-2b61-4875-8a2f-f262e2669c4b-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:20 crc kubenswrapper[4943]: I0307 15:12:20.064075 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c0016496-2b61-4875-8a2f-f262e2669c4b-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:20 crc kubenswrapper[4943]: I0307 15:12:20.064084 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m9v7j\" (UniqueName: \"kubernetes.io/projected/c0016496-2b61-4875-8a2f-f262e2669c4b-kube-api-access-m9v7j\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:20 crc kubenswrapper[4943]: I0307 15:12:20.570362 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8db88eb049985533ace156883ce57ef659f928bf6e2da8af029df05104ec894b" Mar 07 15:12:20 crc kubenswrapper[4943]: I0307 15:12:20.570491 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6n97z" Mar 07 15:12:20 crc kubenswrapper[4943]: I0307 15:12:20.770432 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0016496-2b61-4875-8a2f-f262e2669c4b" path="/var/lib/kubelet/pods/c0016496-2b61-4875-8a2f-f262e2669c4b/volumes" Mar 07 15:12:21 crc kubenswrapper[4943]: I0307 15:12:21.172128 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-b9p9c"] Mar 07 15:12:21 crc kubenswrapper[4943]: E0307 15:12:21.172528 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0016496-2b61-4875-8a2f-f262e2669c4b" containerName="swift-ring-rebalance" Mar 07 15:12:21 crc kubenswrapper[4943]: I0307 15:12:21.172548 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0016496-2b61-4875-8a2f-f262e2669c4b" containerName="swift-ring-rebalance" Mar 07 15:12:21 crc kubenswrapper[4943]: I0307 15:12:21.172865 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0016496-2b61-4875-8a2f-f262e2669c4b" containerName="swift-ring-rebalance" Mar 07 15:12:21 crc kubenswrapper[4943]: I0307 15:12:21.173591 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-b9p9c" Mar 07 15:12:21 crc kubenswrapper[4943]: I0307 15:12:21.181116 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:12:21 crc kubenswrapper[4943]: I0307 15:12:21.181391 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:12:21 crc kubenswrapper[4943]: I0307 15:12:21.181534 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/52b02ad4-9f4c-4c77-881d-116b8aaf03bc-dispersionconf\") pod \"swift-ring-rebalance-debug-b9p9c\" (UID: \"52b02ad4-9f4c-4c77-881d-116b8aaf03bc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-b9p9c" Mar 07 15:12:21 crc kubenswrapper[4943]: I0307 15:12:21.181589 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/52b02ad4-9f4c-4c77-881d-116b8aaf03bc-etc-swift\") pod \"swift-ring-rebalance-debug-b9p9c\" (UID: \"52b02ad4-9f4c-4c77-881d-116b8aaf03bc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-b9p9c" Mar 07 15:12:21 crc kubenswrapper[4943]: I0307 15:12:21.181665 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/52b02ad4-9f4c-4c77-881d-116b8aaf03bc-scripts\") pod \"swift-ring-rebalance-debug-b9p9c\" (UID: \"52b02ad4-9f4c-4c77-881d-116b8aaf03bc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-b9p9c" Mar 07 15:12:21 crc kubenswrapper[4943]: I0307 15:12:21.181697 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/52b02ad4-9f4c-4c77-881d-116b8aaf03bc-ring-data-devices\") pod \"swift-ring-rebalance-debug-b9p9c\" (UID: \"52b02ad4-9f4c-4c77-881d-116b8aaf03bc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-b9p9c" Mar 07 15:12:21 crc kubenswrapper[4943]: I0307 15:12:21.181874 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/52b02ad4-9f4c-4c77-881d-116b8aaf03bc-swiftconf\") pod \"swift-ring-rebalance-debug-b9p9c\" (UID: \"52b02ad4-9f4c-4c77-881d-116b8aaf03bc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-b9p9c" Mar 07 15:12:21 crc kubenswrapper[4943]: I0307 15:12:21.182105 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9cp6n\" (UniqueName: \"kubernetes.io/projected/52b02ad4-9f4c-4c77-881d-116b8aaf03bc-kube-api-access-9cp6n\") pod \"swift-ring-rebalance-debug-b9p9c\" (UID: \"52b02ad4-9f4c-4c77-881d-116b8aaf03bc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-b9p9c" Mar 07 15:12:21 crc kubenswrapper[4943]: I0307 15:12:21.184350 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-b9p9c"] Mar 07 15:12:21 crc kubenswrapper[4943]: I0307 15:12:21.284466 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/52b02ad4-9f4c-4c77-881d-116b8aaf03bc-dispersionconf\") pod \"swift-ring-rebalance-debug-b9p9c\" (UID: \"52b02ad4-9f4c-4c77-881d-116b8aaf03bc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-b9p9c" Mar 07 15:12:21 crc kubenswrapper[4943]: I0307 15:12:21.284544 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/52b02ad4-9f4c-4c77-881d-116b8aaf03bc-etc-swift\") pod \"swift-ring-rebalance-debug-b9p9c\" (UID: \"52b02ad4-9f4c-4c77-881d-116b8aaf03bc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-b9p9c" Mar 07 15:12:21 crc kubenswrapper[4943]: I0307 15:12:21.284606 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/52b02ad4-9f4c-4c77-881d-116b8aaf03bc-scripts\") pod \"swift-ring-rebalance-debug-b9p9c\" (UID: \"52b02ad4-9f4c-4c77-881d-116b8aaf03bc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-b9p9c" Mar 07 15:12:21 crc kubenswrapper[4943]: I0307 15:12:21.284646 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/52b02ad4-9f4c-4c77-881d-116b8aaf03bc-ring-data-devices\") pod \"swift-ring-rebalance-debug-b9p9c\" (UID: \"52b02ad4-9f4c-4c77-881d-116b8aaf03bc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-b9p9c" Mar 07 15:12:21 crc kubenswrapper[4943]: I0307 15:12:21.284697 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/52b02ad4-9f4c-4c77-881d-116b8aaf03bc-swiftconf\") pod \"swift-ring-rebalance-debug-b9p9c\" (UID: \"52b02ad4-9f4c-4c77-881d-116b8aaf03bc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-b9p9c" Mar 07 15:12:21 crc kubenswrapper[4943]: I0307 15:12:21.284752 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9cp6n\" (UniqueName: \"kubernetes.io/projected/52b02ad4-9f4c-4c77-881d-116b8aaf03bc-kube-api-access-9cp6n\") pod \"swift-ring-rebalance-debug-b9p9c\" (UID: \"52b02ad4-9f4c-4c77-881d-116b8aaf03bc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-b9p9c" Mar 07 15:12:21 crc kubenswrapper[4943]: I0307 15:12:21.285484 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/52b02ad4-9f4c-4c77-881d-116b8aaf03bc-etc-swift\") pod \"swift-ring-rebalance-debug-b9p9c\" (UID: \"52b02ad4-9f4c-4c77-881d-116b8aaf03bc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-b9p9c" Mar 07 15:12:21 crc kubenswrapper[4943]: I0307 15:12:21.285851 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/52b02ad4-9f4c-4c77-881d-116b8aaf03bc-ring-data-devices\") pod \"swift-ring-rebalance-debug-b9p9c\" (UID: \"52b02ad4-9f4c-4c77-881d-116b8aaf03bc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-b9p9c" Mar 07 15:12:21 crc kubenswrapper[4943]: I0307 15:12:21.286728 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/52b02ad4-9f4c-4c77-881d-116b8aaf03bc-scripts\") pod \"swift-ring-rebalance-debug-b9p9c\" (UID: \"52b02ad4-9f4c-4c77-881d-116b8aaf03bc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-b9p9c" Mar 07 15:12:21 crc kubenswrapper[4943]: I0307 15:12:21.289162 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/52b02ad4-9f4c-4c77-881d-116b8aaf03bc-dispersionconf\") pod \"swift-ring-rebalance-debug-b9p9c\" (UID: \"52b02ad4-9f4c-4c77-881d-116b8aaf03bc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-b9p9c" Mar 07 15:12:21 crc kubenswrapper[4943]: I0307 15:12:21.289525 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/52b02ad4-9f4c-4c77-881d-116b8aaf03bc-swiftconf\") pod \"swift-ring-rebalance-debug-b9p9c\" (UID: \"52b02ad4-9f4c-4c77-881d-116b8aaf03bc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-b9p9c" Mar 07 15:12:21 crc kubenswrapper[4943]: I0307 15:12:21.309316 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9cp6n\" (UniqueName: \"kubernetes.io/projected/52b02ad4-9f4c-4c77-881d-116b8aaf03bc-kube-api-access-9cp6n\") pod \"swift-ring-rebalance-debug-b9p9c\" (UID: \"52b02ad4-9f4c-4c77-881d-116b8aaf03bc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-b9p9c" Mar 07 15:12:21 crc kubenswrapper[4943]: I0307 15:12:21.387997 4943 scope.go:117] "RemoveContainer" containerID="05fe59cccad959e643fc20072c6b5a061c26fac692faa569a5c6d8519505b13d" Mar 07 15:12:21 crc kubenswrapper[4943]: I0307 15:12:21.541714 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-b9p9c" Mar 07 15:12:22 crc kubenswrapper[4943]: I0307 15:12:22.349207 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-b9p9c"] Mar 07 15:12:22 crc kubenswrapper[4943]: W0307 15:12:22.357894 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod52b02ad4_9f4c_4c77_881d_116b8aaf03bc.slice/crio-c0a06b2b4d2c481a353ae62ba3948bc0b7f2fa2162650a7e4791540169519d23 WatchSource:0}: Error finding container c0a06b2b4d2c481a353ae62ba3948bc0b7f2fa2162650a7e4791540169519d23: Status 404 returned error can't find the container with id c0a06b2b4d2c481a353ae62ba3948bc0b7f2fa2162650a7e4791540169519d23 Mar 07 15:12:22 crc kubenswrapper[4943]: I0307 15:12:22.590515 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-b9p9c" event={"ID":"52b02ad4-9f4c-4c77-881d-116b8aaf03bc","Type":"ContainerStarted","Data":"43aa2e98bb78522353924a86521f72f9a0b8723d89b7db6e30ef90d1cb0e943f"} Mar 07 15:12:22 crc kubenswrapper[4943]: I0307 15:12:22.590557 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-b9p9c" event={"ID":"52b02ad4-9f4c-4c77-881d-116b8aaf03bc","Type":"ContainerStarted","Data":"c0a06b2b4d2c481a353ae62ba3948bc0b7f2fa2162650a7e4791540169519d23"} Mar 07 15:12:22 crc kubenswrapper[4943]: I0307 15:12:22.612277 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-b9p9c" podStartSLOduration=1.612254363 podStartE2EDuration="1.612254363s" podCreationTimestamp="2026-03-07 15:12:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:12:22.605953439 +0000 UTC m=+1984.558089957" watchObservedRunningTime="2026-03-07 15:12:22.612254363 +0000 UTC m=+1984.564390861" Mar 07 15:12:24 crc kubenswrapper[4943]: I0307 15:12:24.610817 4943 generic.go:334] "Generic (PLEG): container finished" podID="52b02ad4-9f4c-4c77-881d-116b8aaf03bc" containerID="43aa2e98bb78522353924a86521f72f9a0b8723d89b7db6e30ef90d1cb0e943f" exitCode=0 Mar 07 15:12:24 crc kubenswrapper[4943]: I0307 15:12:24.610966 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-b9p9c" event={"ID":"52b02ad4-9f4c-4c77-881d-116b8aaf03bc","Type":"ContainerDied","Data":"43aa2e98bb78522353924a86521f72f9a0b8723d89b7db6e30ef90d1cb0e943f"} Mar 07 15:12:25 crc kubenswrapper[4943]: I0307 15:12:25.961485 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-b9p9c" Mar 07 15:12:26 crc kubenswrapper[4943]: I0307 15:12:26.007364 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-b9p9c"] Mar 07 15:12:26 crc kubenswrapper[4943]: I0307 15:12:26.022228 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-b9p9c"] Mar 07 15:12:26 crc kubenswrapper[4943]: I0307 15:12:26.159296 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/52b02ad4-9f4c-4c77-881d-116b8aaf03bc-dispersionconf\") pod \"52b02ad4-9f4c-4c77-881d-116b8aaf03bc\" (UID: \"52b02ad4-9f4c-4c77-881d-116b8aaf03bc\") " Mar 07 15:12:26 crc kubenswrapper[4943]: I0307 15:12:26.159350 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/52b02ad4-9f4c-4c77-881d-116b8aaf03bc-etc-swift\") pod \"52b02ad4-9f4c-4c77-881d-116b8aaf03bc\" (UID: \"52b02ad4-9f4c-4c77-881d-116b8aaf03bc\") " Mar 07 15:12:26 crc kubenswrapper[4943]: I0307 15:12:26.159386 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/52b02ad4-9f4c-4c77-881d-116b8aaf03bc-ring-data-devices\") pod \"52b02ad4-9f4c-4c77-881d-116b8aaf03bc\" (UID: \"52b02ad4-9f4c-4c77-881d-116b8aaf03bc\") " Mar 07 15:12:26 crc kubenswrapper[4943]: I0307 15:12:26.159450 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/52b02ad4-9f4c-4c77-881d-116b8aaf03bc-swiftconf\") pod \"52b02ad4-9f4c-4c77-881d-116b8aaf03bc\" (UID: \"52b02ad4-9f4c-4c77-881d-116b8aaf03bc\") " Mar 07 15:12:26 crc kubenswrapper[4943]: I0307 15:12:26.159489 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/52b02ad4-9f4c-4c77-881d-116b8aaf03bc-scripts\") pod \"52b02ad4-9f4c-4c77-881d-116b8aaf03bc\" (UID: \"52b02ad4-9f4c-4c77-881d-116b8aaf03bc\") " Mar 07 15:12:26 crc kubenswrapper[4943]: I0307 15:12:26.159585 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9cp6n\" (UniqueName: \"kubernetes.io/projected/52b02ad4-9f4c-4c77-881d-116b8aaf03bc-kube-api-access-9cp6n\") pod \"52b02ad4-9f4c-4c77-881d-116b8aaf03bc\" (UID: \"52b02ad4-9f4c-4c77-881d-116b8aaf03bc\") " Mar 07 15:12:26 crc kubenswrapper[4943]: I0307 15:12:26.160410 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/52b02ad4-9f4c-4c77-881d-116b8aaf03bc-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "52b02ad4-9f4c-4c77-881d-116b8aaf03bc" (UID: "52b02ad4-9f4c-4c77-881d-116b8aaf03bc"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:12:26 crc kubenswrapper[4943]: I0307 15:12:26.160755 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52b02ad4-9f4c-4c77-881d-116b8aaf03bc-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "52b02ad4-9f4c-4c77-881d-116b8aaf03bc" (UID: "52b02ad4-9f4c-4c77-881d-116b8aaf03bc"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:12:26 crc kubenswrapper[4943]: I0307 15:12:26.167338 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52b02ad4-9f4c-4c77-881d-116b8aaf03bc-kube-api-access-9cp6n" (OuterVolumeSpecName: "kube-api-access-9cp6n") pod "52b02ad4-9f4c-4c77-881d-116b8aaf03bc" (UID: "52b02ad4-9f4c-4c77-881d-116b8aaf03bc"). InnerVolumeSpecName "kube-api-access-9cp6n". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:12:26 crc kubenswrapper[4943]: I0307 15:12:26.183487 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52b02ad4-9f4c-4c77-881d-116b8aaf03bc-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "52b02ad4-9f4c-4c77-881d-116b8aaf03bc" (UID: "52b02ad4-9f4c-4c77-881d-116b8aaf03bc"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:12:26 crc kubenswrapper[4943]: I0307 15:12:26.184376 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52b02ad4-9f4c-4c77-881d-116b8aaf03bc-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "52b02ad4-9f4c-4c77-881d-116b8aaf03bc" (UID: "52b02ad4-9f4c-4c77-881d-116b8aaf03bc"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:12:26 crc kubenswrapper[4943]: I0307 15:12:26.195014 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52b02ad4-9f4c-4c77-881d-116b8aaf03bc-scripts" (OuterVolumeSpecName: "scripts") pod "52b02ad4-9f4c-4c77-881d-116b8aaf03bc" (UID: "52b02ad4-9f4c-4c77-881d-116b8aaf03bc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:12:26 crc kubenswrapper[4943]: I0307 15:12:26.261170 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9cp6n\" (UniqueName: \"kubernetes.io/projected/52b02ad4-9f4c-4c77-881d-116b8aaf03bc-kube-api-access-9cp6n\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:26 crc kubenswrapper[4943]: I0307 15:12:26.261232 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/52b02ad4-9f4c-4c77-881d-116b8aaf03bc-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:26 crc kubenswrapper[4943]: I0307 15:12:26.261247 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/52b02ad4-9f4c-4c77-881d-116b8aaf03bc-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:26 crc kubenswrapper[4943]: I0307 15:12:26.261261 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/52b02ad4-9f4c-4c77-881d-116b8aaf03bc-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:26 crc kubenswrapper[4943]: I0307 15:12:26.261274 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/52b02ad4-9f4c-4c77-881d-116b8aaf03bc-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:26 crc kubenswrapper[4943]: I0307 15:12:26.261286 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/52b02ad4-9f4c-4c77-881d-116b8aaf03bc-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:26 crc kubenswrapper[4943]: I0307 15:12:26.633543 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c0a06b2b4d2c481a353ae62ba3948bc0b7f2fa2162650a7e4791540169519d23" Mar 07 15:12:26 crc kubenswrapper[4943]: I0307 15:12:26.633601 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-b9p9c" Mar 07 15:12:26 crc kubenswrapper[4943]: I0307 15:12:26.767072 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52b02ad4-9f4c-4c77-881d-116b8aaf03bc" path="/var/lib/kubelet/pods/52b02ad4-9f4c-4c77-881d-116b8aaf03bc/volumes" Mar 07 15:12:27 crc kubenswrapper[4943]: I0307 15:12:27.168154 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-4lx9r"] Mar 07 15:12:27 crc kubenswrapper[4943]: E0307 15:12:27.168435 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52b02ad4-9f4c-4c77-881d-116b8aaf03bc" containerName="swift-ring-rebalance" Mar 07 15:12:27 crc kubenswrapper[4943]: I0307 15:12:27.168449 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="52b02ad4-9f4c-4c77-881d-116b8aaf03bc" containerName="swift-ring-rebalance" Mar 07 15:12:27 crc kubenswrapper[4943]: I0307 15:12:27.168592 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="52b02ad4-9f4c-4c77-881d-116b8aaf03bc" containerName="swift-ring-rebalance" Mar 07 15:12:27 crc kubenswrapper[4943]: I0307 15:12:27.169075 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-4lx9r" Mar 07 15:12:27 crc kubenswrapper[4943]: I0307 15:12:27.170581 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:12:27 crc kubenswrapper[4943]: I0307 15:12:27.171035 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:12:27 crc kubenswrapper[4943]: I0307 15:12:27.183280 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-4lx9r"] Mar 07 15:12:27 crc kubenswrapper[4943]: I0307 15:12:27.277609 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/8b6b168d-d695-4916-a11c-810b697d1e8a-swiftconf\") pod \"swift-ring-rebalance-debug-4lx9r\" (UID: \"8b6b168d-d695-4916-a11c-810b697d1e8a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4lx9r" Mar 07 15:12:27 crc kubenswrapper[4943]: I0307 15:12:27.277918 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8b6b168d-d695-4916-a11c-810b697d1e8a-scripts\") pod \"swift-ring-rebalance-debug-4lx9r\" (UID: \"8b6b168d-d695-4916-a11c-810b697d1e8a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4lx9r" Mar 07 15:12:27 crc kubenswrapper[4943]: I0307 15:12:27.278010 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/8b6b168d-d695-4916-a11c-810b697d1e8a-dispersionconf\") pod \"swift-ring-rebalance-debug-4lx9r\" (UID: \"8b6b168d-d695-4916-a11c-810b697d1e8a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4lx9r" Mar 07 15:12:27 crc kubenswrapper[4943]: I0307 15:12:27.278075 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqd4z\" (UniqueName: \"kubernetes.io/projected/8b6b168d-d695-4916-a11c-810b697d1e8a-kube-api-access-tqd4z\") pod \"swift-ring-rebalance-debug-4lx9r\" (UID: \"8b6b168d-d695-4916-a11c-810b697d1e8a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4lx9r" Mar 07 15:12:27 crc kubenswrapper[4943]: I0307 15:12:27.278145 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/8b6b168d-d695-4916-a11c-810b697d1e8a-etc-swift\") pod \"swift-ring-rebalance-debug-4lx9r\" (UID: \"8b6b168d-d695-4916-a11c-810b697d1e8a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4lx9r" Mar 07 15:12:27 crc kubenswrapper[4943]: I0307 15:12:27.278185 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/8b6b168d-d695-4916-a11c-810b697d1e8a-ring-data-devices\") pod \"swift-ring-rebalance-debug-4lx9r\" (UID: \"8b6b168d-d695-4916-a11c-810b697d1e8a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4lx9r" Mar 07 15:12:27 crc kubenswrapper[4943]: I0307 15:12:27.378994 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/8b6b168d-d695-4916-a11c-810b697d1e8a-etc-swift\") pod \"swift-ring-rebalance-debug-4lx9r\" (UID: \"8b6b168d-d695-4916-a11c-810b697d1e8a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4lx9r" Mar 07 15:12:27 crc kubenswrapper[4943]: I0307 15:12:27.379060 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/8b6b168d-d695-4916-a11c-810b697d1e8a-ring-data-devices\") pod \"swift-ring-rebalance-debug-4lx9r\" (UID: \"8b6b168d-d695-4916-a11c-810b697d1e8a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4lx9r" Mar 07 15:12:27 crc kubenswrapper[4943]: I0307 15:12:27.379102 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/8b6b168d-d695-4916-a11c-810b697d1e8a-swiftconf\") pod \"swift-ring-rebalance-debug-4lx9r\" (UID: \"8b6b168d-d695-4916-a11c-810b697d1e8a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4lx9r" Mar 07 15:12:27 crc kubenswrapper[4943]: I0307 15:12:27.379130 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8b6b168d-d695-4916-a11c-810b697d1e8a-scripts\") pod \"swift-ring-rebalance-debug-4lx9r\" (UID: \"8b6b168d-d695-4916-a11c-810b697d1e8a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4lx9r" Mar 07 15:12:27 crc kubenswrapper[4943]: I0307 15:12:27.379173 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/8b6b168d-d695-4916-a11c-810b697d1e8a-dispersionconf\") pod \"swift-ring-rebalance-debug-4lx9r\" (UID: \"8b6b168d-d695-4916-a11c-810b697d1e8a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4lx9r" Mar 07 15:12:27 crc kubenswrapper[4943]: I0307 15:12:27.379203 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqd4z\" (UniqueName: \"kubernetes.io/projected/8b6b168d-d695-4916-a11c-810b697d1e8a-kube-api-access-tqd4z\") pod \"swift-ring-rebalance-debug-4lx9r\" (UID: \"8b6b168d-d695-4916-a11c-810b697d1e8a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4lx9r" Mar 07 15:12:27 crc kubenswrapper[4943]: I0307 15:12:27.380165 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/8b6b168d-d695-4916-a11c-810b697d1e8a-etc-swift\") pod \"swift-ring-rebalance-debug-4lx9r\" (UID: \"8b6b168d-d695-4916-a11c-810b697d1e8a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4lx9r" Mar 07 15:12:27 crc kubenswrapper[4943]: I0307 15:12:27.380760 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/8b6b168d-d695-4916-a11c-810b697d1e8a-ring-data-devices\") pod \"swift-ring-rebalance-debug-4lx9r\" (UID: \"8b6b168d-d695-4916-a11c-810b697d1e8a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4lx9r" Mar 07 15:12:27 crc kubenswrapper[4943]: I0307 15:12:27.381729 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8b6b168d-d695-4916-a11c-810b697d1e8a-scripts\") pod \"swift-ring-rebalance-debug-4lx9r\" (UID: \"8b6b168d-d695-4916-a11c-810b697d1e8a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4lx9r" Mar 07 15:12:27 crc kubenswrapper[4943]: I0307 15:12:27.385188 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/8b6b168d-d695-4916-a11c-810b697d1e8a-dispersionconf\") pod \"swift-ring-rebalance-debug-4lx9r\" (UID: \"8b6b168d-d695-4916-a11c-810b697d1e8a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4lx9r" Mar 07 15:12:27 crc kubenswrapper[4943]: I0307 15:12:27.385358 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/8b6b168d-d695-4916-a11c-810b697d1e8a-swiftconf\") pod \"swift-ring-rebalance-debug-4lx9r\" (UID: \"8b6b168d-d695-4916-a11c-810b697d1e8a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4lx9r" Mar 07 15:12:27 crc kubenswrapper[4943]: I0307 15:12:27.396488 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqd4z\" (UniqueName: \"kubernetes.io/projected/8b6b168d-d695-4916-a11c-810b697d1e8a-kube-api-access-tqd4z\") pod \"swift-ring-rebalance-debug-4lx9r\" (UID: \"8b6b168d-d695-4916-a11c-810b697d1e8a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4lx9r" Mar 07 15:12:27 crc kubenswrapper[4943]: I0307 15:12:27.492421 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-4lx9r" Mar 07 15:12:27 crc kubenswrapper[4943]: I0307 15:12:27.975348 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-4lx9r"] Mar 07 15:12:28 crc kubenswrapper[4943]: I0307 15:12:28.651198 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-4lx9r" event={"ID":"8b6b168d-d695-4916-a11c-810b697d1e8a","Type":"ContainerStarted","Data":"2d04fcdab766b2b930785d656361aae896447910660dbeb62ee7e4215298cf07"} Mar 07 15:12:28 crc kubenswrapper[4943]: I0307 15:12:28.651541 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-4lx9r" event={"ID":"8b6b168d-d695-4916-a11c-810b697d1e8a","Type":"ContainerStarted","Data":"281c33e62c7ecaf9baf8e691262f810b9f5af2d0e8cdbad8d3fbcf1dcdf6f955"} Mar 07 15:12:28 crc kubenswrapper[4943]: I0307 15:12:28.672454 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-4lx9r" podStartSLOduration=1.672430495 podStartE2EDuration="1.672430495s" podCreationTimestamp="2026-03-07 15:12:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:12:28.66858686 +0000 UTC m=+1990.620723378" watchObservedRunningTime="2026-03-07 15:12:28.672430495 +0000 UTC m=+1990.624567013" Mar 07 15:12:29 crc kubenswrapper[4943]: I0307 15:12:29.664064 4943 generic.go:334] "Generic (PLEG): container finished" podID="8b6b168d-d695-4916-a11c-810b697d1e8a" containerID="2d04fcdab766b2b930785d656361aae896447910660dbeb62ee7e4215298cf07" exitCode=0 Mar 07 15:12:29 crc kubenswrapper[4943]: I0307 15:12:29.665422 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-4lx9r" event={"ID":"8b6b168d-d695-4916-a11c-810b697d1e8a","Type":"ContainerDied","Data":"2d04fcdab766b2b930785d656361aae896447910660dbeb62ee7e4215298cf07"} Mar 07 15:12:31 crc kubenswrapper[4943]: I0307 15:12:31.002392 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-4lx9r" Mar 07 15:12:31 crc kubenswrapper[4943]: I0307 15:12:31.032123 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-4lx9r"] Mar 07 15:12:31 crc kubenswrapper[4943]: I0307 15:12:31.046435 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-4lx9r"] Mar 07 15:12:31 crc kubenswrapper[4943]: I0307 15:12:31.149665 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tqd4z\" (UniqueName: \"kubernetes.io/projected/8b6b168d-d695-4916-a11c-810b697d1e8a-kube-api-access-tqd4z\") pod \"8b6b168d-d695-4916-a11c-810b697d1e8a\" (UID: \"8b6b168d-d695-4916-a11c-810b697d1e8a\") " Mar 07 15:12:31 crc kubenswrapper[4943]: I0307 15:12:31.149746 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/8b6b168d-d695-4916-a11c-810b697d1e8a-ring-data-devices\") pod \"8b6b168d-d695-4916-a11c-810b697d1e8a\" (UID: \"8b6b168d-d695-4916-a11c-810b697d1e8a\") " Mar 07 15:12:31 crc kubenswrapper[4943]: I0307 15:12:31.149790 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/8b6b168d-d695-4916-a11c-810b697d1e8a-dispersionconf\") pod \"8b6b168d-d695-4916-a11c-810b697d1e8a\" (UID: \"8b6b168d-d695-4916-a11c-810b697d1e8a\") " Mar 07 15:12:31 crc kubenswrapper[4943]: I0307 15:12:31.149816 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/8b6b168d-d695-4916-a11c-810b697d1e8a-swiftconf\") pod \"8b6b168d-d695-4916-a11c-810b697d1e8a\" (UID: \"8b6b168d-d695-4916-a11c-810b697d1e8a\") " Mar 07 15:12:31 crc kubenswrapper[4943]: I0307 15:12:31.149861 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8b6b168d-d695-4916-a11c-810b697d1e8a-scripts\") pod \"8b6b168d-d695-4916-a11c-810b697d1e8a\" (UID: \"8b6b168d-d695-4916-a11c-810b697d1e8a\") " Mar 07 15:12:31 crc kubenswrapper[4943]: I0307 15:12:31.149914 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/8b6b168d-d695-4916-a11c-810b697d1e8a-etc-swift\") pod \"8b6b168d-d695-4916-a11c-810b697d1e8a\" (UID: \"8b6b168d-d695-4916-a11c-810b697d1e8a\") " Mar 07 15:12:31 crc kubenswrapper[4943]: I0307 15:12:31.150879 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b6b168d-d695-4916-a11c-810b697d1e8a-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "8b6b168d-d695-4916-a11c-810b697d1e8a" (UID: "8b6b168d-d695-4916-a11c-810b697d1e8a"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:12:31 crc kubenswrapper[4943]: I0307 15:12:31.151401 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b6b168d-d695-4916-a11c-810b697d1e8a-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "8b6b168d-d695-4916-a11c-810b697d1e8a" (UID: "8b6b168d-d695-4916-a11c-810b697d1e8a"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:12:31 crc kubenswrapper[4943]: I0307 15:12:31.156148 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b6b168d-d695-4916-a11c-810b697d1e8a-kube-api-access-tqd4z" (OuterVolumeSpecName: "kube-api-access-tqd4z") pod "8b6b168d-d695-4916-a11c-810b697d1e8a" (UID: "8b6b168d-d695-4916-a11c-810b697d1e8a"). InnerVolumeSpecName "kube-api-access-tqd4z". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:12:31 crc kubenswrapper[4943]: I0307 15:12:31.169346 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b6b168d-d695-4916-a11c-810b697d1e8a-scripts" (OuterVolumeSpecName: "scripts") pod "8b6b168d-d695-4916-a11c-810b697d1e8a" (UID: "8b6b168d-d695-4916-a11c-810b697d1e8a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:12:31 crc kubenswrapper[4943]: I0307 15:12:31.186584 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b6b168d-d695-4916-a11c-810b697d1e8a-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "8b6b168d-d695-4916-a11c-810b697d1e8a" (UID: "8b6b168d-d695-4916-a11c-810b697d1e8a"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:12:31 crc kubenswrapper[4943]: I0307 15:12:31.188532 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b6b168d-d695-4916-a11c-810b697d1e8a-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "8b6b168d-d695-4916-a11c-810b697d1e8a" (UID: "8b6b168d-d695-4916-a11c-810b697d1e8a"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:12:31 crc kubenswrapper[4943]: I0307 15:12:31.251937 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/8b6b168d-d695-4916-a11c-810b697d1e8a-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:31 crc kubenswrapper[4943]: I0307 15:12:31.251967 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/8b6b168d-d695-4916-a11c-810b697d1e8a-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:31 crc kubenswrapper[4943]: I0307 15:12:31.251976 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/8b6b168d-d695-4916-a11c-810b697d1e8a-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:31 crc kubenswrapper[4943]: I0307 15:12:31.251984 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8b6b168d-d695-4916-a11c-810b697d1e8a-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:31 crc kubenswrapper[4943]: I0307 15:12:31.251993 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/8b6b168d-d695-4916-a11c-810b697d1e8a-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:31 crc kubenswrapper[4943]: I0307 15:12:31.252001 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tqd4z\" (UniqueName: \"kubernetes.io/projected/8b6b168d-d695-4916-a11c-810b697d1e8a-kube-api-access-tqd4z\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:31 crc kubenswrapper[4943]: I0307 15:12:31.687619 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="281c33e62c7ecaf9baf8e691262f810b9f5af2d0e8cdbad8d3fbcf1dcdf6f955" Mar 07 15:12:31 crc kubenswrapper[4943]: I0307 15:12:31.687678 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-4lx9r" Mar 07 15:12:32 crc kubenswrapper[4943]: I0307 15:12:32.243319 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7fs6t"] Mar 07 15:12:32 crc kubenswrapper[4943]: E0307 15:12:32.243591 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b6b168d-d695-4916-a11c-810b697d1e8a" containerName="swift-ring-rebalance" Mar 07 15:12:32 crc kubenswrapper[4943]: I0307 15:12:32.243603 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b6b168d-d695-4916-a11c-810b697d1e8a" containerName="swift-ring-rebalance" Mar 07 15:12:32 crc kubenswrapper[4943]: I0307 15:12:32.243750 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b6b168d-d695-4916-a11c-810b697d1e8a" containerName="swift-ring-rebalance" Mar 07 15:12:32 crc kubenswrapper[4943]: I0307 15:12:32.244189 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7fs6t" Mar 07 15:12:32 crc kubenswrapper[4943]: I0307 15:12:32.246281 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:12:32 crc kubenswrapper[4943]: I0307 15:12:32.246787 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:12:32 crc kubenswrapper[4943]: I0307 15:12:32.263505 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7fs6t"] Mar 07 15:12:32 crc kubenswrapper[4943]: I0307 15:12:32.370840 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5ddd49e9-e705-4da2-937c-ead66c880624-etc-swift\") pod \"swift-ring-rebalance-debug-7fs6t\" (UID: \"5ddd49e9-e705-4da2-937c-ead66c880624\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7fs6t" Mar 07 15:12:32 crc kubenswrapper[4943]: I0307 15:12:32.370907 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5ddd49e9-e705-4da2-937c-ead66c880624-dispersionconf\") pod \"swift-ring-rebalance-debug-7fs6t\" (UID: \"5ddd49e9-e705-4da2-937c-ead66c880624\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7fs6t" Mar 07 15:12:32 crc kubenswrapper[4943]: I0307 15:12:32.370972 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5ddd49e9-e705-4da2-937c-ead66c880624-swiftconf\") pod \"swift-ring-rebalance-debug-7fs6t\" (UID: \"5ddd49e9-e705-4da2-937c-ead66c880624\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7fs6t" Mar 07 15:12:32 crc kubenswrapper[4943]: I0307 15:12:32.370994 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5ddd49e9-e705-4da2-937c-ead66c880624-ring-data-devices\") pod \"swift-ring-rebalance-debug-7fs6t\" (UID: \"5ddd49e9-e705-4da2-937c-ead66c880624\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7fs6t" Mar 07 15:12:32 crc kubenswrapper[4943]: I0307 15:12:32.371059 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46f52\" (UniqueName: \"kubernetes.io/projected/5ddd49e9-e705-4da2-937c-ead66c880624-kube-api-access-46f52\") pod \"swift-ring-rebalance-debug-7fs6t\" (UID: \"5ddd49e9-e705-4da2-937c-ead66c880624\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7fs6t" Mar 07 15:12:32 crc kubenswrapper[4943]: I0307 15:12:32.371086 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5ddd49e9-e705-4da2-937c-ead66c880624-scripts\") pod \"swift-ring-rebalance-debug-7fs6t\" (UID: \"5ddd49e9-e705-4da2-937c-ead66c880624\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7fs6t" Mar 07 15:12:32 crc kubenswrapper[4943]: I0307 15:12:32.472479 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46f52\" (UniqueName: \"kubernetes.io/projected/5ddd49e9-e705-4da2-937c-ead66c880624-kube-api-access-46f52\") pod \"swift-ring-rebalance-debug-7fs6t\" (UID: \"5ddd49e9-e705-4da2-937c-ead66c880624\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7fs6t" Mar 07 15:12:32 crc kubenswrapper[4943]: I0307 15:12:32.472559 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5ddd49e9-e705-4da2-937c-ead66c880624-scripts\") pod \"swift-ring-rebalance-debug-7fs6t\" (UID: \"5ddd49e9-e705-4da2-937c-ead66c880624\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7fs6t" Mar 07 15:12:32 crc kubenswrapper[4943]: I0307 15:12:32.472652 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5ddd49e9-e705-4da2-937c-ead66c880624-etc-swift\") pod \"swift-ring-rebalance-debug-7fs6t\" (UID: \"5ddd49e9-e705-4da2-937c-ead66c880624\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7fs6t" Mar 07 15:12:32 crc kubenswrapper[4943]: I0307 15:12:32.472738 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5ddd49e9-e705-4da2-937c-ead66c880624-dispersionconf\") pod \"swift-ring-rebalance-debug-7fs6t\" (UID: \"5ddd49e9-e705-4da2-937c-ead66c880624\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7fs6t" Mar 07 15:12:32 crc kubenswrapper[4943]: I0307 15:12:32.472822 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5ddd49e9-e705-4da2-937c-ead66c880624-swiftconf\") pod \"swift-ring-rebalance-debug-7fs6t\" (UID: \"5ddd49e9-e705-4da2-937c-ead66c880624\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7fs6t" Mar 07 15:12:32 crc kubenswrapper[4943]: I0307 15:12:32.472858 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5ddd49e9-e705-4da2-937c-ead66c880624-ring-data-devices\") pod \"swift-ring-rebalance-debug-7fs6t\" (UID: \"5ddd49e9-e705-4da2-937c-ead66c880624\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7fs6t" Mar 07 15:12:32 crc kubenswrapper[4943]: I0307 15:12:32.473999 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5ddd49e9-e705-4da2-937c-ead66c880624-etc-swift\") pod \"swift-ring-rebalance-debug-7fs6t\" (UID: \"5ddd49e9-e705-4da2-937c-ead66c880624\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7fs6t" Mar 07 15:12:32 crc kubenswrapper[4943]: I0307 15:12:32.474072 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5ddd49e9-e705-4da2-937c-ead66c880624-scripts\") pod \"swift-ring-rebalance-debug-7fs6t\" (UID: \"5ddd49e9-e705-4da2-937c-ead66c880624\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7fs6t" Mar 07 15:12:32 crc kubenswrapper[4943]: I0307 15:12:32.474170 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5ddd49e9-e705-4da2-937c-ead66c880624-ring-data-devices\") pod \"swift-ring-rebalance-debug-7fs6t\" (UID: \"5ddd49e9-e705-4da2-937c-ead66c880624\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7fs6t" Mar 07 15:12:32 crc kubenswrapper[4943]: I0307 15:12:32.480303 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5ddd49e9-e705-4da2-937c-ead66c880624-dispersionconf\") pod \"swift-ring-rebalance-debug-7fs6t\" (UID: \"5ddd49e9-e705-4da2-937c-ead66c880624\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7fs6t" Mar 07 15:12:32 crc kubenswrapper[4943]: I0307 15:12:32.481678 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5ddd49e9-e705-4da2-937c-ead66c880624-swiftconf\") pod \"swift-ring-rebalance-debug-7fs6t\" (UID: \"5ddd49e9-e705-4da2-937c-ead66c880624\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7fs6t" Mar 07 15:12:32 crc kubenswrapper[4943]: I0307 15:12:32.500142 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46f52\" (UniqueName: \"kubernetes.io/projected/5ddd49e9-e705-4da2-937c-ead66c880624-kube-api-access-46f52\") pod \"swift-ring-rebalance-debug-7fs6t\" (UID: \"5ddd49e9-e705-4da2-937c-ead66c880624\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7fs6t" Mar 07 15:12:32 crc kubenswrapper[4943]: I0307 15:12:32.561788 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7fs6t" Mar 07 15:12:32 crc kubenswrapper[4943]: I0307 15:12:32.765596 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b6b168d-d695-4916-a11c-810b697d1e8a" path="/var/lib/kubelet/pods/8b6b168d-d695-4916-a11c-810b697d1e8a/volumes" Mar 07 15:12:32 crc kubenswrapper[4943]: I0307 15:12:32.783955 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7fs6t"] Mar 07 15:12:33 crc kubenswrapper[4943]: I0307 15:12:33.711020 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7fs6t" event={"ID":"5ddd49e9-e705-4da2-937c-ead66c880624","Type":"ContainerStarted","Data":"724090a96661631af14d84e2e4a041922bff6bdf27022c51d53e1d415be01555"} Mar 07 15:12:33 crc kubenswrapper[4943]: I0307 15:12:33.711269 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7fs6t" event={"ID":"5ddd49e9-e705-4da2-937c-ead66c880624","Type":"ContainerStarted","Data":"fe0932da57ed162dcf4aff771e474d2deaba2655c99a7e639726bb42c6990ec0"} Mar 07 15:12:34 crc kubenswrapper[4943]: I0307 15:12:34.724867 4943 generic.go:334] "Generic (PLEG): container finished" podID="5ddd49e9-e705-4da2-937c-ead66c880624" containerID="724090a96661631af14d84e2e4a041922bff6bdf27022c51d53e1d415be01555" exitCode=0 Mar 07 15:12:34 crc kubenswrapper[4943]: I0307 15:12:34.724909 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7fs6t" event={"ID":"5ddd49e9-e705-4da2-937c-ead66c880624","Type":"ContainerDied","Data":"724090a96661631af14d84e2e4a041922bff6bdf27022c51d53e1d415be01555"} Mar 07 15:12:36 crc kubenswrapper[4943]: I0307 15:12:36.074301 4943 patch_prober.go:28] interesting pod/machine-config-daemon-cpdmm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 07 15:12:36 crc kubenswrapper[4943]: I0307 15:12:36.074573 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 07 15:12:36 crc kubenswrapper[4943]: I0307 15:12:36.118740 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7fs6t" Mar 07 15:12:36 crc kubenswrapper[4943]: I0307 15:12:36.157730 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7fs6t"] Mar 07 15:12:36 crc kubenswrapper[4943]: I0307 15:12:36.166698 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7fs6t"] Mar 07 15:12:36 crc kubenswrapper[4943]: I0307 15:12:36.233584 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5ddd49e9-e705-4da2-937c-ead66c880624-dispersionconf\") pod \"5ddd49e9-e705-4da2-937c-ead66c880624\" (UID: \"5ddd49e9-e705-4da2-937c-ead66c880624\") " Mar 07 15:12:36 crc kubenswrapper[4943]: I0307 15:12:36.233711 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5ddd49e9-e705-4da2-937c-ead66c880624-etc-swift\") pod \"5ddd49e9-e705-4da2-937c-ead66c880624\" (UID: \"5ddd49e9-e705-4da2-937c-ead66c880624\") " Mar 07 15:12:36 crc kubenswrapper[4943]: I0307 15:12:36.233736 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5ddd49e9-e705-4da2-937c-ead66c880624-ring-data-devices\") pod \"5ddd49e9-e705-4da2-937c-ead66c880624\" (UID: \"5ddd49e9-e705-4da2-937c-ead66c880624\") " Mar 07 15:12:36 crc kubenswrapper[4943]: I0307 15:12:36.233758 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5ddd49e9-e705-4da2-937c-ead66c880624-swiftconf\") pod \"5ddd49e9-e705-4da2-937c-ead66c880624\" (UID: \"5ddd49e9-e705-4da2-937c-ead66c880624\") " Mar 07 15:12:36 crc kubenswrapper[4943]: I0307 15:12:36.233794 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-46f52\" (UniqueName: \"kubernetes.io/projected/5ddd49e9-e705-4da2-937c-ead66c880624-kube-api-access-46f52\") pod \"5ddd49e9-e705-4da2-937c-ead66c880624\" (UID: \"5ddd49e9-e705-4da2-937c-ead66c880624\") " Mar 07 15:12:36 crc kubenswrapper[4943]: I0307 15:12:36.233866 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5ddd49e9-e705-4da2-937c-ead66c880624-scripts\") pod \"5ddd49e9-e705-4da2-937c-ead66c880624\" (UID: \"5ddd49e9-e705-4da2-937c-ead66c880624\") " Mar 07 15:12:36 crc kubenswrapper[4943]: I0307 15:12:36.234902 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ddd49e9-e705-4da2-937c-ead66c880624-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "5ddd49e9-e705-4da2-937c-ead66c880624" (UID: "5ddd49e9-e705-4da2-937c-ead66c880624"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:12:36 crc kubenswrapper[4943]: I0307 15:12:36.235135 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ddd49e9-e705-4da2-937c-ead66c880624-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "5ddd49e9-e705-4da2-937c-ead66c880624" (UID: "5ddd49e9-e705-4da2-937c-ead66c880624"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:12:36 crc kubenswrapper[4943]: I0307 15:12:36.244104 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ddd49e9-e705-4da2-937c-ead66c880624-kube-api-access-46f52" (OuterVolumeSpecName: "kube-api-access-46f52") pod "5ddd49e9-e705-4da2-937c-ead66c880624" (UID: "5ddd49e9-e705-4da2-937c-ead66c880624"). InnerVolumeSpecName "kube-api-access-46f52". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:12:36 crc kubenswrapper[4943]: I0307 15:12:36.251487 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ddd49e9-e705-4da2-937c-ead66c880624-scripts" (OuterVolumeSpecName: "scripts") pod "5ddd49e9-e705-4da2-937c-ead66c880624" (UID: "5ddd49e9-e705-4da2-937c-ead66c880624"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:12:36 crc kubenswrapper[4943]: I0307 15:12:36.253333 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ddd49e9-e705-4da2-937c-ead66c880624-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "5ddd49e9-e705-4da2-937c-ead66c880624" (UID: "5ddd49e9-e705-4da2-937c-ead66c880624"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:12:36 crc kubenswrapper[4943]: I0307 15:12:36.257358 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ddd49e9-e705-4da2-937c-ead66c880624-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "5ddd49e9-e705-4da2-937c-ead66c880624" (UID: "5ddd49e9-e705-4da2-937c-ead66c880624"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:12:36 crc kubenswrapper[4943]: I0307 15:12:36.335880 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5ddd49e9-e705-4da2-937c-ead66c880624-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:36 crc kubenswrapper[4943]: I0307 15:12:36.335918 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5ddd49e9-e705-4da2-937c-ead66c880624-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:36 crc kubenswrapper[4943]: I0307 15:12:36.335980 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5ddd49e9-e705-4da2-937c-ead66c880624-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:36 crc kubenswrapper[4943]: I0307 15:12:36.335991 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5ddd49e9-e705-4da2-937c-ead66c880624-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:36 crc kubenswrapper[4943]: I0307 15:12:36.336003 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5ddd49e9-e705-4da2-937c-ead66c880624-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:36 crc kubenswrapper[4943]: I0307 15:12:36.336013 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-46f52\" (UniqueName: \"kubernetes.io/projected/5ddd49e9-e705-4da2-937c-ead66c880624-kube-api-access-46f52\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:36 crc kubenswrapper[4943]: I0307 15:12:36.747456 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe0932da57ed162dcf4aff771e474d2deaba2655c99a7e639726bb42c6990ec0" Mar 07 15:12:36 crc kubenswrapper[4943]: I0307 15:12:36.747534 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7fs6t" Mar 07 15:12:36 crc kubenswrapper[4943]: I0307 15:12:36.771303 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ddd49e9-e705-4da2-937c-ead66c880624" path="/var/lib/kubelet/pods/5ddd49e9-e705-4da2-937c-ead66c880624/volumes" Mar 07 15:12:37 crc kubenswrapper[4943]: I0307 15:12:37.380384 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-vqnpf"] Mar 07 15:12:37 crc kubenswrapper[4943]: E0307 15:12:37.380865 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ddd49e9-e705-4da2-937c-ead66c880624" containerName="swift-ring-rebalance" Mar 07 15:12:37 crc kubenswrapper[4943]: I0307 15:12:37.380888 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ddd49e9-e705-4da2-937c-ead66c880624" containerName="swift-ring-rebalance" Mar 07 15:12:37 crc kubenswrapper[4943]: I0307 15:12:37.381292 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ddd49e9-e705-4da2-937c-ead66c880624" containerName="swift-ring-rebalance" Mar 07 15:12:37 crc kubenswrapper[4943]: I0307 15:12:37.382203 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-vqnpf" Mar 07 15:12:37 crc kubenswrapper[4943]: I0307 15:12:37.386491 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:12:37 crc kubenswrapper[4943]: I0307 15:12:37.386540 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:12:37 crc kubenswrapper[4943]: I0307 15:12:37.393127 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-vqnpf"] Mar 07 15:12:37 crc kubenswrapper[4943]: I0307 15:12:37.453820 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c7f7231f-c606-43b7-b4a7-5990f7544198-swiftconf\") pod \"swift-ring-rebalance-debug-vqnpf\" (UID: \"c7f7231f-c606-43b7-b4a7-5990f7544198\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vqnpf" Mar 07 15:12:37 crc kubenswrapper[4943]: I0307 15:12:37.453961 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c7f7231f-c606-43b7-b4a7-5990f7544198-scripts\") pod \"swift-ring-rebalance-debug-vqnpf\" (UID: \"c7f7231f-c606-43b7-b4a7-5990f7544198\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vqnpf" Mar 07 15:12:37 crc kubenswrapper[4943]: I0307 15:12:37.453997 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c7f7231f-c606-43b7-b4a7-5990f7544198-ring-data-devices\") pod \"swift-ring-rebalance-debug-vqnpf\" (UID: \"c7f7231f-c606-43b7-b4a7-5990f7544198\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vqnpf" Mar 07 15:12:37 crc kubenswrapper[4943]: I0307 15:12:37.454024 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qggs9\" (UniqueName: \"kubernetes.io/projected/c7f7231f-c606-43b7-b4a7-5990f7544198-kube-api-access-qggs9\") pod \"swift-ring-rebalance-debug-vqnpf\" (UID: \"c7f7231f-c606-43b7-b4a7-5990f7544198\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vqnpf" Mar 07 15:12:37 crc kubenswrapper[4943]: I0307 15:12:37.454066 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c7f7231f-c606-43b7-b4a7-5990f7544198-dispersionconf\") pod \"swift-ring-rebalance-debug-vqnpf\" (UID: \"c7f7231f-c606-43b7-b4a7-5990f7544198\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vqnpf" Mar 07 15:12:37 crc kubenswrapper[4943]: I0307 15:12:37.454115 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c7f7231f-c606-43b7-b4a7-5990f7544198-etc-swift\") pod \"swift-ring-rebalance-debug-vqnpf\" (UID: \"c7f7231f-c606-43b7-b4a7-5990f7544198\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vqnpf" Mar 07 15:12:37 crc kubenswrapper[4943]: I0307 15:12:37.555107 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c7f7231f-c606-43b7-b4a7-5990f7544198-scripts\") pod \"swift-ring-rebalance-debug-vqnpf\" (UID: \"c7f7231f-c606-43b7-b4a7-5990f7544198\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vqnpf" Mar 07 15:12:37 crc kubenswrapper[4943]: I0307 15:12:37.555169 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c7f7231f-c606-43b7-b4a7-5990f7544198-ring-data-devices\") pod \"swift-ring-rebalance-debug-vqnpf\" (UID: \"c7f7231f-c606-43b7-b4a7-5990f7544198\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vqnpf" Mar 07 15:12:37 crc kubenswrapper[4943]: I0307 15:12:37.555198 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qggs9\" (UniqueName: \"kubernetes.io/projected/c7f7231f-c606-43b7-b4a7-5990f7544198-kube-api-access-qggs9\") pod \"swift-ring-rebalance-debug-vqnpf\" (UID: \"c7f7231f-c606-43b7-b4a7-5990f7544198\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vqnpf" Mar 07 15:12:37 crc kubenswrapper[4943]: I0307 15:12:37.555259 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c7f7231f-c606-43b7-b4a7-5990f7544198-dispersionconf\") pod \"swift-ring-rebalance-debug-vqnpf\" (UID: \"c7f7231f-c606-43b7-b4a7-5990f7544198\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vqnpf" Mar 07 15:12:37 crc kubenswrapper[4943]: I0307 15:12:37.555308 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c7f7231f-c606-43b7-b4a7-5990f7544198-etc-swift\") pod \"swift-ring-rebalance-debug-vqnpf\" (UID: \"c7f7231f-c606-43b7-b4a7-5990f7544198\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vqnpf" Mar 07 15:12:37 crc kubenswrapper[4943]: I0307 15:12:37.555334 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c7f7231f-c606-43b7-b4a7-5990f7544198-swiftconf\") pod \"swift-ring-rebalance-debug-vqnpf\" (UID: \"c7f7231f-c606-43b7-b4a7-5990f7544198\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vqnpf" Mar 07 15:12:37 crc kubenswrapper[4943]: I0307 15:12:37.556089 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c7f7231f-c606-43b7-b4a7-5990f7544198-etc-swift\") pod \"swift-ring-rebalance-debug-vqnpf\" (UID: \"c7f7231f-c606-43b7-b4a7-5990f7544198\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vqnpf" Mar 07 15:12:37 crc kubenswrapper[4943]: I0307 15:12:37.556533 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c7f7231f-c606-43b7-b4a7-5990f7544198-scripts\") pod \"swift-ring-rebalance-debug-vqnpf\" (UID: \"c7f7231f-c606-43b7-b4a7-5990f7544198\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vqnpf" Mar 07 15:12:37 crc kubenswrapper[4943]: I0307 15:12:37.556571 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c7f7231f-c606-43b7-b4a7-5990f7544198-ring-data-devices\") pod \"swift-ring-rebalance-debug-vqnpf\" (UID: \"c7f7231f-c606-43b7-b4a7-5990f7544198\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vqnpf" Mar 07 15:12:37 crc kubenswrapper[4943]: I0307 15:12:37.561651 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c7f7231f-c606-43b7-b4a7-5990f7544198-dispersionconf\") pod \"swift-ring-rebalance-debug-vqnpf\" (UID: \"c7f7231f-c606-43b7-b4a7-5990f7544198\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vqnpf" Mar 07 15:12:37 crc kubenswrapper[4943]: I0307 15:12:37.564266 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c7f7231f-c606-43b7-b4a7-5990f7544198-swiftconf\") pod \"swift-ring-rebalance-debug-vqnpf\" (UID: \"c7f7231f-c606-43b7-b4a7-5990f7544198\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vqnpf" Mar 07 15:12:37 crc kubenswrapper[4943]: I0307 15:12:37.589144 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qggs9\" (UniqueName: \"kubernetes.io/projected/c7f7231f-c606-43b7-b4a7-5990f7544198-kube-api-access-qggs9\") pod \"swift-ring-rebalance-debug-vqnpf\" (UID: \"c7f7231f-c606-43b7-b4a7-5990f7544198\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vqnpf" Mar 07 15:12:37 crc kubenswrapper[4943]: I0307 15:12:37.701853 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-vqnpf" Mar 07 15:12:38 crc kubenswrapper[4943]: I0307 15:12:38.046604 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-vqnpf"] Mar 07 15:12:38 crc kubenswrapper[4943]: I0307 15:12:38.770141 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-vqnpf" event={"ID":"c7f7231f-c606-43b7-b4a7-5990f7544198","Type":"ContainerStarted","Data":"20ffa016619f0915dfcc66196b10b19a28e3969cb9c044c1763d00640683bcba"} Mar 07 15:12:38 crc kubenswrapper[4943]: I0307 15:12:38.770507 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-vqnpf" event={"ID":"c7f7231f-c606-43b7-b4a7-5990f7544198","Type":"ContainerStarted","Data":"533e0e2cdb64bde7efa8fe223b46f3f8d1a1e5063a495ea386d75373a4b7c7e4"} Mar 07 15:12:38 crc kubenswrapper[4943]: I0307 15:12:38.821297 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-vqnpf" podStartSLOduration=1.8212809189999999 podStartE2EDuration="1.821280919s" podCreationTimestamp="2026-03-07 15:12:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:12:38.818858979 +0000 UTC m=+2000.770995477" watchObservedRunningTime="2026-03-07 15:12:38.821280919 +0000 UTC m=+2000.773417417" Mar 07 15:12:39 crc kubenswrapper[4943]: I0307 15:12:39.782611 4943 generic.go:334] "Generic (PLEG): container finished" podID="c7f7231f-c606-43b7-b4a7-5990f7544198" containerID="20ffa016619f0915dfcc66196b10b19a28e3969cb9c044c1763d00640683bcba" exitCode=0 Mar 07 15:12:39 crc kubenswrapper[4943]: I0307 15:12:39.782679 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-vqnpf" event={"ID":"c7f7231f-c606-43b7-b4a7-5990f7544198","Type":"ContainerDied","Data":"20ffa016619f0915dfcc66196b10b19a28e3969cb9c044c1763d00640683bcba"} Mar 07 15:12:41 crc kubenswrapper[4943]: I0307 15:12:41.120110 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-vqnpf" Mar 07 15:12:41 crc kubenswrapper[4943]: I0307 15:12:41.150107 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-vqnpf"] Mar 07 15:12:41 crc kubenswrapper[4943]: I0307 15:12:41.156775 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-vqnpf"] Mar 07 15:12:41 crc kubenswrapper[4943]: I0307 15:12:41.264662 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c7f7231f-c606-43b7-b4a7-5990f7544198-scripts\") pod \"c7f7231f-c606-43b7-b4a7-5990f7544198\" (UID: \"c7f7231f-c606-43b7-b4a7-5990f7544198\") " Mar 07 15:12:41 crc kubenswrapper[4943]: I0307 15:12:41.264720 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c7f7231f-c606-43b7-b4a7-5990f7544198-etc-swift\") pod \"c7f7231f-c606-43b7-b4a7-5990f7544198\" (UID: \"c7f7231f-c606-43b7-b4a7-5990f7544198\") " Mar 07 15:12:41 crc kubenswrapper[4943]: I0307 15:12:41.264807 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c7f7231f-c606-43b7-b4a7-5990f7544198-ring-data-devices\") pod \"c7f7231f-c606-43b7-b4a7-5990f7544198\" (UID: \"c7f7231f-c606-43b7-b4a7-5990f7544198\") " Mar 07 15:12:41 crc kubenswrapper[4943]: I0307 15:12:41.264875 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c7f7231f-c606-43b7-b4a7-5990f7544198-dispersionconf\") pod \"c7f7231f-c606-43b7-b4a7-5990f7544198\" (UID: \"c7f7231f-c606-43b7-b4a7-5990f7544198\") " Mar 07 15:12:41 crc kubenswrapper[4943]: I0307 15:12:41.264904 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c7f7231f-c606-43b7-b4a7-5990f7544198-swiftconf\") pod \"c7f7231f-c606-43b7-b4a7-5990f7544198\" (UID: \"c7f7231f-c606-43b7-b4a7-5990f7544198\") " Mar 07 15:12:41 crc kubenswrapper[4943]: I0307 15:12:41.265013 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qggs9\" (UniqueName: \"kubernetes.io/projected/c7f7231f-c606-43b7-b4a7-5990f7544198-kube-api-access-qggs9\") pod \"c7f7231f-c606-43b7-b4a7-5990f7544198\" (UID: \"c7f7231f-c606-43b7-b4a7-5990f7544198\") " Mar 07 15:12:41 crc kubenswrapper[4943]: I0307 15:12:41.266593 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c7f7231f-c606-43b7-b4a7-5990f7544198-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "c7f7231f-c606-43b7-b4a7-5990f7544198" (UID: "c7f7231f-c606-43b7-b4a7-5990f7544198"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:12:41 crc kubenswrapper[4943]: I0307 15:12:41.266639 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7f7231f-c606-43b7-b4a7-5990f7544198-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "c7f7231f-c606-43b7-b4a7-5990f7544198" (UID: "c7f7231f-c606-43b7-b4a7-5990f7544198"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:12:41 crc kubenswrapper[4943]: I0307 15:12:41.269794 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7f7231f-c606-43b7-b4a7-5990f7544198-kube-api-access-qggs9" (OuterVolumeSpecName: "kube-api-access-qggs9") pod "c7f7231f-c606-43b7-b4a7-5990f7544198" (UID: "c7f7231f-c606-43b7-b4a7-5990f7544198"). InnerVolumeSpecName "kube-api-access-qggs9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:12:41 crc kubenswrapper[4943]: I0307 15:12:41.282327 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c7f7231f-c606-43b7-b4a7-5990f7544198-scripts" (OuterVolumeSpecName: "scripts") pod "c7f7231f-c606-43b7-b4a7-5990f7544198" (UID: "c7f7231f-c606-43b7-b4a7-5990f7544198"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:12:41 crc kubenswrapper[4943]: I0307 15:12:41.307898 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7f7231f-c606-43b7-b4a7-5990f7544198-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "c7f7231f-c606-43b7-b4a7-5990f7544198" (UID: "c7f7231f-c606-43b7-b4a7-5990f7544198"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:12:41 crc kubenswrapper[4943]: I0307 15:12:41.309978 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7f7231f-c606-43b7-b4a7-5990f7544198-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "c7f7231f-c606-43b7-b4a7-5990f7544198" (UID: "c7f7231f-c606-43b7-b4a7-5990f7544198"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:12:41 crc kubenswrapper[4943]: I0307 15:12:41.367389 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c7f7231f-c606-43b7-b4a7-5990f7544198-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:41 crc kubenswrapper[4943]: I0307 15:12:41.367428 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c7f7231f-c606-43b7-b4a7-5990f7544198-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:41 crc kubenswrapper[4943]: I0307 15:12:41.367442 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c7f7231f-c606-43b7-b4a7-5990f7544198-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:41 crc kubenswrapper[4943]: I0307 15:12:41.367453 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c7f7231f-c606-43b7-b4a7-5990f7544198-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:41 crc kubenswrapper[4943]: I0307 15:12:41.367465 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qggs9\" (UniqueName: \"kubernetes.io/projected/c7f7231f-c606-43b7-b4a7-5990f7544198-kube-api-access-qggs9\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:41 crc kubenswrapper[4943]: I0307 15:12:41.367476 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c7f7231f-c606-43b7-b4a7-5990f7544198-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:41 crc kubenswrapper[4943]: I0307 15:12:41.801799 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="533e0e2cdb64bde7efa8fe223b46f3f8d1a1e5063a495ea386d75373a4b7c7e4" Mar 07 15:12:41 crc kubenswrapper[4943]: I0307 15:12:41.801890 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-vqnpf" Mar 07 15:12:42 crc kubenswrapper[4943]: I0307 15:12:42.343683 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-fhd6z"] Mar 07 15:12:42 crc kubenswrapper[4943]: E0307 15:12:42.344300 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7f7231f-c606-43b7-b4a7-5990f7544198" containerName="swift-ring-rebalance" Mar 07 15:12:42 crc kubenswrapper[4943]: I0307 15:12:42.344335 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7f7231f-c606-43b7-b4a7-5990f7544198" containerName="swift-ring-rebalance" Mar 07 15:12:42 crc kubenswrapper[4943]: I0307 15:12:42.344770 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7f7231f-c606-43b7-b4a7-5990f7544198" containerName="swift-ring-rebalance" Mar 07 15:12:42 crc kubenswrapper[4943]: I0307 15:12:42.346013 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-fhd6z" Mar 07 15:12:42 crc kubenswrapper[4943]: I0307 15:12:42.351229 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:12:42 crc kubenswrapper[4943]: I0307 15:12:42.352959 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:12:42 crc kubenswrapper[4943]: I0307 15:12:42.363365 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-fhd6z"] Mar 07 15:12:42 crc kubenswrapper[4943]: I0307 15:12:42.485605 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/de15165d-f055-468e-8f76-23f8bfe7d3eb-ring-data-devices\") pod \"swift-ring-rebalance-debug-fhd6z\" (UID: \"de15165d-f055-468e-8f76-23f8bfe7d3eb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fhd6z" Mar 07 15:12:42 crc kubenswrapper[4943]: I0307 15:12:42.485744 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/de15165d-f055-468e-8f76-23f8bfe7d3eb-dispersionconf\") pod \"swift-ring-rebalance-debug-fhd6z\" (UID: \"de15165d-f055-468e-8f76-23f8bfe7d3eb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fhd6z" Mar 07 15:12:42 crc kubenswrapper[4943]: I0307 15:12:42.485802 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/de15165d-f055-468e-8f76-23f8bfe7d3eb-swiftconf\") pod \"swift-ring-rebalance-debug-fhd6z\" (UID: \"de15165d-f055-468e-8f76-23f8bfe7d3eb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fhd6z" Mar 07 15:12:42 crc kubenswrapper[4943]: I0307 15:12:42.486087 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/de15165d-f055-468e-8f76-23f8bfe7d3eb-scripts\") pod \"swift-ring-rebalance-debug-fhd6z\" (UID: \"de15165d-f055-468e-8f76-23f8bfe7d3eb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fhd6z" Mar 07 15:12:42 crc kubenswrapper[4943]: I0307 15:12:42.486161 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmwfd\" (UniqueName: \"kubernetes.io/projected/de15165d-f055-468e-8f76-23f8bfe7d3eb-kube-api-access-lmwfd\") pod \"swift-ring-rebalance-debug-fhd6z\" (UID: \"de15165d-f055-468e-8f76-23f8bfe7d3eb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fhd6z" Mar 07 15:12:42 crc kubenswrapper[4943]: I0307 15:12:42.486319 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/de15165d-f055-468e-8f76-23f8bfe7d3eb-etc-swift\") pod \"swift-ring-rebalance-debug-fhd6z\" (UID: \"de15165d-f055-468e-8f76-23f8bfe7d3eb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fhd6z" Mar 07 15:12:42 crc kubenswrapper[4943]: I0307 15:12:42.587765 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/de15165d-f055-468e-8f76-23f8bfe7d3eb-dispersionconf\") pod \"swift-ring-rebalance-debug-fhd6z\" (UID: \"de15165d-f055-468e-8f76-23f8bfe7d3eb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fhd6z" Mar 07 15:12:42 crc kubenswrapper[4943]: I0307 15:12:42.587835 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/de15165d-f055-468e-8f76-23f8bfe7d3eb-swiftconf\") pod \"swift-ring-rebalance-debug-fhd6z\" (UID: \"de15165d-f055-468e-8f76-23f8bfe7d3eb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fhd6z" Mar 07 15:12:42 crc kubenswrapper[4943]: I0307 15:12:42.587907 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/de15165d-f055-468e-8f76-23f8bfe7d3eb-scripts\") pod \"swift-ring-rebalance-debug-fhd6z\" (UID: \"de15165d-f055-468e-8f76-23f8bfe7d3eb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fhd6z" Mar 07 15:12:42 crc kubenswrapper[4943]: I0307 15:12:42.587958 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmwfd\" (UniqueName: \"kubernetes.io/projected/de15165d-f055-468e-8f76-23f8bfe7d3eb-kube-api-access-lmwfd\") pod \"swift-ring-rebalance-debug-fhd6z\" (UID: \"de15165d-f055-468e-8f76-23f8bfe7d3eb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fhd6z" Mar 07 15:12:42 crc kubenswrapper[4943]: I0307 15:12:42.587986 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/de15165d-f055-468e-8f76-23f8bfe7d3eb-etc-swift\") pod \"swift-ring-rebalance-debug-fhd6z\" (UID: \"de15165d-f055-468e-8f76-23f8bfe7d3eb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fhd6z" Mar 07 15:12:42 crc kubenswrapper[4943]: I0307 15:12:42.588014 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/de15165d-f055-468e-8f76-23f8bfe7d3eb-ring-data-devices\") pod \"swift-ring-rebalance-debug-fhd6z\" (UID: \"de15165d-f055-468e-8f76-23f8bfe7d3eb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fhd6z" Mar 07 15:12:42 crc kubenswrapper[4943]: I0307 15:12:42.589072 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/de15165d-f055-468e-8f76-23f8bfe7d3eb-ring-data-devices\") pod \"swift-ring-rebalance-debug-fhd6z\" (UID: \"de15165d-f055-468e-8f76-23f8bfe7d3eb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fhd6z" Mar 07 15:12:42 crc kubenswrapper[4943]: I0307 15:12:42.589352 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/de15165d-f055-468e-8f76-23f8bfe7d3eb-scripts\") pod \"swift-ring-rebalance-debug-fhd6z\" (UID: \"de15165d-f055-468e-8f76-23f8bfe7d3eb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fhd6z" Mar 07 15:12:42 crc kubenswrapper[4943]: I0307 15:12:42.589601 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/de15165d-f055-468e-8f76-23f8bfe7d3eb-etc-swift\") pod \"swift-ring-rebalance-debug-fhd6z\" (UID: \"de15165d-f055-468e-8f76-23f8bfe7d3eb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fhd6z" Mar 07 15:12:42 crc kubenswrapper[4943]: I0307 15:12:42.594708 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/de15165d-f055-468e-8f76-23f8bfe7d3eb-swiftconf\") pod \"swift-ring-rebalance-debug-fhd6z\" (UID: \"de15165d-f055-468e-8f76-23f8bfe7d3eb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fhd6z" Mar 07 15:12:42 crc kubenswrapper[4943]: I0307 15:12:42.595253 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/de15165d-f055-468e-8f76-23f8bfe7d3eb-dispersionconf\") pod \"swift-ring-rebalance-debug-fhd6z\" (UID: \"de15165d-f055-468e-8f76-23f8bfe7d3eb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fhd6z" Mar 07 15:12:42 crc kubenswrapper[4943]: I0307 15:12:42.614067 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmwfd\" (UniqueName: \"kubernetes.io/projected/de15165d-f055-468e-8f76-23f8bfe7d3eb-kube-api-access-lmwfd\") pod \"swift-ring-rebalance-debug-fhd6z\" (UID: \"de15165d-f055-468e-8f76-23f8bfe7d3eb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fhd6z" Mar 07 15:12:42 crc kubenswrapper[4943]: I0307 15:12:42.677623 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-fhd6z" Mar 07 15:12:42 crc kubenswrapper[4943]: I0307 15:12:42.767995 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7f7231f-c606-43b7-b4a7-5990f7544198" path="/var/lib/kubelet/pods/c7f7231f-c606-43b7-b4a7-5990f7544198/volumes" Mar 07 15:12:43 crc kubenswrapper[4943]: I0307 15:12:43.162771 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-fhd6z"] Mar 07 15:12:43 crc kubenswrapper[4943]: W0307 15:12:43.169586 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podde15165d_f055_468e_8f76_23f8bfe7d3eb.slice/crio-994d0e6b90fc42244e424eb7ac11f1a7e4fd83174d94677a228fd0ac64f789cc WatchSource:0}: Error finding container 994d0e6b90fc42244e424eb7ac11f1a7e4fd83174d94677a228fd0ac64f789cc: Status 404 returned error can't find the container with id 994d0e6b90fc42244e424eb7ac11f1a7e4fd83174d94677a228fd0ac64f789cc Mar 07 15:12:43 crc kubenswrapper[4943]: I0307 15:12:43.836582 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-fhd6z" event={"ID":"de15165d-f055-468e-8f76-23f8bfe7d3eb","Type":"ContainerStarted","Data":"431d954b53a18c9fb4e5d1c686ada0ecacdd1d316b855d9e4f70fcf6ac8b45ed"} Mar 07 15:12:43 crc kubenswrapper[4943]: I0307 15:12:43.836830 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-fhd6z" event={"ID":"de15165d-f055-468e-8f76-23f8bfe7d3eb","Type":"ContainerStarted","Data":"994d0e6b90fc42244e424eb7ac11f1a7e4fd83174d94677a228fd0ac64f789cc"} Mar 07 15:12:43 crc kubenswrapper[4943]: I0307 15:12:43.859908 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-fhd6z" podStartSLOduration=1.859888153 podStartE2EDuration="1.859888153s" podCreationTimestamp="2026-03-07 15:12:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:12:43.858913659 +0000 UTC m=+2005.811050187" watchObservedRunningTime="2026-03-07 15:12:43.859888153 +0000 UTC m=+2005.812024661" Mar 07 15:12:44 crc kubenswrapper[4943]: I0307 15:12:44.860873 4943 generic.go:334] "Generic (PLEG): container finished" podID="de15165d-f055-468e-8f76-23f8bfe7d3eb" containerID="431d954b53a18c9fb4e5d1c686ada0ecacdd1d316b855d9e4f70fcf6ac8b45ed" exitCode=0 Mar 07 15:12:44 crc kubenswrapper[4943]: I0307 15:12:44.860965 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-fhd6z" event={"ID":"de15165d-f055-468e-8f76-23f8bfe7d3eb","Type":"ContainerDied","Data":"431d954b53a18c9fb4e5d1c686ada0ecacdd1d316b855d9e4f70fcf6ac8b45ed"} Mar 07 15:12:46 crc kubenswrapper[4943]: I0307 15:12:46.243706 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-fhd6z" Mar 07 15:12:46 crc kubenswrapper[4943]: I0307 15:12:46.294124 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-fhd6z"] Mar 07 15:12:46 crc kubenswrapper[4943]: I0307 15:12:46.301214 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-fhd6z"] Mar 07 15:12:46 crc kubenswrapper[4943]: I0307 15:12:46.354576 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/de15165d-f055-468e-8f76-23f8bfe7d3eb-ring-data-devices\") pod \"de15165d-f055-468e-8f76-23f8bfe7d3eb\" (UID: \"de15165d-f055-468e-8f76-23f8bfe7d3eb\") " Mar 07 15:12:46 crc kubenswrapper[4943]: I0307 15:12:46.354626 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/de15165d-f055-468e-8f76-23f8bfe7d3eb-dispersionconf\") pod \"de15165d-f055-468e-8f76-23f8bfe7d3eb\" (UID: \"de15165d-f055-468e-8f76-23f8bfe7d3eb\") " Mar 07 15:12:46 crc kubenswrapper[4943]: I0307 15:12:46.354675 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/de15165d-f055-468e-8f76-23f8bfe7d3eb-scripts\") pod \"de15165d-f055-468e-8f76-23f8bfe7d3eb\" (UID: \"de15165d-f055-468e-8f76-23f8bfe7d3eb\") " Mar 07 15:12:46 crc kubenswrapper[4943]: I0307 15:12:46.354724 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/de15165d-f055-468e-8f76-23f8bfe7d3eb-swiftconf\") pod \"de15165d-f055-468e-8f76-23f8bfe7d3eb\" (UID: \"de15165d-f055-468e-8f76-23f8bfe7d3eb\") " Mar 07 15:12:46 crc kubenswrapper[4943]: I0307 15:12:46.354804 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lmwfd\" (UniqueName: \"kubernetes.io/projected/de15165d-f055-468e-8f76-23f8bfe7d3eb-kube-api-access-lmwfd\") pod \"de15165d-f055-468e-8f76-23f8bfe7d3eb\" (UID: \"de15165d-f055-468e-8f76-23f8bfe7d3eb\") " Mar 07 15:12:46 crc kubenswrapper[4943]: I0307 15:12:46.354836 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/de15165d-f055-468e-8f76-23f8bfe7d3eb-etc-swift\") pod \"de15165d-f055-468e-8f76-23f8bfe7d3eb\" (UID: \"de15165d-f055-468e-8f76-23f8bfe7d3eb\") " Mar 07 15:12:46 crc kubenswrapper[4943]: I0307 15:12:46.355464 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de15165d-f055-468e-8f76-23f8bfe7d3eb-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "de15165d-f055-468e-8f76-23f8bfe7d3eb" (UID: "de15165d-f055-468e-8f76-23f8bfe7d3eb"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:12:46 crc kubenswrapper[4943]: I0307 15:12:46.356184 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de15165d-f055-468e-8f76-23f8bfe7d3eb-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "de15165d-f055-468e-8f76-23f8bfe7d3eb" (UID: "de15165d-f055-468e-8f76-23f8bfe7d3eb"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:12:46 crc kubenswrapper[4943]: I0307 15:12:46.363047 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de15165d-f055-468e-8f76-23f8bfe7d3eb-kube-api-access-lmwfd" (OuterVolumeSpecName: "kube-api-access-lmwfd") pod "de15165d-f055-468e-8f76-23f8bfe7d3eb" (UID: "de15165d-f055-468e-8f76-23f8bfe7d3eb"). InnerVolumeSpecName "kube-api-access-lmwfd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:12:46 crc kubenswrapper[4943]: I0307 15:12:46.375150 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de15165d-f055-468e-8f76-23f8bfe7d3eb-scripts" (OuterVolumeSpecName: "scripts") pod "de15165d-f055-468e-8f76-23f8bfe7d3eb" (UID: "de15165d-f055-468e-8f76-23f8bfe7d3eb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:12:46 crc kubenswrapper[4943]: I0307 15:12:46.386873 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de15165d-f055-468e-8f76-23f8bfe7d3eb-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "de15165d-f055-468e-8f76-23f8bfe7d3eb" (UID: "de15165d-f055-468e-8f76-23f8bfe7d3eb"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:12:46 crc kubenswrapper[4943]: I0307 15:12:46.396362 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de15165d-f055-468e-8f76-23f8bfe7d3eb-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "de15165d-f055-468e-8f76-23f8bfe7d3eb" (UID: "de15165d-f055-468e-8f76-23f8bfe7d3eb"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:12:46 crc kubenswrapper[4943]: I0307 15:12:46.456286 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/de15165d-f055-468e-8f76-23f8bfe7d3eb-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:46 crc kubenswrapper[4943]: I0307 15:12:46.456320 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lmwfd\" (UniqueName: \"kubernetes.io/projected/de15165d-f055-468e-8f76-23f8bfe7d3eb-kube-api-access-lmwfd\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:46 crc kubenswrapper[4943]: I0307 15:12:46.456332 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/de15165d-f055-468e-8f76-23f8bfe7d3eb-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:46 crc kubenswrapper[4943]: I0307 15:12:46.456340 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/de15165d-f055-468e-8f76-23f8bfe7d3eb-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:46 crc kubenswrapper[4943]: I0307 15:12:46.456349 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/de15165d-f055-468e-8f76-23f8bfe7d3eb-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:46 crc kubenswrapper[4943]: I0307 15:12:46.456357 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/de15165d-f055-468e-8f76-23f8bfe7d3eb-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:46 crc kubenswrapper[4943]: I0307 15:12:46.771210 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de15165d-f055-468e-8f76-23f8bfe7d3eb" path="/var/lib/kubelet/pods/de15165d-f055-468e-8f76-23f8bfe7d3eb/volumes" Mar 07 15:12:46 crc kubenswrapper[4943]: I0307 15:12:46.885316 4943 scope.go:117] "RemoveContainer" containerID="431d954b53a18c9fb4e5d1c686ada0ecacdd1d316b855d9e4f70fcf6ac8b45ed" Mar 07 15:12:46 crc kubenswrapper[4943]: I0307 15:12:46.885389 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-fhd6z" Mar 07 15:12:47 crc kubenswrapper[4943]: I0307 15:12:47.456510 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-zk7wc"] Mar 07 15:12:47 crc kubenswrapper[4943]: E0307 15:12:47.459257 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de15165d-f055-468e-8f76-23f8bfe7d3eb" containerName="swift-ring-rebalance" Mar 07 15:12:47 crc kubenswrapper[4943]: I0307 15:12:47.459297 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="de15165d-f055-468e-8f76-23f8bfe7d3eb" containerName="swift-ring-rebalance" Mar 07 15:12:47 crc kubenswrapper[4943]: I0307 15:12:47.459601 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="de15165d-f055-468e-8f76-23f8bfe7d3eb" containerName="swift-ring-rebalance" Mar 07 15:12:47 crc kubenswrapper[4943]: I0307 15:12:47.460524 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-zk7wc" Mar 07 15:12:47 crc kubenswrapper[4943]: I0307 15:12:47.464664 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:12:47 crc kubenswrapper[4943]: I0307 15:12:47.465886 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:12:47 crc kubenswrapper[4943]: I0307 15:12:47.472827 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4sjs\" (UniqueName: \"kubernetes.io/projected/ee60e798-5c79-49bf-b1b0-e8ad5c0564a0-kube-api-access-c4sjs\") pod \"swift-ring-rebalance-debug-zk7wc\" (UID: \"ee60e798-5c79-49bf-b1b0-e8ad5c0564a0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zk7wc" Mar 07 15:12:47 crc kubenswrapper[4943]: I0307 15:12:47.472896 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ee60e798-5c79-49bf-b1b0-e8ad5c0564a0-ring-data-devices\") pod \"swift-ring-rebalance-debug-zk7wc\" (UID: \"ee60e798-5c79-49bf-b1b0-e8ad5c0564a0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zk7wc" Mar 07 15:12:47 crc kubenswrapper[4943]: I0307 15:12:47.473194 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ee60e798-5c79-49bf-b1b0-e8ad5c0564a0-etc-swift\") pod \"swift-ring-rebalance-debug-zk7wc\" (UID: \"ee60e798-5c79-49bf-b1b0-e8ad5c0564a0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zk7wc" Mar 07 15:12:47 crc kubenswrapper[4943]: I0307 15:12:47.473296 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ee60e798-5c79-49bf-b1b0-e8ad5c0564a0-scripts\") pod \"swift-ring-rebalance-debug-zk7wc\" (UID: \"ee60e798-5c79-49bf-b1b0-e8ad5c0564a0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zk7wc" Mar 07 15:12:47 crc kubenswrapper[4943]: I0307 15:12:47.473323 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ee60e798-5c79-49bf-b1b0-e8ad5c0564a0-dispersionconf\") pod \"swift-ring-rebalance-debug-zk7wc\" (UID: \"ee60e798-5c79-49bf-b1b0-e8ad5c0564a0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zk7wc" Mar 07 15:12:47 crc kubenswrapper[4943]: I0307 15:12:47.473402 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ee60e798-5c79-49bf-b1b0-e8ad5c0564a0-swiftconf\") pod \"swift-ring-rebalance-debug-zk7wc\" (UID: \"ee60e798-5c79-49bf-b1b0-e8ad5c0564a0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zk7wc" Mar 07 15:12:47 crc kubenswrapper[4943]: I0307 15:12:47.489127 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-zk7wc"] Mar 07 15:12:47 crc kubenswrapper[4943]: I0307 15:12:47.575097 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ee60e798-5c79-49bf-b1b0-e8ad5c0564a0-ring-data-devices\") pod \"swift-ring-rebalance-debug-zk7wc\" (UID: \"ee60e798-5c79-49bf-b1b0-e8ad5c0564a0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zk7wc" Mar 07 15:12:47 crc kubenswrapper[4943]: I0307 15:12:47.575181 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ee60e798-5c79-49bf-b1b0-e8ad5c0564a0-etc-swift\") pod \"swift-ring-rebalance-debug-zk7wc\" (UID: \"ee60e798-5c79-49bf-b1b0-e8ad5c0564a0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zk7wc" Mar 07 15:12:47 crc kubenswrapper[4943]: I0307 15:12:47.575205 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ee60e798-5c79-49bf-b1b0-e8ad5c0564a0-dispersionconf\") pod \"swift-ring-rebalance-debug-zk7wc\" (UID: \"ee60e798-5c79-49bf-b1b0-e8ad5c0564a0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zk7wc" Mar 07 15:12:47 crc kubenswrapper[4943]: I0307 15:12:47.575223 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ee60e798-5c79-49bf-b1b0-e8ad5c0564a0-scripts\") pod \"swift-ring-rebalance-debug-zk7wc\" (UID: \"ee60e798-5c79-49bf-b1b0-e8ad5c0564a0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zk7wc" Mar 07 15:12:47 crc kubenswrapper[4943]: I0307 15:12:47.575252 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ee60e798-5c79-49bf-b1b0-e8ad5c0564a0-swiftconf\") pod \"swift-ring-rebalance-debug-zk7wc\" (UID: \"ee60e798-5c79-49bf-b1b0-e8ad5c0564a0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zk7wc" Mar 07 15:12:47 crc kubenswrapper[4943]: I0307 15:12:47.575291 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4sjs\" (UniqueName: \"kubernetes.io/projected/ee60e798-5c79-49bf-b1b0-e8ad5c0564a0-kube-api-access-c4sjs\") pod \"swift-ring-rebalance-debug-zk7wc\" (UID: \"ee60e798-5c79-49bf-b1b0-e8ad5c0564a0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zk7wc" Mar 07 15:12:47 crc kubenswrapper[4943]: I0307 15:12:47.576021 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ee60e798-5c79-49bf-b1b0-e8ad5c0564a0-etc-swift\") pod \"swift-ring-rebalance-debug-zk7wc\" (UID: \"ee60e798-5c79-49bf-b1b0-e8ad5c0564a0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zk7wc" Mar 07 15:12:47 crc kubenswrapper[4943]: I0307 15:12:47.577164 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ee60e798-5c79-49bf-b1b0-e8ad5c0564a0-scripts\") pod \"swift-ring-rebalance-debug-zk7wc\" (UID: \"ee60e798-5c79-49bf-b1b0-e8ad5c0564a0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zk7wc" Mar 07 15:12:47 crc kubenswrapper[4943]: I0307 15:12:47.577223 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ee60e798-5c79-49bf-b1b0-e8ad5c0564a0-ring-data-devices\") pod \"swift-ring-rebalance-debug-zk7wc\" (UID: \"ee60e798-5c79-49bf-b1b0-e8ad5c0564a0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zk7wc" Mar 07 15:12:47 crc kubenswrapper[4943]: I0307 15:12:47.583157 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ee60e798-5c79-49bf-b1b0-e8ad5c0564a0-dispersionconf\") pod \"swift-ring-rebalance-debug-zk7wc\" (UID: \"ee60e798-5c79-49bf-b1b0-e8ad5c0564a0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zk7wc" Mar 07 15:12:47 crc kubenswrapper[4943]: I0307 15:12:47.583214 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ee60e798-5c79-49bf-b1b0-e8ad5c0564a0-swiftconf\") pod \"swift-ring-rebalance-debug-zk7wc\" (UID: \"ee60e798-5c79-49bf-b1b0-e8ad5c0564a0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zk7wc" Mar 07 15:12:47 crc kubenswrapper[4943]: I0307 15:12:47.594740 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4sjs\" (UniqueName: \"kubernetes.io/projected/ee60e798-5c79-49bf-b1b0-e8ad5c0564a0-kube-api-access-c4sjs\") pod \"swift-ring-rebalance-debug-zk7wc\" (UID: \"ee60e798-5c79-49bf-b1b0-e8ad5c0564a0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zk7wc" Mar 07 15:12:47 crc kubenswrapper[4943]: I0307 15:12:47.783133 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-zk7wc" Mar 07 15:12:48 crc kubenswrapper[4943]: I0307 15:12:48.027822 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-zk7wc"] Mar 07 15:12:48 crc kubenswrapper[4943]: W0307 15:12:48.028650 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podee60e798_5c79_49bf_b1b0_e8ad5c0564a0.slice/crio-c01f4b7b7d738c444fc04f81deee9c610bd00431c5beb990fc9aff5ace29466d WatchSource:0}: Error finding container c01f4b7b7d738c444fc04f81deee9c610bd00431c5beb990fc9aff5ace29466d: Status 404 returned error can't find the container with id c01f4b7b7d738c444fc04f81deee9c610bd00431c5beb990fc9aff5ace29466d Mar 07 15:12:48 crc kubenswrapper[4943]: I0307 15:12:48.919669 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-zk7wc" event={"ID":"ee60e798-5c79-49bf-b1b0-e8ad5c0564a0","Type":"ContainerStarted","Data":"5ef4437771ebe7f583d725cefda637736677ea6c6a02e7b62dac24c36a88ebf1"} Mar 07 15:12:48 crc kubenswrapper[4943]: I0307 15:12:48.919984 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-zk7wc" event={"ID":"ee60e798-5c79-49bf-b1b0-e8ad5c0564a0","Type":"ContainerStarted","Data":"c01f4b7b7d738c444fc04f81deee9c610bd00431c5beb990fc9aff5ace29466d"} Mar 07 15:12:48 crc kubenswrapper[4943]: I0307 15:12:48.937822 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-zk7wc" podStartSLOduration=1.937804662 podStartE2EDuration="1.937804662s" podCreationTimestamp="2026-03-07 15:12:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:12:48.932868501 +0000 UTC m=+2010.885004999" watchObservedRunningTime="2026-03-07 15:12:48.937804662 +0000 UTC m=+2010.889941160" Mar 07 15:12:49 crc kubenswrapper[4943]: I0307 15:12:49.935730 4943 generic.go:334] "Generic (PLEG): container finished" podID="ee60e798-5c79-49bf-b1b0-e8ad5c0564a0" containerID="5ef4437771ebe7f583d725cefda637736677ea6c6a02e7b62dac24c36a88ebf1" exitCode=0 Mar 07 15:12:49 crc kubenswrapper[4943]: I0307 15:12:49.935824 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-zk7wc" event={"ID":"ee60e798-5c79-49bf-b1b0-e8ad5c0564a0","Type":"ContainerDied","Data":"5ef4437771ebe7f583d725cefda637736677ea6c6a02e7b62dac24c36a88ebf1"} Mar 07 15:12:51 crc kubenswrapper[4943]: I0307 15:12:51.320503 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-zk7wc" Mar 07 15:12:51 crc kubenswrapper[4943]: I0307 15:12:51.355011 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-zk7wc"] Mar 07 15:12:51 crc kubenswrapper[4943]: I0307 15:12:51.359471 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-zk7wc"] Mar 07 15:12:51 crc kubenswrapper[4943]: I0307 15:12:51.453858 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ee60e798-5c79-49bf-b1b0-e8ad5c0564a0-scripts\") pod \"ee60e798-5c79-49bf-b1b0-e8ad5c0564a0\" (UID: \"ee60e798-5c79-49bf-b1b0-e8ad5c0564a0\") " Mar 07 15:12:51 crc kubenswrapper[4943]: I0307 15:12:51.453907 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ee60e798-5c79-49bf-b1b0-e8ad5c0564a0-swiftconf\") pod \"ee60e798-5c79-49bf-b1b0-e8ad5c0564a0\" (UID: \"ee60e798-5c79-49bf-b1b0-e8ad5c0564a0\") " Mar 07 15:12:51 crc kubenswrapper[4943]: I0307 15:12:51.453951 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ee60e798-5c79-49bf-b1b0-e8ad5c0564a0-dispersionconf\") pod \"ee60e798-5c79-49bf-b1b0-e8ad5c0564a0\" (UID: \"ee60e798-5c79-49bf-b1b0-e8ad5c0564a0\") " Mar 07 15:12:51 crc kubenswrapper[4943]: I0307 15:12:51.454010 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ee60e798-5c79-49bf-b1b0-e8ad5c0564a0-etc-swift\") pod \"ee60e798-5c79-49bf-b1b0-e8ad5c0564a0\" (UID: \"ee60e798-5c79-49bf-b1b0-e8ad5c0564a0\") " Mar 07 15:12:51 crc kubenswrapper[4943]: I0307 15:12:51.454064 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c4sjs\" (UniqueName: \"kubernetes.io/projected/ee60e798-5c79-49bf-b1b0-e8ad5c0564a0-kube-api-access-c4sjs\") pod \"ee60e798-5c79-49bf-b1b0-e8ad5c0564a0\" (UID: \"ee60e798-5c79-49bf-b1b0-e8ad5c0564a0\") " Mar 07 15:12:51 crc kubenswrapper[4943]: I0307 15:12:51.454175 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ee60e798-5c79-49bf-b1b0-e8ad5c0564a0-ring-data-devices\") pod \"ee60e798-5c79-49bf-b1b0-e8ad5c0564a0\" (UID: \"ee60e798-5c79-49bf-b1b0-e8ad5c0564a0\") " Mar 07 15:12:51 crc kubenswrapper[4943]: I0307 15:12:51.454820 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee60e798-5c79-49bf-b1b0-e8ad5c0564a0-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "ee60e798-5c79-49bf-b1b0-e8ad5c0564a0" (UID: "ee60e798-5c79-49bf-b1b0-e8ad5c0564a0"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:12:51 crc kubenswrapper[4943]: I0307 15:12:51.454847 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee60e798-5c79-49bf-b1b0-e8ad5c0564a0-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "ee60e798-5c79-49bf-b1b0-e8ad5c0564a0" (UID: "ee60e798-5c79-49bf-b1b0-e8ad5c0564a0"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:12:51 crc kubenswrapper[4943]: I0307 15:12:51.461360 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee60e798-5c79-49bf-b1b0-e8ad5c0564a0-kube-api-access-c4sjs" (OuterVolumeSpecName: "kube-api-access-c4sjs") pod "ee60e798-5c79-49bf-b1b0-e8ad5c0564a0" (UID: "ee60e798-5c79-49bf-b1b0-e8ad5c0564a0"). InnerVolumeSpecName "kube-api-access-c4sjs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:12:51 crc kubenswrapper[4943]: I0307 15:12:51.478293 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee60e798-5c79-49bf-b1b0-e8ad5c0564a0-scripts" (OuterVolumeSpecName: "scripts") pod "ee60e798-5c79-49bf-b1b0-e8ad5c0564a0" (UID: "ee60e798-5c79-49bf-b1b0-e8ad5c0564a0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:12:51 crc kubenswrapper[4943]: I0307 15:12:51.481894 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee60e798-5c79-49bf-b1b0-e8ad5c0564a0-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "ee60e798-5c79-49bf-b1b0-e8ad5c0564a0" (UID: "ee60e798-5c79-49bf-b1b0-e8ad5c0564a0"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:12:51 crc kubenswrapper[4943]: I0307 15:12:51.504510 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee60e798-5c79-49bf-b1b0-e8ad5c0564a0-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "ee60e798-5c79-49bf-b1b0-e8ad5c0564a0" (UID: "ee60e798-5c79-49bf-b1b0-e8ad5c0564a0"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:12:51 crc kubenswrapper[4943]: I0307 15:12:51.556063 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ee60e798-5c79-49bf-b1b0-e8ad5c0564a0-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:51 crc kubenswrapper[4943]: I0307 15:12:51.556100 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ee60e798-5c79-49bf-b1b0-e8ad5c0564a0-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:51 crc kubenswrapper[4943]: I0307 15:12:51.556112 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ee60e798-5c79-49bf-b1b0-e8ad5c0564a0-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:51 crc kubenswrapper[4943]: I0307 15:12:51.556126 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ee60e798-5c79-49bf-b1b0-e8ad5c0564a0-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:51 crc kubenswrapper[4943]: I0307 15:12:51.556138 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ee60e798-5c79-49bf-b1b0-e8ad5c0564a0-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:51 crc kubenswrapper[4943]: I0307 15:12:51.556151 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c4sjs\" (UniqueName: \"kubernetes.io/projected/ee60e798-5c79-49bf-b1b0-e8ad5c0564a0-kube-api-access-c4sjs\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:51 crc kubenswrapper[4943]: I0307 15:12:51.970672 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c01f4b7b7d738c444fc04f81deee9c610bd00431c5beb990fc9aff5ace29466d" Mar 07 15:12:51 crc kubenswrapper[4943]: I0307 15:12:51.970783 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-zk7wc" Mar 07 15:12:52 crc kubenswrapper[4943]: I0307 15:12:52.540219 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-b46t6"] Mar 07 15:12:52 crc kubenswrapper[4943]: E0307 15:12:52.540635 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee60e798-5c79-49bf-b1b0-e8ad5c0564a0" containerName="swift-ring-rebalance" Mar 07 15:12:52 crc kubenswrapper[4943]: I0307 15:12:52.540655 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee60e798-5c79-49bf-b1b0-e8ad5c0564a0" containerName="swift-ring-rebalance" Mar 07 15:12:52 crc kubenswrapper[4943]: I0307 15:12:52.540966 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee60e798-5c79-49bf-b1b0-e8ad5c0564a0" containerName="swift-ring-rebalance" Mar 07 15:12:52 crc kubenswrapper[4943]: I0307 15:12:52.541699 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-b46t6" Mar 07 15:12:52 crc kubenswrapper[4943]: I0307 15:12:52.547716 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:12:52 crc kubenswrapper[4943]: I0307 15:12:52.550238 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:12:52 crc kubenswrapper[4943]: I0307 15:12:52.586484 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-b46t6"] Mar 07 15:12:52 crc kubenswrapper[4943]: I0307 15:12:52.676565 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19-scripts\") pod \"swift-ring-rebalance-debug-b46t6\" (UID: \"aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-b46t6" Mar 07 15:12:52 crc kubenswrapper[4943]: I0307 15:12:52.676666 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19-dispersionconf\") pod \"swift-ring-rebalance-debug-b46t6\" (UID: \"aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-b46t6" Mar 07 15:12:52 crc kubenswrapper[4943]: I0307 15:12:52.676776 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19-swiftconf\") pod \"swift-ring-rebalance-debug-b46t6\" (UID: \"aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-b46t6" Mar 07 15:12:52 crc kubenswrapper[4943]: I0307 15:12:52.676828 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19-ring-data-devices\") pod \"swift-ring-rebalance-debug-b46t6\" (UID: \"aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-b46t6" Mar 07 15:12:52 crc kubenswrapper[4943]: I0307 15:12:52.677012 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lkj8\" (UniqueName: \"kubernetes.io/projected/aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19-kube-api-access-8lkj8\") pod \"swift-ring-rebalance-debug-b46t6\" (UID: \"aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-b46t6" Mar 07 15:12:52 crc kubenswrapper[4943]: I0307 15:12:52.677056 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19-etc-swift\") pod \"swift-ring-rebalance-debug-b46t6\" (UID: \"aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-b46t6" Mar 07 15:12:52 crc kubenswrapper[4943]: I0307 15:12:52.770374 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee60e798-5c79-49bf-b1b0-e8ad5c0564a0" path="/var/lib/kubelet/pods/ee60e798-5c79-49bf-b1b0-e8ad5c0564a0/volumes" Mar 07 15:12:52 crc kubenswrapper[4943]: I0307 15:12:52.779064 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lkj8\" (UniqueName: \"kubernetes.io/projected/aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19-kube-api-access-8lkj8\") pod \"swift-ring-rebalance-debug-b46t6\" (UID: \"aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-b46t6" Mar 07 15:12:52 crc kubenswrapper[4943]: I0307 15:12:52.779110 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19-etc-swift\") pod \"swift-ring-rebalance-debug-b46t6\" (UID: \"aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-b46t6" Mar 07 15:12:52 crc kubenswrapper[4943]: I0307 15:12:52.779164 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19-scripts\") pod \"swift-ring-rebalance-debug-b46t6\" (UID: \"aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-b46t6" Mar 07 15:12:52 crc kubenswrapper[4943]: I0307 15:12:52.779185 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19-dispersionconf\") pod \"swift-ring-rebalance-debug-b46t6\" (UID: \"aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-b46t6" Mar 07 15:12:52 crc kubenswrapper[4943]: I0307 15:12:52.779210 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19-swiftconf\") pod \"swift-ring-rebalance-debug-b46t6\" (UID: \"aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-b46t6" Mar 07 15:12:52 crc kubenswrapper[4943]: I0307 15:12:52.779231 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19-ring-data-devices\") pod \"swift-ring-rebalance-debug-b46t6\" (UID: \"aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-b46t6" Mar 07 15:12:52 crc kubenswrapper[4943]: I0307 15:12:52.779985 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19-ring-data-devices\") pod \"swift-ring-rebalance-debug-b46t6\" (UID: \"aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-b46t6" Mar 07 15:12:52 crc kubenswrapper[4943]: I0307 15:12:52.780452 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19-scripts\") pod \"swift-ring-rebalance-debug-b46t6\" (UID: \"aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-b46t6" Mar 07 15:12:52 crc kubenswrapper[4943]: I0307 15:12:52.781231 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19-etc-swift\") pod \"swift-ring-rebalance-debug-b46t6\" (UID: \"aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-b46t6" Mar 07 15:12:52 crc kubenswrapper[4943]: I0307 15:12:52.785992 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19-swiftconf\") pod \"swift-ring-rebalance-debug-b46t6\" (UID: \"aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-b46t6" Mar 07 15:12:52 crc kubenswrapper[4943]: I0307 15:12:52.794494 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19-dispersionconf\") pod \"swift-ring-rebalance-debug-b46t6\" (UID: \"aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-b46t6" Mar 07 15:12:52 crc kubenswrapper[4943]: I0307 15:12:52.813219 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lkj8\" (UniqueName: \"kubernetes.io/projected/aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19-kube-api-access-8lkj8\") pod \"swift-ring-rebalance-debug-b46t6\" (UID: \"aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-b46t6" Mar 07 15:12:52 crc kubenswrapper[4943]: I0307 15:12:52.868477 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-b46t6" Mar 07 15:12:53 crc kubenswrapper[4943]: I0307 15:12:53.347122 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-b46t6"] Mar 07 15:12:53 crc kubenswrapper[4943]: W0307 15:12:53.347906 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaedbc5bc_f6c7_4ac4_b1f8_32dbe4efff19.slice/crio-51788398e82f2170fe3feb90c6832a9a0468ae8c997d24d842ae77300f8c84e7 WatchSource:0}: Error finding container 51788398e82f2170fe3feb90c6832a9a0468ae8c997d24d842ae77300f8c84e7: Status 404 returned error can't find the container with id 51788398e82f2170fe3feb90c6832a9a0468ae8c997d24d842ae77300f8c84e7 Mar 07 15:12:54 crc kubenswrapper[4943]: I0307 15:12:54.008284 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-b46t6" event={"ID":"aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19","Type":"ContainerStarted","Data":"2527b6d3cb9bf1028972989d92691aaee552776ec1ae574fbced497ff1bbf606"} Mar 07 15:12:54 crc kubenswrapper[4943]: I0307 15:12:54.008374 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-b46t6" event={"ID":"aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19","Type":"ContainerStarted","Data":"51788398e82f2170fe3feb90c6832a9a0468ae8c997d24d842ae77300f8c84e7"} Mar 07 15:12:54 crc kubenswrapper[4943]: I0307 15:12:54.036522 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-b46t6" podStartSLOduration=2.036506669 podStartE2EDuration="2.036506669s" podCreationTimestamp="2026-03-07 15:12:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:12:54.033648029 +0000 UTC m=+2015.985784587" watchObservedRunningTime="2026-03-07 15:12:54.036506669 +0000 UTC m=+2015.988643157" Mar 07 15:12:55 crc kubenswrapper[4943]: I0307 15:12:55.024602 4943 generic.go:334] "Generic (PLEG): container finished" podID="aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19" containerID="2527b6d3cb9bf1028972989d92691aaee552776ec1ae574fbced497ff1bbf606" exitCode=0 Mar 07 15:12:55 crc kubenswrapper[4943]: I0307 15:12:55.024701 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-b46t6" event={"ID":"aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19","Type":"ContainerDied","Data":"2527b6d3cb9bf1028972989d92691aaee552776ec1ae574fbced497ff1bbf606"} Mar 07 15:12:56 crc kubenswrapper[4943]: I0307 15:12:56.426342 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-b46t6" Mar 07 15:12:56 crc kubenswrapper[4943]: I0307 15:12:56.455150 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-b46t6"] Mar 07 15:12:56 crc kubenswrapper[4943]: I0307 15:12:56.460107 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-b46t6"] Mar 07 15:12:56 crc kubenswrapper[4943]: I0307 15:12:56.538959 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19-ring-data-devices\") pod \"aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19\" (UID: \"aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19\") " Mar 07 15:12:56 crc kubenswrapper[4943]: I0307 15:12:56.539004 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19-scripts\") pod \"aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19\" (UID: \"aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19\") " Mar 07 15:12:56 crc kubenswrapper[4943]: I0307 15:12:56.539057 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19-swiftconf\") pod \"aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19\" (UID: \"aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19\") " Mar 07 15:12:56 crc kubenswrapper[4943]: I0307 15:12:56.539079 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19-etc-swift\") pod \"aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19\" (UID: \"aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19\") " Mar 07 15:12:56 crc kubenswrapper[4943]: I0307 15:12:56.539110 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8lkj8\" (UniqueName: \"kubernetes.io/projected/aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19-kube-api-access-8lkj8\") pod \"aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19\" (UID: \"aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19\") " Mar 07 15:12:56 crc kubenswrapper[4943]: I0307 15:12:56.539167 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19-dispersionconf\") pod \"aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19\" (UID: \"aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19\") " Mar 07 15:12:56 crc kubenswrapper[4943]: I0307 15:12:56.540114 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19" (UID: "aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:12:56 crc kubenswrapper[4943]: I0307 15:12:56.540529 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19" (UID: "aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:12:56 crc kubenswrapper[4943]: I0307 15:12:56.544322 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19-kube-api-access-8lkj8" (OuterVolumeSpecName: "kube-api-access-8lkj8") pod "aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19" (UID: "aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19"). InnerVolumeSpecName "kube-api-access-8lkj8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:12:56 crc kubenswrapper[4943]: I0307 15:12:56.564794 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19-scripts" (OuterVolumeSpecName: "scripts") pod "aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19" (UID: "aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:12:56 crc kubenswrapper[4943]: I0307 15:12:56.578551 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19" (UID: "aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:12:56 crc kubenswrapper[4943]: I0307 15:12:56.588095 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19" (UID: "aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:12:56 crc kubenswrapper[4943]: I0307 15:12:56.641290 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:56 crc kubenswrapper[4943]: I0307 15:12:56.641325 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:56 crc kubenswrapper[4943]: I0307 15:12:56.641338 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:56 crc kubenswrapper[4943]: I0307 15:12:56.641352 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:56 crc kubenswrapper[4943]: I0307 15:12:56.641367 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8lkj8\" (UniqueName: \"kubernetes.io/projected/aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19-kube-api-access-8lkj8\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:56 crc kubenswrapper[4943]: I0307 15:12:56.641522 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:12:56 crc kubenswrapper[4943]: I0307 15:12:56.769849 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19" path="/var/lib/kubelet/pods/aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19/volumes" Mar 07 15:12:57 crc kubenswrapper[4943]: I0307 15:12:57.057066 4943 scope.go:117] "RemoveContainer" containerID="2527b6d3cb9bf1028972989d92691aaee552776ec1ae574fbced497ff1bbf606" Mar 07 15:12:57 crc kubenswrapper[4943]: I0307 15:12:57.057266 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-b46t6" Mar 07 15:12:57 crc kubenswrapper[4943]: I0307 15:12:57.785901 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-wkvvv"] Mar 07 15:12:57 crc kubenswrapper[4943]: E0307 15:12:57.786390 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19" containerName="swift-ring-rebalance" Mar 07 15:12:57 crc kubenswrapper[4943]: I0307 15:12:57.786407 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19" containerName="swift-ring-rebalance" Mar 07 15:12:57 crc kubenswrapper[4943]: I0307 15:12:57.786586 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="aedbc5bc-f6c7-4ac4-b1f8-32dbe4efff19" containerName="swift-ring-rebalance" Mar 07 15:12:57 crc kubenswrapper[4943]: I0307 15:12:57.787171 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wkvvv" Mar 07 15:12:57 crc kubenswrapper[4943]: I0307 15:12:57.789608 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:12:57 crc kubenswrapper[4943]: I0307 15:12:57.792397 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:12:57 crc kubenswrapper[4943]: I0307 15:12:57.824269 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-wkvvv"] Mar 07 15:12:57 crc kubenswrapper[4943]: I0307 15:12:57.870353 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ca9a6b9c-6d95-4a78-9bdc-8890afd35eba-swiftconf\") pod \"swift-ring-rebalance-debug-wkvvv\" (UID: \"ca9a6b9c-6d95-4a78-9bdc-8890afd35eba\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wkvvv" Mar 07 15:12:57 crc kubenswrapper[4943]: I0307 15:12:57.870412 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwrls\" (UniqueName: \"kubernetes.io/projected/ca9a6b9c-6d95-4a78-9bdc-8890afd35eba-kube-api-access-zwrls\") pod \"swift-ring-rebalance-debug-wkvvv\" (UID: \"ca9a6b9c-6d95-4a78-9bdc-8890afd35eba\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wkvvv" Mar 07 15:12:57 crc kubenswrapper[4943]: I0307 15:12:57.870448 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ca9a6b9c-6d95-4a78-9bdc-8890afd35eba-ring-data-devices\") pod \"swift-ring-rebalance-debug-wkvvv\" (UID: \"ca9a6b9c-6d95-4a78-9bdc-8890afd35eba\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wkvvv" Mar 07 15:12:57 crc kubenswrapper[4943]: I0307 15:12:57.870475 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ca9a6b9c-6d95-4a78-9bdc-8890afd35eba-scripts\") pod \"swift-ring-rebalance-debug-wkvvv\" (UID: \"ca9a6b9c-6d95-4a78-9bdc-8890afd35eba\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wkvvv" Mar 07 15:12:57 crc kubenswrapper[4943]: I0307 15:12:57.870562 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ca9a6b9c-6d95-4a78-9bdc-8890afd35eba-dispersionconf\") pod \"swift-ring-rebalance-debug-wkvvv\" (UID: \"ca9a6b9c-6d95-4a78-9bdc-8890afd35eba\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wkvvv" Mar 07 15:12:57 crc kubenswrapper[4943]: I0307 15:12:57.870726 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ca9a6b9c-6d95-4a78-9bdc-8890afd35eba-etc-swift\") pod \"swift-ring-rebalance-debug-wkvvv\" (UID: \"ca9a6b9c-6d95-4a78-9bdc-8890afd35eba\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wkvvv" Mar 07 15:12:57 crc kubenswrapper[4943]: I0307 15:12:57.972288 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ca9a6b9c-6d95-4a78-9bdc-8890afd35eba-dispersionconf\") pod \"swift-ring-rebalance-debug-wkvvv\" (UID: \"ca9a6b9c-6d95-4a78-9bdc-8890afd35eba\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wkvvv" Mar 07 15:12:57 crc kubenswrapper[4943]: I0307 15:12:57.972425 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ca9a6b9c-6d95-4a78-9bdc-8890afd35eba-etc-swift\") pod \"swift-ring-rebalance-debug-wkvvv\" (UID: \"ca9a6b9c-6d95-4a78-9bdc-8890afd35eba\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wkvvv" Mar 07 15:12:57 crc kubenswrapper[4943]: I0307 15:12:57.972575 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ca9a6b9c-6d95-4a78-9bdc-8890afd35eba-swiftconf\") pod \"swift-ring-rebalance-debug-wkvvv\" (UID: \"ca9a6b9c-6d95-4a78-9bdc-8890afd35eba\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wkvvv" Mar 07 15:12:57 crc kubenswrapper[4943]: I0307 15:12:57.972619 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwrls\" (UniqueName: \"kubernetes.io/projected/ca9a6b9c-6d95-4a78-9bdc-8890afd35eba-kube-api-access-zwrls\") pod \"swift-ring-rebalance-debug-wkvvv\" (UID: \"ca9a6b9c-6d95-4a78-9bdc-8890afd35eba\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wkvvv" Mar 07 15:12:57 crc kubenswrapper[4943]: I0307 15:12:57.973039 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ca9a6b9c-6d95-4a78-9bdc-8890afd35eba-etc-swift\") pod \"swift-ring-rebalance-debug-wkvvv\" (UID: \"ca9a6b9c-6d95-4a78-9bdc-8890afd35eba\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wkvvv" Mar 07 15:12:57 crc kubenswrapper[4943]: I0307 15:12:57.973152 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ca9a6b9c-6d95-4a78-9bdc-8890afd35eba-ring-data-devices\") pod \"swift-ring-rebalance-debug-wkvvv\" (UID: \"ca9a6b9c-6d95-4a78-9bdc-8890afd35eba\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wkvvv" Mar 07 15:12:57 crc kubenswrapper[4943]: I0307 15:12:57.973899 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ca9a6b9c-6d95-4a78-9bdc-8890afd35eba-ring-data-devices\") pod \"swift-ring-rebalance-debug-wkvvv\" (UID: \"ca9a6b9c-6d95-4a78-9bdc-8890afd35eba\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wkvvv" Mar 07 15:12:57 crc kubenswrapper[4943]: I0307 15:12:57.974051 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ca9a6b9c-6d95-4a78-9bdc-8890afd35eba-scripts\") pod \"swift-ring-rebalance-debug-wkvvv\" (UID: \"ca9a6b9c-6d95-4a78-9bdc-8890afd35eba\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wkvvv" Mar 07 15:12:57 crc kubenswrapper[4943]: I0307 15:12:57.974187 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ca9a6b9c-6d95-4a78-9bdc-8890afd35eba-scripts\") pod \"swift-ring-rebalance-debug-wkvvv\" (UID: \"ca9a6b9c-6d95-4a78-9bdc-8890afd35eba\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wkvvv" Mar 07 15:12:57 crc kubenswrapper[4943]: I0307 15:12:57.976984 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ca9a6b9c-6d95-4a78-9bdc-8890afd35eba-swiftconf\") pod \"swift-ring-rebalance-debug-wkvvv\" (UID: \"ca9a6b9c-6d95-4a78-9bdc-8890afd35eba\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wkvvv" Mar 07 15:12:57 crc kubenswrapper[4943]: I0307 15:12:57.977224 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ca9a6b9c-6d95-4a78-9bdc-8890afd35eba-dispersionconf\") pod \"swift-ring-rebalance-debug-wkvvv\" (UID: \"ca9a6b9c-6d95-4a78-9bdc-8890afd35eba\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wkvvv" Mar 07 15:12:58 crc kubenswrapper[4943]: I0307 15:12:58.003062 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwrls\" (UniqueName: \"kubernetes.io/projected/ca9a6b9c-6d95-4a78-9bdc-8890afd35eba-kube-api-access-zwrls\") pod \"swift-ring-rebalance-debug-wkvvv\" (UID: \"ca9a6b9c-6d95-4a78-9bdc-8890afd35eba\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wkvvv" Mar 07 15:12:58 crc kubenswrapper[4943]: I0307 15:12:58.105546 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wkvvv" Mar 07 15:12:58 crc kubenswrapper[4943]: I0307 15:12:58.380145 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-wkvvv"] Mar 07 15:12:58 crc kubenswrapper[4943]: W0307 15:12:58.384822 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podca9a6b9c_6d95_4a78_9bdc_8890afd35eba.slice/crio-31dcf8cd73e914f3a524183c2ba89213432e40dcc70c3471dd07aa67b7ece1f0 WatchSource:0}: Error finding container 31dcf8cd73e914f3a524183c2ba89213432e40dcc70c3471dd07aa67b7ece1f0: Status 404 returned error can't find the container with id 31dcf8cd73e914f3a524183c2ba89213432e40dcc70c3471dd07aa67b7ece1f0 Mar 07 15:12:59 crc kubenswrapper[4943]: I0307 15:12:59.086723 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wkvvv" event={"ID":"ca9a6b9c-6d95-4a78-9bdc-8890afd35eba","Type":"ContainerStarted","Data":"8dba4aa6f09b0fe5fd0bfa8f4abad1b0555ab0d6d2cf190e7b2a9f200fbddb08"} Mar 07 15:12:59 crc kubenswrapper[4943]: I0307 15:12:59.087096 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wkvvv" event={"ID":"ca9a6b9c-6d95-4a78-9bdc-8890afd35eba","Type":"ContainerStarted","Data":"31dcf8cd73e914f3a524183c2ba89213432e40dcc70c3471dd07aa67b7ece1f0"} Mar 07 15:12:59 crc kubenswrapper[4943]: I0307 15:12:59.110814 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wkvvv" podStartSLOduration=2.110791108 podStartE2EDuration="2.110791108s" podCreationTimestamp="2026-03-07 15:12:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:12:59.103840607 +0000 UTC m=+2021.055977145" watchObservedRunningTime="2026-03-07 15:12:59.110791108 +0000 UTC m=+2021.062927646" Mar 07 15:13:00 crc kubenswrapper[4943]: I0307 15:13:00.118996 4943 generic.go:334] "Generic (PLEG): container finished" podID="ca9a6b9c-6d95-4a78-9bdc-8890afd35eba" containerID="8dba4aa6f09b0fe5fd0bfa8f4abad1b0555ab0d6d2cf190e7b2a9f200fbddb08" exitCode=0 Mar 07 15:13:00 crc kubenswrapper[4943]: I0307 15:13:00.119059 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wkvvv" event={"ID":"ca9a6b9c-6d95-4a78-9bdc-8890afd35eba","Type":"ContainerDied","Data":"8dba4aa6f09b0fe5fd0bfa8f4abad1b0555ab0d6d2cf190e7b2a9f200fbddb08"} Mar 07 15:13:01 crc kubenswrapper[4943]: I0307 15:13:01.737613 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wkvvv" Mar 07 15:13:01 crc kubenswrapper[4943]: I0307 15:13:01.797758 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-wkvvv"] Mar 07 15:13:01 crc kubenswrapper[4943]: I0307 15:13:01.801262 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ca9a6b9c-6d95-4a78-9bdc-8890afd35eba-etc-swift\") pod \"ca9a6b9c-6d95-4a78-9bdc-8890afd35eba\" (UID: \"ca9a6b9c-6d95-4a78-9bdc-8890afd35eba\") " Mar 07 15:13:01 crc kubenswrapper[4943]: I0307 15:13:01.801361 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zwrls\" (UniqueName: \"kubernetes.io/projected/ca9a6b9c-6d95-4a78-9bdc-8890afd35eba-kube-api-access-zwrls\") pod \"ca9a6b9c-6d95-4a78-9bdc-8890afd35eba\" (UID: \"ca9a6b9c-6d95-4a78-9bdc-8890afd35eba\") " Mar 07 15:13:01 crc kubenswrapper[4943]: I0307 15:13:01.801500 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ca9a6b9c-6d95-4a78-9bdc-8890afd35eba-ring-data-devices\") pod \"ca9a6b9c-6d95-4a78-9bdc-8890afd35eba\" (UID: \"ca9a6b9c-6d95-4a78-9bdc-8890afd35eba\") " Mar 07 15:13:01 crc kubenswrapper[4943]: I0307 15:13:01.801610 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ca9a6b9c-6d95-4a78-9bdc-8890afd35eba-swiftconf\") pod \"ca9a6b9c-6d95-4a78-9bdc-8890afd35eba\" (UID: \"ca9a6b9c-6d95-4a78-9bdc-8890afd35eba\") " Mar 07 15:13:01 crc kubenswrapper[4943]: I0307 15:13:01.801695 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ca9a6b9c-6d95-4a78-9bdc-8890afd35eba-dispersionconf\") pod \"ca9a6b9c-6d95-4a78-9bdc-8890afd35eba\" (UID: \"ca9a6b9c-6d95-4a78-9bdc-8890afd35eba\") " Mar 07 15:13:01 crc kubenswrapper[4943]: I0307 15:13:01.801743 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ca9a6b9c-6d95-4a78-9bdc-8890afd35eba-scripts\") pod \"ca9a6b9c-6d95-4a78-9bdc-8890afd35eba\" (UID: \"ca9a6b9c-6d95-4a78-9bdc-8890afd35eba\") " Mar 07 15:13:01 crc kubenswrapper[4943]: I0307 15:13:01.801970 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca9a6b9c-6d95-4a78-9bdc-8890afd35eba-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "ca9a6b9c-6d95-4a78-9bdc-8890afd35eba" (UID: "ca9a6b9c-6d95-4a78-9bdc-8890afd35eba"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:13:01 crc kubenswrapper[4943]: I0307 15:13:01.802161 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca9a6b9c-6d95-4a78-9bdc-8890afd35eba-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "ca9a6b9c-6d95-4a78-9bdc-8890afd35eba" (UID: "ca9a6b9c-6d95-4a78-9bdc-8890afd35eba"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:13:01 crc kubenswrapper[4943]: I0307 15:13:01.802237 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ca9a6b9c-6d95-4a78-9bdc-8890afd35eba-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:01 crc kubenswrapper[4943]: I0307 15:13:01.807282 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-wkvvv"] Mar 07 15:13:01 crc kubenswrapper[4943]: I0307 15:13:01.817077 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca9a6b9c-6d95-4a78-9bdc-8890afd35eba-kube-api-access-zwrls" (OuterVolumeSpecName: "kube-api-access-zwrls") pod "ca9a6b9c-6d95-4a78-9bdc-8890afd35eba" (UID: "ca9a6b9c-6d95-4a78-9bdc-8890afd35eba"). InnerVolumeSpecName "kube-api-access-zwrls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:13:01 crc kubenswrapper[4943]: I0307 15:13:01.839021 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca9a6b9c-6d95-4a78-9bdc-8890afd35eba-scripts" (OuterVolumeSpecName: "scripts") pod "ca9a6b9c-6d95-4a78-9bdc-8890afd35eba" (UID: "ca9a6b9c-6d95-4a78-9bdc-8890afd35eba"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:13:01 crc kubenswrapper[4943]: I0307 15:13:01.840250 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca9a6b9c-6d95-4a78-9bdc-8890afd35eba-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "ca9a6b9c-6d95-4a78-9bdc-8890afd35eba" (UID: "ca9a6b9c-6d95-4a78-9bdc-8890afd35eba"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:13:01 crc kubenswrapper[4943]: I0307 15:13:01.843232 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca9a6b9c-6d95-4a78-9bdc-8890afd35eba-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "ca9a6b9c-6d95-4a78-9bdc-8890afd35eba" (UID: "ca9a6b9c-6d95-4a78-9bdc-8890afd35eba"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:13:01 crc kubenswrapper[4943]: I0307 15:13:01.903816 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ca9a6b9c-6d95-4a78-9bdc-8890afd35eba-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:01 crc kubenswrapper[4943]: I0307 15:13:01.903853 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zwrls\" (UniqueName: \"kubernetes.io/projected/ca9a6b9c-6d95-4a78-9bdc-8890afd35eba-kube-api-access-zwrls\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:01 crc kubenswrapper[4943]: I0307 15:13:01.903868 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ca9a6b9c-6d95-4a78-9bdc-8890afd35eba-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:01 crc kubenswrapper[4943]: I0307 15:13:01.903878 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ca9a6b9c-6d95-4a78-9bdc-8890afd35eba-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:01 crc kubenswrapper[4943]: I0307 15:13:01.903893 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ca9a6b9c-6d95-4a78-9bdc-8890afd35eba-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:02 crc kubenswrapper[4943]: I0307 15:13:02.318446 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="31dcf8cd73e914f3a524183c2ba89213432e40dcc70c3471dd07aa67b7ece1f0" Mar 07 15:13:02 crc kubenswrapper[4943]: I0307 15:13:02.318562 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wkvvv" Mar 07 15:13:02 crc kubenswrapper[4943]: I0307 15:13:02.771850 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca9a6b9c-6d95-4a78-9bdc-8890afd35eba" path="/var/lib/kubelet/pods/ca9a6b9c-6d95-4a78-9bdc-8890afd35eba/volumes" Mar 07 15:13:03 crc kubenswrapper[4943]: I0307 15:13:03.000834 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-zhngn"] Mar 07 15:13:03 crc kubenswrapper[4943]: E0307 15:13:03.001296 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca9a6b9c-6d95-4a78-9bdc-8890afd35eba" containerName="swift-ring-rebalance" Mar 07 15:13:03 crc kubenswrapper[4943]: I0307 15:13:03.001327 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca9a6b9c-6d95-4a78-9bdc-8890afd35eba" containerName="swift-ring-rebalance" Mar 07 15:13:03 crc kubenswrapper[4943]: I0307 15:13:03.001646 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca9a6b9c-6d95-4a78-9bdc-8890afd35eba" containerName="swift-ring-rebalance" Mar 07 15:13:03 crc kubenswrapper[4943]: I0307 15:13:03.002637 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-zhngn" Mar 07 15:13:03 crc kubenswrapper[4943]: I0307 15:13:03.006370 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:13:03 crc kubenswrapper[4943]: I0307 15:13:03.007096 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:13:03 crc kubenswrapper[4943]: I0307 15:13:03.017331 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-zhngn"] Mar 07 15:13:03 crc kubenswrapper[4943]: I0307 15:13:03.126969 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e2a5a7d3-795d-42a1-8411-d3ee479944cd-dispersionconf\") pod \"swift-ring-rebalance-debug-zhngn\" (UID: \"e2a5a7d3-795d-42a1-8411-d3ee479944cd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zhngn" Mar 07 15:13:03 crc kubenswrapper[4943]: I0307 15:13:03.127380 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e2a5a7d3-795d-42a1-8411-d3ee479944cd-scripts\") pod \"swift-ring-rebalance-debug-zhngn\" (UID: \"e2a5a7d3-795d-42a1-8411-d3ee479944cd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zhngn" Mar 07 15:13:03 crc kubenswrapper[4943]: I0307 15:13:03.127645 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e2a5a7d3-795d-42a1-8411-d3ee479944cd-ring-data-devices\") pod \"swift-ring-rebalance-debug-zhngn\" (UID: \"e2a5a7d3-795d-42a1-8411-d3ee479944cd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zhngn" Mar 07 15:13:03 crc kubenswrapper[4943]: I0307 15:13:03.127760 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkbpz\" (UniqueName: \"kubernetes.io/projected/e2a5a7d3-795d-42a1-8411-d3ee479944cd-kube-api-access-pkbpz\") pod \"swift-ring-rebalance-debug-zhngn\" (UID: \"e2a5a7d3-795d-42a1-8411-d3ee479944cd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zhngn" Mar 07 15:13:03 crc kubenswrapper[4943]: I0307 15:13:03.127895 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e2a5a7d3-795d-42a1-8411-d3ee479944cd-etc-swift\") pod \"swift-ring-rebalance-debug-zhngn\" (UID: \"e2a5a7d3-795d-42a1-8411-d3ee479944cd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zhngn" Mar 07 15:13:03 crc kubenswrapper[4943]: I0307 15:13:03.127999 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e2a5a7d3-795d-42a1-8411-d3ee479944cd-swiftconf\") pod \"swift-ring-rebalance-debug-zhngn\" (UID: \"e2a5a7d3-795d-42a1-8411-d3ee479944cd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zhngn" Mar 07 15:13:03 crc kubenswrapper[4943]: I0307 15:13:03.229082 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e2a5a7d3-795d-42a1-8411-d3ee479944cd-swiftconf\") pod \"swift-ring-rebalance-debug-zhngn\" (UID: \"e2a5a7d3-795d-42a1-8411-d3ee479944cd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zhngn" Mar 07 15:13:03 crc kubenswrapper[4943]: I0307 15:13:03.229166 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e2a5a7d3-795d-42a1-8411-d3ee479944cd-dispersionconf\") pod \"swift-ring-rebalance-debug-zhngn\" (UID: \"e2a5a7d3-795d-42a1-8411-d3ee479944cd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zhngn" Mar 07 15:13:03 crc kubenswrapper[4943]: I0307 15:13:03.229199 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e2a5a7d3-795d-42a1-8411-d3ee479944cd-scripts\") pod \"swift-ring-rebalance-debug-zhngn\" (UID: \"e2a5a7d3-795d-42a1-8411-d3ee479944cd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zhngn" Mar 07 15:13:03 crc kubenswrapper[4943]: I0307 15:13:03.229285 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e2a5a7d3-795d-42a1-8411-d3ee479944cd-ring-data-devices\") pod \"swift-ring-rebalance-debug-zhngn\" (UID: \"e2a5a7d3-795d-42a1-8411-d3ee479944cd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zhngn" Mar 07 15:13:03 crc kubenswrapper[4943]: I0307 15:13:03.229312 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkbpz\" (UniqueName: \"kubernetes.io/projected/e2a5a7d3-795d-42a1-8411-d3ee479944cd-kube-api-access-pkbpz\") pod \"swift-ring-rebalance-debug-zhngn\" (UID: \"e2a5a7d3-795d-42a1-8411-d3ee479944cd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zhngn" Mar 07 15:13:03 crc kubenswrapper[4943]: I0307 15:13:03.229340 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e2a5a7d3-795d-42a1-8411-d3ee479944cd-etc-swift\") pod \"swift-ring-rebalance-debug-zhngn\" (UID: \"e2a5a7d3-795d-42a1-8411-d3ee479944cd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zhngn" Mar 07 15:13:03 crc kubenswrapper[4943]: I0307 15:13:03.229900 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e2a5a7d3-795d-42a1-8411-d3ee479944cd-etc-swift\") pod \"swift-ring-rebalance-debug-zhngn\" (UID: \"e2a5a7d3-795d-42a1-8411-d3ee479944cd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zhngn" Mar 07 15:13:03 crc kubenswrapper[4943]: I0307 15:13:03.230634 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e2a5a7d3-795d-42a1-8411-d3ee479944cd-ring-data-devices\") pod \"swift-ring-rebalance-debug-zhngn\" (UID: \"e2a5a7d3-795d-42a1-8411-d3ee479944cd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zhngn" Mar 07 15:13:03 crc kubenswrapper[4943]: I0307 15:13:03.231722 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e2a5a7d3-795d-42a1-8411-d3ee479944cd-scripts\") pod \"swift-ring-rebalance-debug-zhngn\" (UID: \"e2a5a7d3-795d-42a1-8411-d3ee479944cd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zhngn" Mar 07 15:13:03 crc kubenswrapper[4943]: I0307 15:13:03.236681 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e2a5a7d3-795d-42a1-8411-d3ee479944cd-swiftconf\") pod \"swift-ring-rebalance-debug-zhngn\" (UID: \"e2a5a7d3-795d-42a1-8411-d3ee479944cd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zhngn" Mar 07 15:13:03 crc kubenswrapper[4943]: I0307 15:13:03.239229 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e2a5a7d3-795d-42a1-8411-d3ee479944cd-dispersionconf\") pod \"swift-ring-rebalance-debug-zhngn\" (UID: \"e2a5a7d3-795d-42a1-8411-d3ee479944cd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zhngn" Mar 07 15:13:03 crc kubenswrapper[4943]: I0307 15:13:03.251110 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkbpz\" (UniqueName: \"kubernetes.io/projected/e2a5a7d3-795d-42a1-8411-d3ee479944cd-kube-api-access-pkbpz\") pod \"swift-ring-rebalance-debug-zhngn\" (UID: \"e2a5a7d3-795d-42a1-8411-d3ee479944cd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zhngn" Mar 07 15:13:03 crc kubenswrapper[4943]: I0307 15:13:03.329497 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-zhngn" Mar 07 15:13:03 crc kubenswrapper[4943]: I0307 15:13:03.613945 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-zhngn"] Mar 07 15:13:03 crc kubenswrapper[4943]: W0307 15:13:03.619474 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode2a5a7d3_795d_42a1_8411_d3ee479944cd.slice/crio-bc97f7c134503714f5edb4cdc9c4e01d5b140a994fb4c1946db5caf32f83c95d WatchSource:0}: Error finding container bc97f7c134503714f5edb4cdc9c4e01d5b140a994fb4c1946db5caf32f83c95d: Status 404 returned error can't find the container with id bc97f7c134503714f5edb4cdc9c4e01d5b140a994fb4c1946db5caf32f83c95d Mar 07 15:13:04 crc kubenswrapper[4943]: I0307 15:13:04.353541 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-zhngn" event={"ID":"e2a5a7d3-795d-42a1-8411-d3ee479944cd","Type":"ContainerStarted","Data":"02739a3a73e71193174188d25806e2c39d3e53193138b2878dd656daca6cff81"} Mar 07 15:13:04 crc kubenswrapper[4943]: I0307 15:13:04.353611 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-zhngn" event={"ID":"e2a5a7d3-795d-42a1-8411-d3ee479944cd","Type":"ContainerStarted","Data":"bc97f7c134503714f5edb4cdc9c4e01d5b140a994fb4c1946db5caf32f83c95d"} Mar 07 15:13:04 crc kubenswrapper[4943]: I0307 15:13:04.392681 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-zhngn" podStartSLOduration=2.392653886 podStartE2EDuration="2.392653886s" podCreationTimestamp="2026-03-07 15:13:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:13:04.381431151 +0000 UTC m=+2026.333567689" watchObservedRunningTime="2026-03-07 15:13:04.392653886 +0000 UTC m=+2026.344790414" Mar 07 15:13:05 crc kubenswrapper[4943]: I0307 15:13:05.363525 4943 generic.go:334] "Generic (PLEG): container finished" podID="e2a5a7d3-795d-42a1-8411-d3ee479944cd" containerID="02739a3a73e71193174188d25806e2c39d3e53193138b2878dd656daca6cff81" exitCode=0 Mar 07 15:13:05 crc kubenswrapper[4943]: I0307 15:13:05.363565 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-zhngn" event={"ID":"e2a5a7d3-795d-42a1-8411-d3ee479944cd","Type":"ContainerDied","Data":"02739a3a73e71193174188d25806e2c39d3e53193138b2878dd656daca6cff81"} Mar 07 15:13:06 crc kubenswrapper[4943]: I0307 15:13:06.074323 4943 patch_prober.go:28] interesting pod/machine-config-daemon-cpdmm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 07 15:13:06 crc kubenswrapper[4943]: I0307 15:13:06.074420 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 07 15:13:06 crc kubenswrapper[4943]: I0307 15:13:06.682503 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-zhngn" Mar 07 15:13:06 crc kubenswrapper[4943]: I0307 15:13:06.710648 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-zhngn"] Mar 07 15:13:06 crc kubenswrapper[4943]: I0307 15:13:06.717823 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-zhngn"] Mar 07 15:13:06 crc kubenswrapper[4943]: I0307 15:13:06.793486 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e2a5a7d3-795d-42a1-8411-d3ee479944cd-etc-swift\") pod \"e2a5a7d3-795d-42a1-8411-d3ee479944cd\" (UID: \"e2a5a7d3-795d-42a1-8411-d3ee479944cd\") " Mar 07 15:13:06 crc kubenswrapper[4943]: I0307 15:13:06.793556 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e2a5a7d3-795d-42a1-8411-d3ee479944cd-swiftconf\") pod \"e2a5a7d3-795d-42a1-8411-d3ee479944cd\" (UID: \"e2a5a7d3-795d-42a1-8411-d3ee479944cd\") " Mar 07 15:13:06 crc kubenswrapper[4943]: I0307 15:13:06.793592 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e2a5a7d3-795d-42a1-8411-d3ee479944cd-dispersionconf\") pod \"e2a5a7d3-795d-42a1-8411-d3ee479944cd\" (UID: \"e2a5a7d3-795d-42a1-8411-d3ee479944cd\") " Mar 07 15:13:06 crc kubenswrapper[4943]: I0307 15:13:06.793682 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e2a5a7d3-795d-42a1-8411-d3ee479944cd-ring-data-devices\") pod \"e2a5a7d3-795d-42a1-8411-d3ee479944cd\" (UID: \"e2a5a7d3-795d-42a1-8411-d3ee479944cd\") " Mar 07 15:13:06 crc kubenswrapper[4943]: I0307 15:13:06.793867 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e2a5a7d3-795d-42a1-8411-d3ee479944cd-scripts\") pod \"e2a5a7d3-795d-42a1-8411-d3ee479944cd\" (UID: \"e2a5a7d3-795d-42a1-8411-d3ee479944cd\") " Mar 07 15:13:06 crc kubenswrapper[4943]: I0307 15:13:06.793911 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pkbpz\" (UniqueName: \"kubernetes.io/projected/e2a5a7d3-795d-42a1-8411-d3ee479944cd-kube-api-access-pkbpz\") pod \"e2a5a7d3-795d-42a1-8411-d3ee479944cd\" (UID: \"e2a5a7d3-795d-42a1-8411-d3ee479944cd\") " Mar 07 15:13:06 crc kubenswrapper[4943]: I0307 15:13:06.794493 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2a5a7d3-795d-42a1-8411-d3ee479944cd-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "e2a5a7d3-795d-42a1-8411-d3ee479944cd" (UID: "e2a5a7d3-795d-42a1-8411-d3ee479944cd"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:13:06 crc kubenswrapper[4943]: I0307 15:13:06.795119 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e2a5a7d3-795d-42a1-8411-d3ee479944cd-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "e2a5a7d3-795d-42a1-8411-d3ee479944cd" (UID: "e2a5a7d3-795d-42a1-8411-d3ee479944cd"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:13:06 crc kubenswrapper[4943]: I0307 15:13:06.803183 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2a5a7d3-795d-42a1-8411-d3ee479944cd-kube-api-access-pkbpz" (OuterVolumeSpecName: "kube-api-access-pkbpz") pod "e2a5a7d3-795d-42a1-8411-d3ee479944cd" (UID: "e2a5a7d3-795d-42a1-8411-d3ee479944cd"). InnerVolumeSpecName "kube-api-access-pkbpz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:13:06 crc kubenswrapper[4943]: I0307 15:13:06.812640 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2a5a7d3-795d-42a1-8411-d3ee479944cd-scripts" (OuterVolumeSpecName: "scripts") pod "e2a5a7d3-795d-42a1-8411-d3ee479944cd" (UID: "e2a5a7d3-795d-42a1-8411-d3ee479944cd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:13:06 crc kubenswrapper[4943]: I0307 15:13:06.814805 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2a5a7d3-795d-42a1-8411-d3ee479944cd-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "e2a5a7d3-795d-42a1-8411-d3ee479944cd" (UID: "e2a5a7d3-795d-42a1-8411-d3ee479944cd"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:13:06 crc kubenswrapper[4943]: I0307 15:13:06.831356 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2a5a7d3-795d-42a1-8411-d3ee479944cd-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "e2a5a7d3-795d-42a1-8411-d3ee479944cd" (UID: "e2a5a7d3-795d-42a1-8411-d3ee479944cd"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:13:06 crc kubenswrapper[4943]: I0307 15:13:06.896216 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e2a5a7d3-795d-42a1-8411-d3ee479944cd-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:06 crc kubenswrapper[4943]: I0307 15:13:06.896244 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e2a5a7d3-795d-42a1-8411-d3ee479944cd-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:06 crc kubenswrapper[4943]: I0307 15:13:06.896274 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pkbpz\" (UniqueName: \"kubernetes.io/projected/e2a5a7d3-795d-42a1-8411-d3ee479944cd-kube-api-access-pkbpz\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:06 crc kubenswrapper[4943]: I0307 15:13:06.896285 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e2a5a7d3-795d-42a1-8411-d3ee479944cd-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:06 crc kubenswrapper[4943]: I0307 15:13:06.896293 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e2a5a7d3-795d-42a1-8411-d3ee479944cd-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:06 crc kubenswrapper[4943]: I0307 15:13:06.896301 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e2a5a7d3-795d-42a1-8411-d3ee479944cd-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:07 crc kubenswrapper[4943]: I0307 15:13:07.389382 4943 scope.go:117] "RemoveContainer" containerID="02739a3a73e71193174188d25806e2c39d3e53193138b2878dd656daca6cff81" Mar 07 15:13:07 crc kubenswrapper[4943]: I0307 15:13:07.389422 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-zhngn" Mar 07 15:13:07 crc kubenswrapper[4943]: I0307 15:13:07.914689 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-cxt2f"] Mar 07 15:13:07 crc kubenswrapper[4943]: E0307 15:13:07.915166 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2a5a7d3-795d-42a1-8411-d3ee479944cd" containerName="swift-ring-rebalance" Mar 07 15:13:07 crc kubenswrapper[4943]: I0307 15:13:07.915188 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2a5a7d3-795d-42a1-8411-d3ee479944cd" containerName="swift-ring-rebalance" Mar 07 15:13:07 crc kubenswrapper[4943]: I0307 15:13:07.915479 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2a5a7d3-795d-42a1-8411-d3ee479944cd" containerName="swift-ring-rebalance" Mar 07 15:13:07 crc kubenswrapper[4943]: I0307 15:13:07.916255 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-cxt2f" Mar 07 15:13:07 crc kubenswrapper[4943]: I0307 15:13:07.953643 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:13:07 crc kubenswrapper[4943]: I0307 15:13:07.954233 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:13:07 crc kubenswrapper[4943]: I0307 15:13:07.960379 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-cxt2f"] Mar 07 15:13:08 crc kubenswrapper[4943]: I0307 15:13:08.015838 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/70153fd2-21a3-4895-be3a-7fef7e17f97c-swiftconf\") pod \"swift-ring-rebalance-debug-cxt2f\" (UID: \"70153fd2-21a3-4895-be3a-7fef7e17f97c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-cxt2f" Mar 07 15:13:08 crc kubenswrapper[4943]: I0307 15:13:08.015902 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/70153fd2-21a3-4895-be3a-7fef7e17f97c-dispersionconf\") pod \"swift-ring-rebalance-debug-cxt2f\" (UID: \"70153fd2-21a3-4895-be3a-7fef7e17f97c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-cxt2f" Mar 07 15:13:08 crc kubenswrapper[4943]: I0307 15:13:08.016011 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-496gb\" (UniqueName: \"kubernetes.io/projected/70153fd2-21a3-4895-be3a-7fef7e17f97c-kube-api-access-496gb\") pod \"swift-ring-rebalance-debug-cxt2f\" (UID: \"70153fd2-21a3-4895-be3a-7fef7e17f97c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-cxt2f" Mar 07 15:13:08 crc kubenswrapper[4943]: I0307 15:13:08.016216 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/70153fd2-21a3-4895-be3a-7fef7e17f97c-etc-swift\") pod \"swift-ring-rebalance-debug-cxt2f\" (UID: \"70153fd2-21a3-4895-be3a-7fef7e17f97c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-cxt2f" Mar 07 15:13:08 crc kubenswrapper[4943]: I0307 15:13:08.016440 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/70153fd2-21a3-4895-be3a-7fef7e17f97c-ring-data-devices\") pod \"swift-ring-rebalance-debug-cxt2f\" (UID: \"70153fd2-21a3-4895-be3a-7fef7e17f97c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-cxt2f" Mar 07 15:13:08 crc kubenswrapper[4943]: I0307 15:13:08.016594 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/70153fd2-21a3-4895-be3a-7fef7e17f97c-scripts\") pod \"swift-ring-rebalance-debug-cxt2f\" (UID: \"70153fd2-21a3-4895-be3a-7fef7e17f97c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-cxt2f" Mar 07 15:13:08 crc kubenswrapper[4943]: I0307 15:13:08.118109 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/70153fd2-21a3-4895-be3a-7fef7e17f97c-swiftconf\") pod \"swift-ring-rebalance-debug-cxt2f\" (UID: \"70153fd2-21a3-4895-be3a-7fef7e17f97c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-cxt2f" Mar 07 15:13:08 crc kubenswrapper[4943]: I0307 15:13:08.118697 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/70153fd2-21a3-4895-be3a-7fef7e17f97c-dispersionconf\") pod \"swift-ring-rebalance-debug-cxt2f\" (UID: \"70153fd2-21a3-4895-be3a-7fef7e17f97c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-cxt2f" Mar 07 15:13:08 crc kubenswrapper[4943]: I0307 15:13:08.118863 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-496gb\" (UniqueName: \"kubernetes.io/projected/70153fd2-21a3-4895-be3a-7fef7e17f97c-kube-api-access-496gb\") pod \"swift-ring-rebalance-debug-cxt2f\" (UID: \"70153fd2-21a3-4895-be3a-7fef7e17f97c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-cxt2f" Mar 07 15:13:08 crc kubenswrapper[4943]: I0307 15:13:08.118966 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/70153fd2-21a3-4895-be3a-7fef7e17f97c-etc-swift\") pod \"swift-ring-rebalance-debug-cxt2f\" (UID: \"70153fd2-21a3-4895-be3a-7fef7e17f97c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-cxt2f" Mar 07 15:13:08 crc kubenswrapper[4943]: I0307 15:13:08.119077 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/70153fd2-21a3-4895-be3a-7fef7e17f97c-ring-data-devices\") pod \"swift-ring-rebalance-debug-cxt2f\" (UID: \"70153fd2-21a3-4895-be3a-7fef7e17f97c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-cxt2f" Mar 07 15:13:08 crc kubenswrapper[4943]: I0307 15:13:08.119160 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/70153fd2-21a3-4895-be3a-7fef7e17f97c-scripts\") pod \"swift-ring-rebalance-debug-cxt2f\" (UID: \"70153fd2-21a3-4895-be3a-7fef7e17f97c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-cxt2f" Mar 07 15:13:08 crc kubenswrapper[4943]: I0307 15:13:08.119685 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/70153fd2-21a3-4895-be3a-7fef7e17f97c-etc-swift\") pod \"swift-ring-rebalance-debug-cxt2f\" (UID: \"70153fd2-21a3-4895-be3a-7fef7e17f97c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-cxt2f" Mar 07 15:13:08 crc kubenswrapper[4943]: I0307 15:13:08.120654 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/70153fd2-21a3-4895-be3a-7fef7e17f97c-scripts\") pod \"swift-ring-rebalance-debug-cxt2f\" (UID: \"70153fd2-21a3-4895-be3a-7fef7e17f97c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-cxt2f" Mar 07 15:13:08 crc kubenswrapper[4943]: I0307 15:13:08.120733 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/70153fd2-21a3-4895-be3a-7fef7e17f97c-ring-data-devices\") pod \"swift-ring-rebalance-debug-cxt2f\" (UID: \"70153fd2-21a3-4895-be3a-7fef7e17f97c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-cxt2f" Mar 07 15:13:08 crc kubenswrapper[4943]: I0307 15:13:08.124033 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/70153fd2-21a3-4895-be3a-7fef7e17f97c-swiftconf\") pod \"swift-ring-rebalance-debug-cxt2f\" (UID: \"70153fd2-21a3-4895-be3a-7fef7e17f97c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-cxt2f" Mar 07 15:13:08 crc kubenswrapper[4943]: I0307 15:13:08.125689 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/70153fd2-21a3-4895-be3a-7fef7e17f97c-dispersionconf\") pod \"swift-ring-rebalance-debug-cxt2f\" (UID: \"70153fd2-21a3-4895-be3a-7fef7e17f97c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-cxt2f" Mar 07 15:13:08 crc kubenswrapper[4943]: I0307 15:13:08.142225 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-496gb\" (UniqueName: \"kubernetes.io/projected/70153fd2-21a3-4895-be3a-7fef7e17f97c-kube-api-access-496gb\") pod \"swift-ring-rebalance-debug-cxt2f\" (UID: \"70153fd2-21a3-4895-be3a-7fef7e17f97c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-cxt2f" Mar 07 15:13:08 crc kubenswrapper[4943]: I0307 15:13:08.279515 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-cxt2f" Mar 07 15:13:08 crc kubenswrapper[4943]: I0307 15:13:08.554635 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-cxt2f"] Mar 07 15:13:08 crc kubenswrapper[4943]: I0307 15:13:08.789407 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2a5a7d3-795d-42a1-8411-d3ee479944cd" path="/var/lib/kubelet/pods/e2a5a7d3-795d-42a1-8411-d3ee479944cd/volumes" Mar 07 15:13:09 crc kubenswrapper[4943]: I0307 15:13:09.410888 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-cxt2f" event={"ID":"70153fd2-21a3-4895-be3a-7fef7e17f97c","Type":"ContainerStarted","Data":"bd8ffc2cc6457280f506fef1ee2317694d39880dd2efc16a67fc94cb69e6f936"} Mar 07 15:13:09 crc kubenswrapper[4943]: I0307 15:13:09.411172 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-cxt2f" event={"ID":"70153fd2-21a3-4895-be3a-7fef7e17f97c","Type":"ContainerStarted","Data":"1d48727adf6c6cbe1ef255b56c2bd7007147d0936ae254631f5fdbc617f7d804"} Mar 07 15:13:09 crc kubenswrapper[4943]: I0307 15:13:09.435131 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-cxt2f" podStartSLOduration=2.435093413 podStartE2EDuration="2.435093413s" podCreationTimestamp="2026-03-07 15:13:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:13:09.429499136 +0000 UTC m=+2031.381635674" watchObservedRunningTime="2026-03-07 15:13:09.435093413 +0000 UTC m=+2031.387229951" Mar 07 15:13:10 crc kubenswrapper[4943]: I0307 15:13:10.421536 4943 generic.go:334] "Generic (PLEG): container finished" podID="70153fd2-21a3-4895-be3a-7fef7e17f97c" containerID="bd8ffc2cc6457280f506fef1ee2317694d39880dd2efc16a67fc94cb69e6f936" exitCode=0 Mar 07 15:13:10 crc kubenswrapper[4943]: I0307 15:13:10.421593 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-cxt2f" event={"ID":"70153fd2-21a3-4895-be3a-7fef7e17f97c","Type":"ContainerDied","Data":"bd8ffc2cc6457280f506fef1ee2317694d39880dd2efc16a67fc94cb69e6f936"} Mar 07 15:13:11 crc kubenswrapper[4943]: I0307 15:13:11.850858 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-cxt2f" Mar 07 15:13:11 crc kubenswrapper[4943]: I0307 15:13:11.888437 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-cxt2f"] Mar 07 15:13:11 crc kubenswrapper[4943]: I0307 15:13:11.896553 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-cxt2f"] Mar 07 15:13:11 crc kubenswrapper[4943]: I0307 15:13:11.994010 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/70153fd2-21a3-4895-be3a-7fef7e17f97c-scripts\") pod \"70153fd2-21a3-4895-be3a-7fef7e17f97c\" (UID: \"70153fd2-21a3-4895-be3a-7fef7e17f97c\") " Mar 07 15:13:11 crc kubenswrapper[4943]: I0307 15:13:11.994162 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/70153fd2-21a3-4895-be3a-7fef7e17f97c-etc-swift\") pod \"70153fd2-21a3-4895-be3a-7fef7e17f97c\" (UID: \"70153fd2-21a3-4895-be3a-7fef7e17f97c\") " Mar 07 15:13:11 crc kubenswrapper[4943]: I0307 15:13:11.994198 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/70153fd2-21a3-4895-be3a-7fef7e17f97c-ring-data-devices\") pod \"70153fd2-21a3-4895-be3a-7fef7e17f97c\" (UID: \"70153fd2-21a3-4895-be3a-7fef7e17f97c\") " Mar 07 15:13:11 crc kubenswrapper[4943]: I0307 15:13:11.994239 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-496gb\" (UniqueName: \"kubernetes.io/projected/70153fd2-21a3-4895-be3a-7fef7e17f97c-kube-api-access-496gb\") pod \"70153fd2-21a3-4895-be3a-7fef7e17f97c\" (UID: \"70153fd2-21a3-4895-be3a-7fef7e17f97c\") " Mar 07 15:13:11 crc kubenswrapper[4943]: I0307 15:13:11.994326 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/70153fd2-21a3-4895-be3a-7fef7e17f97c-dispersionconf\") pod \"70153fd2-21a3-4895-be3a-7fef7e17f97c\" (UID: \"70153fd2-21a3-4895-be3a-7fef7e17f97c\") " Mar 07 15:13:11 crc kubenswrapper[4943]: I0307 15:13:11.994386 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/70153fd2-21a3-4895-be3a-7fef7e17f97c-swiftconf\") pod \"70153fd2-21a3-4895-be3a-7fef7e17f97c\" (UID: \"70153fd2-21a3-4895-be3a-7fef7e17f97c\") " Mar 07 15:13:11 crc kubenswrapper[4943]: I0307 15:13:11.995445 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70153fd2-21a3-4895-be3a-7fef7e17f97c-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "70153fd2-21a3-4895-be3a-7fef7e17f97c" (UID: "70153fd2-21a3-4895-be3a-7fef7e17f97c"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:13:11 crc kubenswrapper[4943]: I0307 15:13:11.996683 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/70153fd2-21a3-4895-be3a-7fef7e17f97c-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:11 crc kubenswrapper[4943]: I0307 15:13:11.997067 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/70153fd2-21a3-4895-be3a-7fef7e17f97c-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "70153fd2-21a3-4895-be3a-7fef7e17f97c" (UID: "70153fd2-21a3-4895-be3a-7fef7e17f97c"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:13:12 crc kubenswrapper[4943]: I0307 15:13:12.002152 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70153fd2-21a3-4895-be3a-7fef7e17f97c-kube-api-access-496gb" (OuterVolumeSpecName: "kube-api-access-496gb") pod "70153fd2-21a3-4895-be3a-7fef7e17f97c" (UID: "70153fd2-21a3-4895-be3a-7fef7e17f97c"). InnerVolumeSpecName "kube-api-access-496gb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:13:12 crc kubenswrapper[4943]: I0307 15:13:12.016824 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70153fd2-21a3-4895-be3a-7fef7e17f97c-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "70153fd2-21a3-4895-be3a-7fef7e17f97c" (UID: "70153fd2-21a3-4895-be3a-7fef7e17f97c"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:13:12 crc kubenswrapper[4943]: I0307 15:13:12.032792 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70153fd2-21a3-4895-be3a-7fef7e17f97c-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "70153fd2-21a3-4895-be3a-7fef7e17f97c" (UID: "70153fd2-21a3-4895-be3a-7fef7e17f97c"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:13:12 crc kubenswrapper[4943]: I0307 15:13:12.036490 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/70153fd2-21a3-4895-be3a-7fef7e17f97c-scripts" (OuterVolumeSpecName: "scripts") pod "70153fd2-21a3-4895-be3a-7fef7e17f97c" (UID: "70153fd2-21a3-4895-be3a-7fef7e17f97c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:13:12 crc kubenswrapper[4943]: I0307 15:13:12.098336 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/70153fd2-21a3-4895-be3a-7fef7e17f97c-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:12 crc kubenswrapper[4943]: I0307 15:13:12.098379 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/70153fd2-21a3-4895-be3a-7fef7e17f97c-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:12 crc kubenswrapper[4943]: I0307 15:13:12.098396 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-496gb\" (UniqueName: \"kubernetes.io/projected/70153fd2-21a3-4895-be3a-7fef7e17f97c-kube-api-access-496gb\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:12 crc kubenswrapper[4943]: I0307 15:13:12.098408 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/70153fd2-21a3-4895-be3a-7fef7e17f97c-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:12 crc kubenswrapper[4943]: I0307 15:13:12.098420 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/70153fd2-21a3-4895-be3a-7fef7e17f97c-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:12 crc kubenswrapper[4943]: I0307 15:13:12.441792 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1d48727adf6c6cbe1ef255b56c2bd7007147d0936ae254631f5fdbc617f7d804" Mar 07 15:13:12 crc kubenswrapper[4943]: I0307 15:13:12.441831 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-cxt2f" Mar 07 15:13:12 crc kubenswrapper[4943]: I0307 15:13:12.769644 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70153fd2-21a3-4895-be3a-7fef7e17f97c" path="/var/lib/kubelet/pods/70153fd2-21a3-4895-be3a-7fef7e17f97c/volumes" Mar 07 15:13:13 crc kubenswrapper[4943]: I0307 15:13:13.055958 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-zhk9t"] Mar 07 15:13:13 crc kubenswrapper[4943]: E0307 15:13:13.056416 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70153fd2-21a3-4895-be3a-7fef7e17f97c" containerName="swift-ring-rebalance" Mar 07 15:13:13 crc kubenswrapper[4943]: I0307 15:13:13.056437 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="70153fd2-21a3-4895-be3a-7fef7e17f97c" containerName="swift-ring-rebalance" Mar 07 15:13:13 crc kubenswrapper[4943]: I0307 15:13:13.056701 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="70153fd2-21a3-4895-be3a-7fef7e17f97c" containerName="swift-ring-rebalance" Mar 07 15:13:13 crc kubenswrapper[4943]: I0307 15:13:13.057463 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-zhk9t" Mar 07 15:13:13 crc kubenswrapper[4943]: I0307 15:13:13.060428 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:13:13 crc kubenswrapper[4943]: I0307 15:13:13.061593 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:13:13 crc kubenswrapper[4943]: I0307 15:13:13.077060 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-zhk9t"] Mar 07 15:13:13 crc kubenswrapper[4943]: I0307 15:13:13.216040 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhk46\" (UniqueName: \"kubernetes.io/projected/c45b3cea-474a-437a-8e9a-2fc96e1d7200-kube-api-access-zhk46\") pod \"swift-ring-rebalance-debug-zhk9t\" (UID: \"c45b3cea-474a-437a-8e9a-2fc96e1d7200\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zhk9t" Mar 07 15:13:13 crc kubenswrapper[4943]: I0307 15:13:13.216236 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c45b3cea-474a-437a-8e9a-2fc96e1d7200-ring-data-devices\") pod \"swift-ring-rebalance-debug-zhk9t\" (UID: \"c45b3cea-474a-437a-8e9a-2fc96e1d7200\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zhk9t" Mar 07 15:13:13 crc kubenswrapper[4943]: I0307 15:13:13.216296 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c45b3cea-474a-437a-8e9a-2fc96e1d7200-swiftconf\") pod \"swift-ring-rebalance-debug-zhk9t\" (UID: \"c45b3cea-474a-437a-8e9a-2fc96e1d7200\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zhk9t" Mar 07 15:13:13 crc kubenswrapper[4943]: I0307 15:13:13.216410 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c45b3cea-474a-437a-8e9a-2fc96e1d7200-scripts\") pod \"swift-ring-rebalance-debug-zhk9t\" (UID: \"c45b3cea-474a-437a-8e9a-2fc96e1d7200\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zhk9t" Mar 07 15:13:13 crc kubenswrapper[4943]: I0307 15:13:13.216476 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c45b3cea-474a-437a-8e9a-2fc96e1d7200-dispersionconf\") pod \"swift-ring-rebalance-debug-zhk9t\" (UID: \"c45b3cea-474a-437a-8e9a-2fc96e1d7200\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zhk9t" Mar 07 15:13:13 crc kubenswrapper[4943]: I0307 15:13:13.216632 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c45b3cea-474a-437a-8e9a-2fc96e1d7200-etc-swift\") pod \"swift-ring-rebalance-debug-zhk9t\" (UID: \"c45b3cea-474a-437a-8e9a-2fc96e1d7200\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zhk9t" Mar 07 15:13:13 crc kubenswrapper[4943]: I0307 15:13:13.318397 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhk46\" (UniqueName: \"kubernetes.io/projected/c45b3cea-474a-437a-8e9a-2fc96e1d7200-kube-api-access-zhk46\") pod \"swift-ring-rebalance-debug-zhk9t\" (UID: \"c45b3cea-474a-437a-8e9a-2fc96e1d7200\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zhk9t" Mar 07 15:13:13 crc kubenswrapper[4943]: I0307 15:13:13.318845 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c45b3cea-474a-437a-8e9a-2fc96e1d7200-ring-data-devices\") pod \"swift-ring-rebalance-debug-zhk9t\" (UID: \"c45b3cea-474a-437a-8e9a-2fc96e1d7200\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zhk9t" Mar 07 15:13:13 crc kubenswrapper[4943]: I0307 15:13:13.318919 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c45b3cea-474a-437a-8e9a-2fc96e1d7200-swiftconf\") pod \"swift-ring-rebalance-debug-zhk9t\" (UID: \"c45b3cea-474a-437a-8e9a-2fc96e1d7200\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zhk9t" Mar 07 15:13:13 crc kubenswrapper[4943]: I0307 15:13:13.319016 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c45b3cea-474a-437a-8e9a-2fc96e1d7200-scripts\") pod \"swift-ring-rebalance-debug-zhk9t\" (UID: \"c45b3cea-474a-437a-8e9a-2fc96e1d7200\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zhk9t" Mar 07 15:13:13 crc kubenswrapper[4943]: I0307 15:13:13.319056 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c45b3cea-474a-437a-8e9a-2fc96e1d7200-dispersionconf\") pod \"swift-ring-rebalance-debug-zhk9t\" (UID: \"c45b3cea-474a-437a-8e9a-2fc96e1d7200\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zhk9t" Mar 07 15:13:13 crc kubenswrapper[4943]: I0307 15:13:13.319148 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c45b3cea-474a-437a-8e9a-2fc96e1d7200-etc-swift\") pod \"swift-ring-rebalance-debug-zhk9t\" (UID: \"c45b3cea-474a-437a-8e9a-2fc96e1d7200\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zhk9t" Mar 07 15:13:13 crc kubenswrapper[4943]: I0307 15:13:13.319821 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c45b3cea-474a-437a-8e9a-2fc96e1d7200-etc-swift\") pod \"swift-ring-rebalance-debug-zhk9t\" (UID: \"c45b3cea-474a-437a-8e9a-2fc96e1d7200\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zhk9t" Mar 07 15:13:13 crc kubenswrapper[4943]: I0307 15:13:13.320090 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c45b3cea-474a-437a-8e9a-2fc96e1d7200-ring-data-devices\") pod \"swift-ring-rebalance-debug-zhk9t\" (UID: \"c45b3cea-474a-437a-8e9a-2fc96e1d7200\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zhk9t" Mar 07 15:13:13 crc kubenswrapper[4943]: I0307 15:13:13.320322 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c45b3cea-474a-437a-8e9a-2fc96e1d7200-scripts\") pod \"swift-ring-rebalance-debug-zhk9t\" (UID: \"c45b3cea-474a-437a-8e9a-2fc96e1d7200\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zhk9t" Mar 07 15:13:13 crc kubenswrapper[4943]: I0307 15:13:13.329704 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c45b3cea-474a-437a-8e9a-2fc96e1d7200-swiftconf\") pod \"swift-ring-rebalance-debug-zhk9t\" (UID: \"c45b3cea-474a-437a-8e9a-2fc96e1d7200\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zhk9t" Mar 07 15:13:13 crc kubenswrapper[4943]: I0307 15:13:13.330176 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c45b3cea-474a-437a-8e9a-2fc96e1d7200-dispersionconf\") pod \"swift-ring-rebalance-debug-zhk9t\" (UID: \"c45b3cea-474a-437a-8e9a-2fc96e1d7200\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zhk9t" Mar 07 15:13:13 crc kubenswrapper[4943]: I0307 15:13:13.337249 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhk46\" (UniqueName: \"kubernetes.io/projected/c45b3cea-474a-437a-8e9a-2fc96e1d7200-kube-api-access-zhk46\") pod \"swift-ring-rebalance-debug-zhk9t\" (UID: \"c45b3cea-474a-437a-8e9a-2fc96e1d7200\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zhk9t" Mar 07 15:13:13 crc kubenswrapper[4943]: I0307 15:13:13.397155 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-zhk9t" Mar 07 15:13:13 crc kubenswrapper[4943]: I0307 15:13:13.949221 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-zhk9t"] Mar 07 15:13:14 crc kubenswrapper[4943]: I0307 15:13:14.468588 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-zhk9t" event={"ID":"c45b3cea-474a-437a-8e9a-2fc96e1d7200","Type":"ContainerStarted","Data":"7ab86a755eb914c5752287227d41b72269aced522ad634d946657faf7f98e060"} Mar 07 15:13:14 crc kubenswrapper[4943]: I0307 15:13:14.468651 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-zhk9t" event={"ID":"c45b3cea-474a-437a-8e9a-2fc96e1d7200","Type":"ContainerStarted","Data":"03b37b9c2a686a5638dc9f4289a40fca025ede7c21bc09a66ebed58524933ab7"} Mar 07 15:13:14 crc kubenswrapper[4943]: I0307 15:13:14.526158 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-zhk9t" podStartSLOduration=1.526139122 podStartE2EDuration="1.526139122s" podCreationTimestamp="2026-03-07 15:13:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:13:14.508202532 +0000 UTC m=+2036.460339070" watchObservedRunningTime="2026-03-07 15:13:14.526139122 +0000 UTC m=+2036.478275610" Mar 07 15:13:16 crc kubenswrapper[4943]: I0307 15:13:16.490321 4943 generic.go:334] "Generic (PLEG): container finished" podID="c45b3cea-474a-437a-8e9a-2fc96e1d7200" containerID="7ab86a755eb914c5752287227d41b72269aced522ad634d946657faf7f98e060" exitCode=0 Mar 07 15:13:16 crc kubenswrapper[4943]: I0307 15:13:16.490387 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-zhk9t" event={"ID":"c45b3cea-474a-437a-8e9a-2fc96e1d7200","Type":"ContainerDied","Data":"7ab86a755eb914c5752287227d41b72269aced522ad634d946657faf7f98e060"} Mar 07 15:13:17 crc kubenswrapper[4943]: I0307 15:13:17.894126 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-zhk9t" Mar 07 15:13:17 crc kubenswrapper[4943]: I0307 15:13:17.947869 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-zhk9t"] Mar 07 15:13:17 crc kubenswrapper[4943]: I0307 15:13:17.957703 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-zhk9t"] Mar 07 15:13:18 crc kubenswrapper[4943]: I0307 15:13:18.012033 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c45b3cea-474a-437a-8e9a-2fc96e1d7200-ring-data-devices\") pod \"c45b3cea-474a-437a-8e9a-2fc96e1d7200\" (UID: \"c45b3cea-474a-437a-8e9a-2fc96e1d7200\") " Mar 07 15:13:18 crc kubenswrapper[4943]: I0307 15:13:18.012084 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c45b3cea-474a-437a-8e9a-2fc96e1d7200-scripts\") pod \"c45b3cea-474a-437a-8e9a-2fc96e1d7200\" (UID: \"c45b3cea-474a-437a-8e9a-2fc96e1d7200\") " Mar 07 15:13:18 crc kubenswrapper[4943]: I0307 15:13:18.012106 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c45b3cea-474a-437a-8e9a-2fc96e1d7200-etc-swift\") pod \"c45b3cea-474a-437a-8e9a-2fc96e1d7200\" (UID: \"c45b3cea-474a-437a-8e9a-2fc96e1d7200\") " Mar 07 15:13:18 crc kubenswrapper[4943]: I0307 15:13:18.012182 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c45b3cea-474a-437a-8e9a-2fc96e1d7200-swiftconf\") pod \"c45b3cea-474a-437a-8e9a-2fc96e1d7200\" (UID: \"c45b3cea-474a-437a-8e9a-2fc96e1d7200\") " Mar 07 15:13:18 crc kubenswrapper[4943]: I0307 15:13:18.012224 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zhk46\" (UniqueName: \"kubernetes.io/projected/c45b3cea-474a-437a-8e9a-2fc96e1d7200-kube-api-access-zhk46\") pod \"c45b3cea-474a-437a-8e9a-2fc96e1d7200\" (UID: \"c45b3cea-474a-437a-8e9a-2fc96e1d7200\") " Mar 07 15:13:18 crc kubenswrapper[4943]: I0307 15:13:18.012240 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c45b3cea-474a-437a-8e9a-2fc96e1d7200-dispersionconf\") pod \"c45b3cea-474a-437a-8e9a-2fc96e1d7200\" (UID: \"c45b3cea-474a-437a-8e9a-2fc96e1d7200\") " Mar 07 15:13:18 crc kubenswrapper[4943]: I0307 15:13:18.013028 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c45b3cea-474a-437a-8e9a-2fc96e1d7200-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "c45b3cea-474a-437a-8e9a-2fc96e1d7200" (UID: "c45b3cea-474a-437a-8e9a-2fc96e1d7200"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:13:18 crc kubenswrapper[4943]: I0307 15:13:18.013403 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c45b3cea-474a-437a-8e9a-2fc96e1d7200-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "c45b3cea-474a-437a-8e9a-2fc96e1d7200" (UID: "c45b3cea-474a-437a-8e9a-2fc96e1d7200"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:13:18 crc kubenswrapper[4943]: I0307 15:13:18.019595 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c45b3cea-474a-437a-8e9a-2fc96e1d7200-kube-api-access-zhk46" (OuterVolumeSpecName: "kube-api-access-zhk46") pod "c45b3cea-474a-437a-8e9a-2fc96e1d7200" (UID: "c45b3cea-474a-437a-8e9a-2fc96e1d7200"). InnerVolumeSpecName "kube-api-access-zhk46". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:13:18 crc kubenswrapper[4943]: I0307 15:13:18.032384 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c45b3cea-474a-437a-8e9a-2fc96e1d7200-scripts" (OuterVolumeSpecName: "scripts") pod "c45b3cea-474a-437a-8e9a-2fc96e1d7200" (UID: "c45b3cea-474a-437a-8e9a-2fc96e1d7200"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:13:18 crc kubenswrapper[4943]: I0307 15:13:18.043859 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c45b3cea-474a-437a-8e9a-2fc96e1d7200-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "c45b3cea-474a-437a-8e9a-2fc96e1d7200" (UID: "c45b3cea-474a-437a-8e9a-2fc96e1d7200"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:13:18 crc kubenswrapper[4943]: I0307 15:13:18.050484 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c45b3cea-474a-437a-8e9a-2fc96e1d7200-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "c45b3cea-474a-437a-8e9a-2fc96e1d7200" (UID: "c45b3cea-474a-437a-8e9a-2fc96e1d7200"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:13:18 crc kubenswrapper[4943]: I0307 15:13:18.114081 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c45b3cea-474a-437a-8e9a-2fc96e1d7200-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:18 crc kubenswrapper[4943]: I0307 15:13:18.114133 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c45b3cea-474a-437a-8e9a-2fc96e1d7200-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:18 crc kubenswrapper[4943]: I0307 15:13:18.114151 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c45b3cea-474a-437a-8e9a-2fc96e1d7200-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:18 crc kubenswrapper[4943]: I0307 15:13:18.114172 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c45b3cea-474a-437a-8e9a-2fc96e1d7200-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:18 crc kubenswrapper[4943]: I0307 15:13:18.114189 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zhk46\" (UniqueName: \"kubernetes.io/projected/c45b3cea-474a-437a-8e9a-2fc96e1d7200-kube-api-access-zhk46\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:18 crc kubenswrapper[4943]: I0307 15:13:18.114206 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c45b3cea-474a-437a-8e9a-2fc96e1d7200-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:18 crc kubenswrapper[4943]: I0307 15:13:18.513029 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="03b37b9c2a686a5638dc9f4289a40fca025ede7c21bc09a66ebed58524933ab7" Mar 07 15:13:18 crc kubenswrapper[4943]: I0307 15:13:18.513130 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-zhk9t" Mar 07 15:13:18 crc kubenswrapper[4943]: I0307 15:13:18.768613 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c45b3cea-474a-437a-8e9a-2fc96e1d7200" path="/var/lib/kubelet/pods/c45b3cea-474a-437a-8e9a-2fc96e1d7200/volumes" Mar 07 15:13:19 crc kubenswrapper[4943]: I0307 15:13:19.106678 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-wtm8b"] Mar 07 15:13:19 crc kubenswrapper[4943]: E0307 15:13:19.107200 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c45b3cea-474a-437a-8e9a-2fc96e1d7200" containerName="swift-ring-rebalance" Mar 07 15:13:19 crc kubenswrapper[4943]: I0307 15:13:19.107218 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="c45b3cea-474a-437a-8e9a-2fc96e1d7200" containerName="swift-ring-rebalance" Mar 07 15:13:19 crc kubenswrapper[4943]: I0307 15:13:19.107534 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="c45b3cea-474a-437a-8e9a-2fc96e1d7200" containerName="swift-ring-rebalance" Mar 07 15:13:19 crc kubenswrapper[4943]: I0307 15:13:19.108322 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wtm8b" Mar 07 15:13:19 crc kubenswrapper[4943]: I0307 15:13:19.111780 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:13:19 crc kubenswrapper[4943]: I0307 15:13:19.112066 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:13:19 crc kubenswrapper[4943]: I0307 15:13:19.117378 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-wtm8b"] Mar 07 15:13:19 crc kubenswrapper[4943]: I0307 15:13:19.231907 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d3079fd0-e173-4c2c-a8ce-7775f8ca98f9-dispersionconf\") pod \"swift-ring-rebalance-debug-wtm8b\" (UID: \"d3079fd0-e173-4c2c-a8ce-7775f8ca98f9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wtm8b" Mar 07 15:13:19 crc kubenswrapper[4943]: I0307 15:13:19.232007 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d3079fd0-e173-4c2c-a8ce-7775f8ca98f9-etc-swift\") pod \"swift-ring-rebalance-debug-wtm8b\" (UID: \"d3079fd0-e173-4c2c-a8ce-7775f8ca98f9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wtm8b" Mar 07 15:13:19 crc kubenswrapper[4943]: I0307 15:13:19.232076 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7mcv\" (UniqueName: \"kubernetes.io/projected/d3079fd0-e173-4c2c-a8ce-7775f8ca98f9-kube-api-access-n7mcv\") pod \"swift-ring-rebalance-debug-wtm8b\" (UID: \"d3079fd0-e173-4c2c-a8ce-7775f8ca98f9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wtm8b" Mar 07 15:13:19 crc kubenswrapper[4943]: I0307 15:13:19.232100 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d3079fd0-e173-4c2c-a8ce-7775f8ca98f9-ring-data-devices\") pod \"swift-ring-rebalance-debug-wtm8b\" (UID: \"d3079fd0-e173-4c2c-a8ce-7775f8ca98f9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wtm8b" Mar 07 15:13:19 crc kubenswrapper[4943]: I0307 15:13:19.232124 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d3079fd0-e173-4c2c-a8ce-7775f8ca98f9-scripts\") pod \"swift-ring-rebalance-debug-wtm8b\" (UID: \"d3079fd0-e173-4c2c-a8ce-7775f8ca98f9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wtm8b" Mar 07 15:13:19 crc kubenswrapper[4943]: I0307 15:13:19.232160 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d3079fd0-e173-4c2c-a8ce-7775f8ca98f9-swiftconf\") pod \"swift-ring-rebalance-debug-wtm8b\" (UID: \"d3079fd0-e173-4c2c-a8ce-7775f8ca98f9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wtm8b" Mar 07 15:13:19 crc kubenswrapper[4943]: I0307 15:13:19.333521 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d3079fd0-e173-4c2c-a8ce-7775f8ca98f9-dispersionconf\") pod \"swift-ring-rebalance-debug-wtm8b\" (UID: \"d3079fd0-e173-4c2c-a8ce-7775f8ca98f9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wtm8b" Mar 07 15:13:19 crc kubenswrapper[4943]: I0307 15:13:19.333566 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d3079fd0-e173-4c2c-a8ce-7775f8ca98f9-etc-swift\") pod \"swift-ring-rebalance-debug-wtm8b\" (UID: \"d3079fd0-e173-4c2c-a8ce-7775f8ca98f9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wtm8b" Mar 07 15:13:19 crc kubenswrapper[4943]: I0307 15:13:19.333616 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7mcv\" (UniqueName: \"kubernetes.io/projected/d3079fd0-e173-4c2c-a8ce-7775f8ca98f9-kube-api-access-n7mcv\") pod \"swift-ring-rebalance-debug-wtm8b\" (UID: \"d3079fd0-e173-4c2c-a8ce-7775f8ca98f9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wtm8b" Mar 07 15:13:19 crc kubenswrapper[4943]: I0307 15:13:19.333632 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d3079fd0-e173-4c2c-a8ce-7775f8ca98f9-ring-data-devices\") pod \"swift-ring-rebalance-debug-wtm8b\" (UID: \"d3079fd0-e173-4c2c-a8ce-7775f8ca98f9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wtm8b" Mar 07 15:13:19 crc kubenswrapper[4943]: I0307 15:13:19.333650 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d3079fd0-e173-4c2c-a8ce-7775f8ca98f9-scripts\") pod \"swift-ring-rebalance-debug-wtm8b\" (UID: \"d3079fd0-e173-4c2c-a8ce-7775f8ca98f9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wtm8b" Mar 07 15:13:19 crc kubenswrapper[4943]: I0307 15:13:19.333678 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d3079fd0-e173-4c2c-a8ce-7775f8ca98f9-swiftconf\") pod \"swift-ring-rebalance-debug-wtm8b\" (UID: \"d3079fd0-e173-4c2c-a8ce-7775f8ca98f9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wtm8b" Mar 07 15:13:19 crc kubenswrapper[4943]: I0307 15:13:19.334245 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d3079fd0-e173-4c2c-a8ce-7775f8ca98f9-etc-swift\") pod \"swift-ring-rebalance-debug-wtm8b\" (UID: \"d3079fd0-e173-4c2c-a8ce-7775f8ca98f9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wtm8b" Mar 07 15:13:19 crc kubenswrapper[4943]: I0307 15:13:19.334618 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d3079fd0-e173-4c2c-a8ce-7775f8ca98f9-scripts\") pod \"swift-ring-rebalance-debug-wtm8b\" (UID: \"d3079fd0-e173-4c2c-a8ce-7775f8ca98f9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wtm8b" Mar 07 15:13:19 crc kubenswrapper[4943]: I0307 15:13:19.334782 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d3079fd0-e173-4c2c-a8ce-7775f8ca98f9-ring-data-devices\") pod \"swift-ring-rebalance-debug-wtm8b\" (UID: \"d3079fd0-e173-4c2c-a8ce-7775f8ca98f9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wtm8b" Mar 07 15:13:19 crc kubenswrapper[4943]: I0307 15:13:19.337459 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d3079fd0-e173-4c2c-a8ce-7775f8ca98f9-dispersionconf\") pod \"swift-ring-rebalance-debug-wtm8b\" (UID: \"d3079fd0-e173-4c2c-a8ce-7775f8ca98f9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wtm8b" Mar 07 15:13:19 crc kubenswrapper[4943]: I0307 15:13:19.338552 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d3079fd0-e173-4c2c-a8ce-7775f8ca98f9-swiftconf\") pod \"swift-ring-rebalance-debug-wtm8b\" (UID: \"d3079fd0-e173-4c2c-a8ce-7775f8ca98f9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wtm8b" Mar 07 15:13:19 crc kubenswrapper[4943]: I0307 15:13:19.352568 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7mcv\" (UniqueName: \"kubernetes.io/projected/d3079fd0-e173-4c2c-a8ce-7775f8ca98f9-kube-api-access-n7mcv\") pod \"swift-ring-rebalance-debug-wtm8b\" (UID: \"d3079fd0-e173-4c2c-a8ce-7775f8ca98f9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wtm8b" Mar 07 15:13:19 crc kubenswrapper[4943]: I0307 15:13:19.433095 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wtm8b" Mar 07 15:13:19 crc kubenswrapper[4943]: I0307 15:13:19.875940 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-wtm8b"] Mar 07 15:13:20 crc kubenswrapper[4943]: I0307 15:13:20.545884 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wtm8b" event={"ID":"d3079fd0-e173-4c2c-a8ce-7775f8ca98f9","Type":"ContainerStarted","Data":"d6a31cdb138264f6507f5de9000244ab3aeca8f599d047e044379d49425395da"} Mar 07 15:13:20 crc kubenswrapper[4943]: I0307 15:13:20.545964 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wtm8b" event={"ID":"d3079fd0-e173-4c2c-a8ce-7775f8ca98f9","Type":"ContainerStarted","Data":"0f84fc5dd902f82c37d3ec8bd210282480ddf0a2cc3d70d68543723f295c1fbb"} Mar 07 15:13:20 crc kubenswrapper[4943]: I0307 15:13:20.570413 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wtm8b" podStartSLOduration=1.5703861639999999 podStartE2EDuration="1.570386164s" podCreationTimestamp="2026-03-07 15:13:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:13:20.567313018 +0000 UTC m=+2042.519449516" watchObservedRunningTime="2026-03-07 15:13:20.570386164 +0000 UTC m=+2042.522522662" Mar 07 15:13:21 crc kubenswrapper[4943]: I0307 15:13:21.555369 4943 generic.go:334] "Generic (PLEG): container finished" podID="d3079fd0-e173-4c2c-a8ce-7775f8ca98f9" containerID="d6a31cdb138264f6507f5de9000244ab3aeca8f599d047e044379d49425395da" exitCode=0 Mar 07 15:13:21 crc kubenswrapper[4943]: I0307 15:13:21.555432 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wtm8b" event={"ID":"d3079fd0-e173-4c2c-a8ce-7775f8ca98f9","Type":"ContainerDied","Data":"d6a31cdb138264f6507f5de9000244ab3aeca8f599d047e044379d49425395da"} Mar 07 15:13:22 crc kubenswrapper[4943]: I0307 15:13:22.936840 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wtm8b" Mar 07 15:13:22 crc kubenswrapper[4943]: I0307 15:13:22.978316 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-wtm8b"] Mar 07 15:13:22 crc kubenswrapper[4943]: I0307 15:13:22.989843 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-wtm8b"] Mar 07 15:13:23 crc kubenswrapper[4943]: I0307 15:13:23.094597 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d3079fd0-e173-4c2c-a8ce-7775f8ca98f9-ring-data-devices\") pod \"d3079fd0-e173-4c2c-a8ce-7775f8ca98f9\" (UID: \"d3079fd0-e173-4c2c-a8ce-7775f8ca98f9\") " Mar 07 15:13:23 crc kubenswrapper[4943]: I0307 15:13:23.094678 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d3079fd0-e173-4c2c-a8ce-7775f8ca98f9-swiftconf\") pod \"d3079fd0-e173-4c2c-a8ce-7775f8ca98f9\" (UID: \"d3079fd0-e173-4c2c-a8ce-7775f8ca98f9\") " Mar 07 15:13:23 crc kubenswrapper[4943]: I0307 15:13:23.094757 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n7mcv\" (UniqueName: \"kubernetes.io/projected/d3079fd0-e173-4c2c-a8ce-7775f8ca98f9-kube-api-access-n7mcv\") pod \"d3079fd0-e173-4c2c-a8ce-7775f8ca98f9\" (UID: \"d3079fd0-e173-4c2c-a8ce-7775f8ca98f9\") " Mar 07 15:13:23 crc kubenswrapper[4943]: I0307 15:13:23.094799 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d3079fd0-e173-4c2c-a8ce-7775f8ca98f9-scripts\") pod \"d3079fd0-e173-4c2c-a8ce-7775f8ca98f9\" (UID: \"d3079fd0-e173-4c2c-a8ce-7775f8ca98f9\") " Mar 07 15:13:23 crc kubenswrapper[4943]: I0307 15:13:23.094883 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d3079fd0-e173-4c2c-a8ce-7775f8ca98f9-etc-swift\") pod \"d3079fd0-e173-4c2c-a8ce-7775f8ca98f9\" (UID: \"d3079fd0-e173-4c2c-a8ce-7775f8ca98f9\") " Mar 07 15:13:23 crc kubenswrapper[4943]: I0307 15:13:23.094960 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d3079fd0-e173-4c2c-a8ce-7775f8ca98f9-dispersionconf\") pod \"d3079fd0-e173-4c2c-a8ce-7775f8ca98f9\" (UID: \"d3079fd0-e173-4c2c-a8ce-7775f8ca98f9\") " Mar 07 15:13:23 crc kubenswrapper[4943]: I0307 15:13:23.096076 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3079fd0-e173-4c2c-a8ce-7775f8ca98f9-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "d3079fd0-e173-4c2c-a8ce-7775f8ca98f9" (UID: "d3079fd0-e173-4c2c-a8ce-7775f8ca98f9"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:13:23 crc kubenswrapper[4943]: I0307 15:13:23.096306 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3079fd0-e173-4c2c-a8ce-7775f8ca98f9-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "d3079fd0-e173-4c2c-a8ce-7775f8ca98f9" (UID: "d3079fd0-e173-4c2c-a8ce-7775f8ca98f9"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:13:23 crc kubenswrapper[4943]: I0307 15:13:23.101482 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3079fd0-e173-4c2c-a8ce-7775f8ca98f9-kube-api-access-n7mcv" (OuterVolumeSpecName: "kube-api-access-n7mcv") pod "d3079fd0-e173-4c2c-a8ce-7775f8ca98f9" (UID: "d3079fd0-e173-4c2c-a8ce-7775f8ca98f9"). InnerVolumeSpecName "kube-api-access-n7mcv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:13:23 crc kubenswrapper[4943]: I0307 15:13:23.114435 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3079fd0-e173-4c2c-a8ce-7775f8ca98f9-scripts" (OuterVolumeSpecName: "scripts") pod "d3079fd0-e173-4c2c-a8ce-7775f8ca98f9" (UID: "d3079fd0-e173-4c2c-a8ce-7775f8ca98f9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:13:23 crc kubenswrapper[4943]: I0307 15:13:23.125180 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3079fd0-e173-4c2c-a8ce-7775f8ca98f9-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "d3079fd0-e173-4c2c-a8ce-7775f8ca98f9" (UID: "d3079fd0-e173-4c2c-a8ce-7775f8ca98f9"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:13:23 crc kubenswrapper[4943]: I0307 15:13:23.129989 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3079fd0-e173-4c2c-a8ce-7775f8ca98f9-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "d3079fd0-e173-4c2c-a8ce-7775f8ca98f9" (UID: "d3079fd0-e173-4c2c-a8ce-7775f8ca98f9"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:13:23 crc kubenswrapper[4943]: I0307 15:13:23.196901 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n7mcv\" (UniqueName: \"kubernetes.io/projected/d3079fd0-e173-4c2c-a8ce-7775f8ca98f9-kube-api-access-n7mcv\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:23 crc kubenswrapper[4943]: I0307 15:13:23.196952 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d3079fd0-e173-4c2c-a8ce-7775f8ca98f9-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:23 crc kubenswrapper[4943]: I0307 15:13:23.196964 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d3079fd0-e173-4c2c-a8ce-7775f8ca98f9-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:23 crc kubenswrapper[4943]: I0307 15:13:23.196978 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d3079fd0-e173-4c2c-a8ce-7775f8ca98f9-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:23 crc kubenswrapper[4943]: I0307 15:13:23.196988 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d3079fd0-e173-4c2c-a8ce-7775f8ca98f9-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:23 crc kubenswrapper[4943]: I0307 15:13:23.196997 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d3079fd0-e173-4c2c-a8ce-7775f8ca98f9-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:23 crc kubenswrapper[4943]: I0307 15:13:23.585421 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0f84fc5dd902f82c37d3ec8bd210282480ddf0a2cc3d70d68543723f295c1fbb" Mar 07 15:13:23 crc kubenswrapper[4943]: I0307 15:13:23.585490 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wtm8b" Mar 07 15:13:24 crc kubenswrapper[4943]: I0307 15:13:24.161071 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-z2h4l"] Mar 07 15:13:24 crc kubenswrapper[4943]: E0307 15:13:24.161450 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3079fd0-e173-4c2c-a8ce-7775f8ca98f9" containerName="swift-ring-rebalance" Mar 07 15:13:24 crc kubenswrapper[4943]: I0307 15:13:24.161472 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3079fd0-e173-4c2c-a8ce-7775f8ca98f9" containerName="swift-ring-rebalance" Mar 07 15:13:24 crc kubenswrapper[4943]: I0307 15:13:24.161726 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3079fd0-e173-4c2c-a8ce-7775f8ca98f9" containerName="swift-ring-rebalance" Mar 07 15:13:24 crc kubenswrapper[4943]: I0307 15:13:24.162808 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-z2h4l" Mar 07 15:13:24 crc kubenswrapper[4943]: I0307 15:13:24.166284 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:13:24 crc kubenswrapper[4943]: I0307 15:13:24.167852 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:13:24 crc kubenswrapper[4943]: I0307 15:13:24.181717 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-z2h4l"] Mar 07 15:13:24 crc kubenswrapper[4943]: I0307 15:13:24.313064 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9108a5d2-66ef-4be6-9503-7f006e13c992-swiftconf\") pod \"swift-ring-rebalance-debug-z2h4l\" (UID: \"9108a5d2-66ef-4be6-9503-7f006e13c992\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-z2h4l" Mar 07 15:13:24 crc kubenswrapper[4943]: I0307 15:13:24.313150 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbl79\" (UniqueName: \"kubernetes.io/projected/9108a5d2-66ef-4be6-9503-7f006e13c992-kube-api-access-kbl79\") pod \"swift-ring-rebalance-debug-z2h4l\" (UID: \"9108a5d2-66ef-4be6-9503-7f006e13c992\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-z2h4l" Mar 07 15:13:24 crc kubenswrapper[4943]: I0307 15:13:24.313197 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9108a5d2-66ef-4be6-9503-7f006e13c992-ring-data-devices\") pod \"swift-ring-rebalance-debug-z2h4l\" (UID: \"9108a5d2-66ef-4be6-9503-7f006e13c992\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-z2h4l" Mar 07 15:13:24 crc kubenswrapper[4943]: I0307 15:13:24.313223 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9108a5d2-66ef-4be6-9503-7f006e13c992-dispersionconf\") pod \"swift-ring-rebalance-debug-z2h4l\" (UID: \"9108a5d2-66ef-4be6-9503-7f006e13c992\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-z2h4l" Mar 07 15:13:24 crc kubenswrapper[4943]: I0307 15:13:24.313267 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9108a5d2-66ef-4be6-9503-7f006e13c992-scripts\") pod \"swift-ring-rebalance-debug-z2h4l\" (UID: \"9108a5d2-66ef-4be6-9503-7f006e13c992\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-z2h4l" Mar 07 15:13:24 crc kubenswrapper[4943]: I0307 15:13:24.313568 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9108a5d2-66ef-4be6-9503-7f006e13c992-etc-swift\") pod \"swift-ring-rebalance-debug-z2h4l\" (UID: \"9108a5d2-66ef-4be6-9503-7f006e13c992\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-z2h4l" Mar 07 15:13:24 crc kubenswrapper[4943]: I0307 15:13:24.415716 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9108a5d2-66ef-4be6-9503-7f006e13c992-swiftconf\") pod \"swift-ring-rebalance-debug-z2h4l\" (UID: \"9108a5d2-66ef-4be6-9503-7f006e13c992\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-z2h4l" Mar 07 15:13:24 crc kubenswrapper[4943]: I0307 15:13:24.415825 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kbl79\" (UniqueName: \"kubernetes.io/projected/9108a5d2-66ef-4be6-9503-7f006e13c992-kube-api-access-kbl79\") pod \"swift-ring-rebalance-debug-z2h4l\" (UID: \"9108a5d2-66ef-4be6-9503-7f006e13c992\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-z2h4l" Mar 07 15:13:24 crc kubenswrapper[4943]: I0307 15:13:24.415868 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9108a5d2-66ef-4be6-9503-7f006e13c992-ring-data-devices\") pod \"swift-ring-rebalance-debug-z2h4l\" (UID: \"9108a5d2-66ef-4be6-9503-7f006e13c992\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-z2h4l" Mar 07 15:13:24 crc kubenswrapper[4943]: I0307 15:13:24.415909 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9108a5d2-66ef-4be6-9503-7f006e13c992-dispersionconf\") pod \"swift-ring-rebalance-debug-z2h4l\" (UID: \"9108a5d2-66ef-4be6-9503-7f006e13c992\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-z2h4l" Mar 07 15:13:24 crc kubenswrapper[4943]: I0307 15:13:24.415993 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9108a5d2-66ef-4be6-9503-7f006e13c992-scripts\") pod \"swift-ring-rebalance-debug-z2h4l\" (UID: \"9108a5d2-66ef-4be6-9503-7f006e13c992\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-z2h4l" Mar 07 15:13:24 crc kubenswrapper[4943]: I0307 15:13:24.416115 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9108a5d2-66ef-4be6-9503-7f006e13c992-etc-swift\") pod \"swift-ring-rebalance-debug-z2h4l\" (UID: \"9108a5d2-66ef-4be6-9503-7f006e13c992\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-z2h4l" Mar 07 15:13:24 crc kubenswrapper[4943]: I0307 15:13:24.417297 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9108a5d2-66ef-4be6-9503-7f006e13c992-etc-swift\") pod \"swift-ring-rebalance-debug-z2h4l\" (UID: \"9108a5d2-66ef-4be6-9503-7f006e13c992\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-z2h4l" Mar 07 15:13:24 crc kubenswrapper[4943]: I0307 15:13:24.417548 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9108a5d2-66ef-4be6-9503-7f006e13c992-ring-data-devices\") pod \"swift-ring-rebalance-debug-z2h4l\" (UID: \"9108a5d2-66ef-4be6-9503-7f006e13c992\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-z2h4l" Mar 07 15:13:24 crc kubenswrapper[4943]: I0307 15:13:24.420100 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9108a5d2-66ef-4be6-9503-7f006e13c992-scripts\") pod \"swift-ring-rebalance-debug-z2h4l\" (UID: \"9108a5d2-66ef-4be6-9503-7f006e13c992\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-z2h4l" Mar 07 15:13:24 crc kubenswrapper[4943]: I0307 15:13:24.424119 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9108a5d2-66ef-4be6-9503-7f006e13c992-swiftconf\") pod \"swift-ring-rebalance-debug-z2h4l\" (UID: \"9108a5d2-66ef-4be6-9503-7f006e13c992\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-z2h4l" Mar 07 15:13:24 crc kubenswrapper[4943]: I0307 15:13:24.430206 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9108a5d2-66ef-4be6-9503-7f006e13c992-dispersionconf\") pod \"swift-ring-rebalance-debug-z2h4l\" (UID: \"9108a5d2-66ef-4be6-9503-7f006e13c992\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-z2h4l" Mar 07 15:13:24 crc kubenswrapper[4943]: I0307 15:13:24.450590 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kbl79\" (UniqueName: \"kubernetes.io/projected/9108a5d2-66ef-4be6-9503-7f006e13c992-kube-api-access-kbl79\") pod \"swift-ring-rebalance-debug-z2h4l\" (UID: \"9108a5d2-66ef-4be6-9503-7f006e13c992\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-z2h4l" Mar 07 15:13:24 crc kubenswrapper[4943]: I0307 15:13:24.504908 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-z2h4l" Mar 07 15:13:24 crc kubenswrapper[4943]: I0307 15:13:24.749674 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-z2h4l"] Mar 07 15:13:24 crc kubenswrapper[4943]: I0307 15:13:24.777262 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3079fd0-e173-4c2c-a8ce-7775f8ca98f9" path="/var/lib/kubelet/pods/d3079fd0-e173-4c2c-a8ce-7775f8ca98f9/volumes" Mar 07 15:13:24 crc kubenswrapper[4943]: W0307 15:13:24.789355 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9108a5d2_66ef_4be6_9503_7f006e13c992.slice/crio-910546f1b3754de8c94c0c15f40c70de7b1f40df0b3707d41835699ef90e7ecb WatchSource:0}: Error finding container 910546f1b3754de8c94c0c15f40c70de7b1f40df0b3707d41835699ef90e7ecb: Status 404 returned error can't find the container with id 910546f1b3754de8c94c0c15f40c70de7b1f40df0b3707d41835699ef90e7ecb Mar 07 15:13:25 crc kubenswrapper[4943]: I0307 15:13:25.604237 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-z2h4l" event={"ID":"9108a5d2-66ef-4be6-9503-7f006e13c992","Type":"ContainerStarted","Data":"19aeee34f55b8a67a45502efa5df20f87d64a97e23eb243996a57ad16dda0bb1"} Mar 07 15:13:25 crc kubenswrapper[4943]: I0307 15:13:25.604584 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-z2h4l" event={"ID":"9108a5d2-66ef-4be6-9503-7f006e13c992","Type":"ContainerStarted","Data":"910546f1b3754de8c94c0c15f40c70de7b1f40df0b3707d41835699ef90e7ecb"} Mar 07 15:13:25 crc kubenswrapper[4943]: I0307 15:13:25.625341 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-z2h4l" podStartSLOduration=1.625318488 podStartE2EDuration="1.625318488s" podCreationTimestamp="2026-03-07 15:13:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:13:25.623551574 +0000 UTC m=+2047.575688082" watchObservedRunningTime="2026-03-07 15:13:25.625318488 +0000 UTC m=+2047.577454996" Mar 07 15:13:26 crc kubenswrapper[4943]: I0307 15:13:26.616859 4943 generic.go:334] "Generic (PLEG): container finished" podID="9108a5d2-66ef-4be6-9503-7f006e13c992" containerID="19aeee34f55b8a67a45502efa5df20f87d64a97e23eb243996a57ad16dda0bb1" exitCode=0 Mar 07 15:13:26 crc kubenswrapper[4943]: I0307 15:13:26.616920 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-z2h4l" event={"ID":"9108a5d2-66ef-4be6-9503-7f006e13c992","Type":"ContainerDied","Data":"19aeee34f55b8a67a45502efa5df20f87d64a97e23eb243996a57ad16dda0bb1"} Mar 07 15:13:28 crc kubenswrapper[4943]: I0307 15:13:28.025987 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-z2h4l" Mar 07 15:13:28 crc kubenswrapper[4943]: I0307 15:13:28.066418 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-z2h4l"] Mar 07 15:13:28 crc kubenswrapper[4943]: I0307 15:13:28.072291 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-z2h4l"] Mar 07 15:13:28 crc kubenswrapper[4943]: I0307 15:13:28.187958 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9108a5d2-66ef-4be6-9503-7f006e13c992-etc-swift\") pod \"9108a5d2-66ef-4be6-9503-7f006e13c992\" (UID: \"9108a5d2-66ef-4be6-9503-7f006e13c992\") " Mar 07 15:13:28 crc kubenswrapper[4943]: I0307 15:13:28.188046 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9108a5d2-66ef-4be6-9503-7f006e13c992-scripts\") pod \"9108a5d2-66ef-4be6-9503-7f006e13c992\" (UID: \"9108a5d2-66ef-4be6-9503-7f006e13c992\") " Mar 07 15:13:28 crc kubenswrapper[4943]: I0307 15:13:28.188093 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kbl79\" (UniqueName: \"kubernetes.io/projected/9108a5d2-66ef-4be6-9503-7f006e13c992-kube-api-access-kbl79\") pod \"9108a5d2-66ef-4be6-9503-7f006e13c992\" (UID: \"9108a5d2-66ef-4be6-9503-7f006e13c992\") " Mar 07 15:13:28 crc kubenswrapper[4943]: I0307 15:13:28.188150 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9108a5d2-66ef-4be6-9503-7f006e13c992-dispersionconf\") pod \"9108a5d2-66ef-4be6-9503-7f006e13c992\" (UID: \"9108a5d2-66ef-4be6-9503-7f006e13c992\") " Mar 07 15:13:28 crc kubenswrapper[4943]: I0307 15:13:28.188167 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9108a5d2-66ef-4be6-9503-7f006e13c992-swiftconf\") pod \"9108a5d2-66ef-4be6-9503-7f006e13c992\" (UID: \"9108a5d2-66ef-4be6-9503-7f006e13c992\") " Mar 07 15:13:28 crc kubenswrapper[4943]: I0307 15:13:28.188196 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9108a5d2-66ef-4be6-9503-7f006e13c992-ring-data-devices\") pod \"9108a5d2-66ef-4be6-9503-7f006e13c992\" (UID: \"9108a5d2-66ef-4be6-9503-7f006e13c992\") " Mar 07 15:13:28 crc kubenswrapper[4943]: I0307 15:13:28.188567 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9108a5d2-66ef-4be6-9503-7f006e13c992-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "9108a5d2-66ef-4be6-9503-7f006e13c992" (UID: "9108a5d2-66ef-4be6-9503-7f006e13c992"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:13:28 crc kubenswrapper[4943]: I0307 15:13:28.188918 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9108a5d2-66ef-4be6-9503-7f006e13c992-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "9108a5d2-66ef-4be6-9503-7f006e13c992" (UID: "9108a5d2-66ef-4be6-9503-7f006e13c992"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:13:28 crc kubenswrapper[4943]: I0307 15:13:28.195139 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9108a5d2-66ef-4be6-9503-7f006e13c992-kube-api-access-kbl79" (OuterVolumeSpecName: "kube-api-access-kbl79") pod "9108a5d2-66ef-4be6-9503-7f006e13c992" (UID: "9108a5d2-66ef-4be6-9503-7f006e13c992"). InnerVolumeSpecName "kube-api-access-kbl79". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:13:28 crc kubenswrapper[4943]: I0307 15:13:28.209724 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9108a5d2-66ef-4be6-9503-7f006e13c992-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "9108a5d2-66ef-4be6-9503-7f006e13c992" (UID: "9108a5d2-66ef-4be6-9503-7f006e13c992"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:13:28 crc kubenswrapper[4943]: I0307 15:13:28.220822 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9108a5d2-66ef-4be6-9503-7f006e13c992-scripts" (OuterVolumeSpecName: "scripts") pod "9108a5d2-66ef-4be6-9503-7f006e13c992" (UID: "9108a5d2-66ef-4be6-9503-7f006e13c992"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:13:28 crc kubenswrapper[4943]: I0307 15:13:28.224188 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9108a5d2-66ef-4be6-9503-7f006e13c992-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "9108a5d2-66ef-4be6-9503-7f006e13c992" (UID: "9108a5d2-66ef-4be6-9503-7f006e13c992"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:13:28 crc kubenswrapper[4943]: I0307 15:13:28.290639 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kbl79\" (UniqueName: \"kubernetes.io/projected/9108a5d2-66ef-4be6-9503-7f006e13c992-kube-api-access-kbl79\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:28 crc kubenswrapper[4943]: I0307 15:13:28.290715 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9108a5d2-66ef-4be6-9503-7f006e13c992-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:28 crc kubenswrapper[4943]: I0307 15:13:28.290747 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9108a5d2-66ef-4be6-9503-7f006e13c992-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:28 crc kubenswrapper[4943]: I0307 15:13:28.290770 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9108a5d2-66ef-4be6-9503-7f006e13c992-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:28 crc kubenswrapper[4943]: I0307 15:13:28.290791 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9108a5d2-66ef-4be6-9503-7f006e13c992-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:28 crc kubenswrapper[4943]: I0307 15:13:28.290815 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9108a5d2-66ef-4be6-9503-7f006e13c992-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:28 crc kubenswrapper[4943]: I0307 15:13:28.640615 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="910546f1b3754de8c94c0c15f40c70de7b1f40df0b3707d41835699ef90e7ecb" Mar 07 15:13:28 crc kubenswrapper[4943]: I0307 15:13:28.640719 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-z2h4l" Mar 07 15:13:28 crc kubenswrapper[4943]: I0307 15:13:28.769749 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9108a5d2-66ef-4be6-9503-7f006e13c992" path="/var/lib/kubelet/pods/9108a5d2-66ef-4be6-9503-7f006e13c992/volumes" Mar 07 15:13:29 crc kubenswrapper[4943]: I0307 15:13:29.250667 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-76qn9"] Mar 07 15:13:29 crc kubenswrapper[4943]: E0307 15:13:29.251654 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9108a5d2-66ef-4be6-9503-7f006e13c992" containerName="swift-ring-rebalance" Mar 07 15:13:29 crc kubenswrapper[4943]: I0307 15:13:29.251687 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="9108a5d2-66ef-4be6-9503-7f006e13c992" containerName="swift-ring-rebalance" Mar 07 15:13:29 crc kubenswrapper[4943]: I0307 15:13:29.252005 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="9108a5d2-66ef-4be6-9503-7f006e13c992" containerName="swift-ring-rebalance" Mar 07 15:13:29 crc kubenswrapper[4943]: I0307 15:13:29.252774 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-76qn9" Mar 07 15:13:29 crc kubenswrapper[4943]: I0307 15:13:29.254996 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:13:29 crc kubenswrapper[4943]: I0307 15:13:29.256266 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:13:29 crc kubenswrapper[4943]: I0307 15:13:29.271574 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-76qn9"] Mar 07 15:13:29 crc kubenswrapper[4943]: I0307 15:13:29.407124 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3247c7e7-b192-4f48-9f6c-494dd33578c6-etc-swift\") pod \"swift-ring-rebalance-debug-76qn9\" (UID: \"3247c7e7-b192-4f48-9f6c-494dd33578c6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-76qn9" Mar 07 15:13:29 crc kubenswrapper[4943]: I0307 15:13:29.407210 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9msc\" (UniqueName: \"kubernetes.io/projected/3247c7e7-b192-4f48-9f6c-494dd33578c6-kube-api-access-w9msc\") pod \"swift-ring-rebalance-debug-76qn9\" (UID: \"3247c7e7-b192-4f48-9f6c-494dd33578c6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-76qn9" Mar 07 15:13:29 crc kubenswrapper[4943]: I0307 15:13:29.407268 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3247c7e7-b192-4f48-9f6c-494dd33578c6-dispersionconf\") pod \"swift-ring-rebalance-debug-76qn9\" (UID: \"3247c7e7-b192-4f48-9f6c-494dd33578c6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-76qn9" Mar 07 15:13:29 crc kubenswrapper[4943]: I0307 15:13:29.407316 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3247c7e7-b192-4f48-9f6c-494dd33578c6-swiftconf\") pod \"swift-ring-rebalance-debug-76qn9\" (UID: \"3247c7e7-b192-4f48-9f6c-494dd33578c6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-76qn9" Mar 07 15:13:29 crc kubenswrapper[4943]: I0307 15:13:29.407494 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3247c7e7-b192-4f48-9f6c-494dd33578c6-scripts\") pod \"swift-ring-rebalance-debug-76qn9\" (UID: \"3247c7e7-b192-4f48-9f6c-494dd33578c6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-76qn9" Mar 07 15:13:29 crc kubenswrapper[4943]: I0307 15:13:29.407589 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3247c7e7-b192-4f48-9f6c-494dd33578c6-ring-data-devices\") pod \"swift-ring-rebalance-debug-76qn9\" (UID: \"3247c7e7-b192-4f48-9f6c-494dd33578c6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-76qn9" Mar 07 15:13:29 crc kubenswrapper[4943]: I0307 15:13:29.509666 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9msc\" (UniqueName: \"kubernetes.io/projected/3247c7e7-b192-4f48-9f6c-494dd33578c6-kube-api-access-w9msc\") pod \"swift-ring-rebalance-debug-76qn9\" (UID: \"3247c7e7-b192-4f48-9f6c-494dd33578c6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-76qn9" Mar 07 15:13:29 crc kubenswrapper[4943]: I0307 15:13:29.509782 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3247c7e7-b192-4f48-9f6c-494dd33578c6-dispersionconf\") pod \"swift-ring-rebalance-debug-76qn9\" (UID: \"3247c7e7-b192-4f48-9f6c-494dd33578c6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-76qn9" Mar 07 15:13:29 crc kubenswrapper[4943]: I0307 15:13:29.509832 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3247c7e7-b192-4f48-9f6c-494dd33578c6-swiftconf\") pod \"swift-ring-rebalance-debug-76qn9\" (UID: \"3247c7e7-b192-4f48-9f6c-494dd33578c6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-76qn9" Mar 07 15:13:29 crc kubenswrapper[4943]: I0307 15:13:29.510072 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3247c7e7-b192-4f48-9f6c-494dd33578c6-scripts\") pod \"swift-ring-rebalance-debug-76qn9\" (UID: \"3247c7e7-b192-4f48-9f6c-494dd33578c6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-76qn9" Mar 07 15:13:29 crc kubenswrapper[4943]: I0307 15:13:29.510150 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3247c7e7-b192-4f48-9f6c-494dd33578c6-ring-data-devices\") pod \"swift-ring-rebalance-debug-76qn9\" (UID: \"3247c7e7-b192-4f48-9f6c-494dd33578c6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-76qn9" Mar 07 15:13:29 crc kubenswrapper[4943]: I0307 15:13:29.510230 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3247c7e7-b192-4f48-9f6c-494dd33578c6-etc-swift\") pod \"swift-ring-rebalance-debug-76qn9\" (UID: \"3247c7e7-b192-4f48-9f6c-494dd33578c6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-76qn9" Mar 07 15:13:29 crc kubenswrapper[4943]: I0307 15:13:29.510770 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3247c7e7-b192-4f48-9f6c-494dd33578c6-scripts\") pod \"swift-ring-rebalance-debug-76qn9\" (UID: \"3247c7e7-b192-4f48-9f6c-494dd33578c6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-76qn9" Mar 07 15:13:29 crc kubenswrapper[4943]: I0307 15:13:29.510777 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3247c7e7-b192-4f48-9f6c-494dd33578c6-etc-swift\") pod \"swift-ring-rebalance-debug-76qn9\" (UID: \"3247c7e7-b192-4f48-9f6c-494dd33578c6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-76qn9" Mar 07 15:13:29 crc kubenswrapper[4943]: I0307 15:13:29.510852 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3247c7e7-b192-4f48-9f6c-494dd33578c6-ring-data-devices\") pod \"swift-ring-rebalance-debug-76qn9\" (UID: \"3247c7e7-b192-4f48-9f6c-494dd33578c6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-76qn9" Mar 07 15:13:29 crc kubenswrapper[4943]: I0307 15:13:29.515552 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3247c7e7-b192-4f48-9f6c-494dd33578c6-swiftconf\") pod \"swift-ring-rebalance-debug-76qn9\" (UID: \"3247c7e7-b192-4f48-9f6c-494dd33578c6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-76qn9" Mar 07 15:13:29 crc kubenswrapper[4943]: I0307 15:13:29.516272 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3247c7e7-b192-4f48-9f6c-494dd33578c6-dispersionconf\") pod \"swift-ring-rebalance-debug-76qn9\" (UID: \"3247c7e7-b192-4f48-9f6c-494dd33578c6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-76qn9" Mar 07 15:13:29 crc kubenswrapper[4943]: I0307 15:13:29.531321 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9msc\" (UniqueName: \"kubernetes.io/projected/3247c7e7-b192-4f48-9f6c-494dd33578c6-kube-api-access-w9msc\") pod \"swift-ring-rebalance-debug-76qn9\" (UID: \"3247c7e7-b192-4f48-9f6c-494dd33578c6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-76qn9" Mar 07 15:13:29 crc kubenswrapper[4943]: I0307 15:13:29.578330 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-76qn9" Mar 07 15:13:30 crc kubenswrapper[4943]: I0307 15:13:30.126203 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-76qn9"] Mar 07 15:13:30 crc kubenswrapper[4943]: I0307 15:13:30.671333 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-76qn9" event={"ID":"3247c7e7-b192-4f48-9f6c-494dd33578c6","Type":"ContainerStarted","Data":"4c521497d06b0b4870529d7fe6ba7b461ebfa9e691dd8007da6a3c477d54a0b4"} Mar 07 15:13:30 crc kubenswrapper[4943]: I0307 15:13:30.671597 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-76qn9" event={"ID":"3247c7e7-b192-4f48-9f6c-494dd33578c6","Type":"ContainerStarted","Data":"b162c332b5df6770a527174d69858d160efb9f4d78c38dff9cbe2c231591b15d"} Mar 07 15:13:30 crc kubenswrapper[4943]: I0307 15:13:30.699729 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-76qn9" podStartSLOduration=1.69971296 podStartE2EDuration="1.69971296s" podCreationTimestamp="2026-03-07 15:13:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:13:30.693130508 +0000 UTC m=+2052.645267006" watchObservedRunningTime="2026-03-07 15:13:30.69971296 +0000 UTC m=+2052.651849458" Mar 07 15:13:31 crc kubenswrapper[4943]: I0307 15:13:31.682755 4943 generic.go:334] "Generic (PLEG): container finished" podID="3247c7e7-b192-4f48-9f6c-494dd33578c6" containerID="4c521497d06b0b4870529d7fe6ba7b461ebfa9e691dd8007da6a3c477d54a0b4" exitCode=0 Mar 07 15:13:31 crc kubenswrapper[4943]: I0307 15:13:31.682819 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-76qn9" event={"ID":"3247c7e7-b192-4f48-9f6c-494dd33578c6","Type":"ContainerDied","Data":"4c521497d06b0b4870529d7fe6ba7b461ebfa9e691dd8007da6a3c477d54a0b4"} Mar 07 15:13:33 crc kubenswrapper[4943]: I0307 15:13:33.028691 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-76qn9" Mar 07 15:13:33 crc kubenswrapper[4943]: I0307 15:13:33.066058 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-76qn9"] Mar 07 15:13:33 crc kubenswrapper[4943]: I0307 15:13:33.070567 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-76qn9"] Mar 07 15:13:33 crc kubenswrapper[4943]: I0307 15:13:33.183465 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3247c7e7-b192-4f48-9f6c-494dd33578c6-dispersionconf\") pod \"3247c7e7-b192-4f48-9f6c-494dd33578c6\" (UID: \"3247c7e7-b192-4f48-9f6c-494dd33578c6\") " Mar 07 15:13:33 crc kubenswrapper[4943]: I0307 15:13:33.183755 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9msc\" (UniqueName: \"kubernetes.io/projected/3247c7e7-b192-4f48-9f6c-494dd33578c6-kube-api-access-w9msc\") pod \"3247c7e7-b192-4f48-9f6c-494dd33578c6\" (UID: \"3247c7e7-b192-4f48-9f6c-494dd33578c6\") " Mar 07 15:13:33 crc kubenswrapper[4943]: I0307 15:13:33.183829 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3247c7e7-b192-4f48-9f6c-494dd33578c6-etc-swift\") pod \"3247c7e7-b192-4f48-9f6c-494dd33578c6\" (UID: \"3247c7e7-b192-4f48-9f6c-494dd33578c6\") " Mar 07 15:13:33 crc kubenswrapper[4943]: I0307 15:13:33.183874 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3247c7e7-b192-4f48-9f6c-494dd33578c6-swiftconf\") pod \"3247c7e7-b192-4f48-9f6c-494dd33578c6\" (UID: \"3247c7e7-b192-4f48-9f6c-494dd33578c6\") " Mar 07 15:13:33 crc kubenswrapper[4943]: I0307 15:13:33.183918 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3247c7e7-b192-4f48-9f6c-494dd33578c6-ring-data-devices\") pod \"3247c7e7-b192-4f48-9f6c-494dd33578c6\" (UID: \"3247c7e7-b192-4f48-9f6c-494dd33578c6\") " Mar 07 15:13:33 crc kubenswrapper[4943]: I0307 15:13:33.183979 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3247c7e7-b192-4f48-9f6c-494dd33578c6-scripts\") pod \"3247c7e7-b192-4f48-9f6c-494dd33578c6\" (UID: \"3247c7e7-b192-4f48-9f6c-494dd33578c6\") " Mar 07 15:13:33 crc kubenswrapper[4943]: I0307 15:13:33.184408 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3247c7e7-b192-4f48-9f6c-494dd33578c6-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "3247c7e7-b192-4f48-9f6c-494dd33578c6" (UID: "3247c7e7-b192-4f48-9f6c-494dd33578c6"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:13:33 crc kubenswrapper[4943]: I0307 15:13:33.184547 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3247c7e7-b192-4f48-9f6c-494dd33578c6-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:33 crc kubenswrapper[4943]: I0307 15:13:33.184573 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3247c7e7-b192-4f48-9f6c-494dd33578c6-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "3247c7e7-b192-4f48-9f6c-494dd33578c6" (UID: "3247c7e7-b192-4f48-9f6c-494dd33578c6"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:13:33 crc kubenswrapper[4943]: I0307 15:13:33.190199 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3247c7e7-b192-4f48-9f6c-494dd33578c6-kube-api-access-w9msc" (OuterVolumeSpecName: "kube-api-access-w9msc") pod "3247c7e7-b192-4f48-9f6c-494dd33578c6" (UID: "3247c7e7-b192-4f48-9f6c-494dd33578c6"). InnerVolumeSpecName "kube-api-access-w9msc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:13:33 crc kubenswrapper[4943]: I0307 15:13:33.208963 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3247c7e7-b192-4f48-9f6c-494dd33578c6-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "3247c7e7-b192-4f48-9f6c-494dd33578c6" (UID: "3247c7e7-b192-4f48-9f6c-494dd33578c6"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:13:33 crc kubenswrapper[4943]: I0307 15:13:33.211960 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3247c7e7-b192-4f48-9f6c-494dd33578c6-scripts" (OuterVolumeSpecName: "scripts") pod "3247c7e7-b192-4f48-9f6c-494dd33578c6" (UID: "3247c7e7-b192-4f48-9f6c-494dd33578c6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:13:33 crc kubenswrapper[4943]: I0307 15:13:33.217243 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3247c7e7-b192-4f48-9f6c-494dd33578c6-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "3247c7e7-b192-4f48-9f6c-494dd33578c6" (UID: "3247c7e7-b192-4f48-9f6c-494dd33578c6"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:13:33 crc kubenswrapper[4943]: I0307 15:13:33.285708 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9msc\" (UniqueName: \"kubernetes.io/projected/3247c7e7-b192-4f48-9f6c-494dd33578c6-kube-api-access-w9msc\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:33 crc kubenswrapper[4943]: I0307 15:13:33.285741 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3247c7e7-b192-4f48-9f6c-494dd33578c6-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:33 crc kubenswrapper[4943]: I0307 15:13:33.285751 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3247c7e7-b192-4f48-9f6c-494dd33578c6-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:33 crc kubenswrapper[4943]: I0307 15:13:33.285761 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3247c7e7-b192-4f48-9f6c-494dd33578c6-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:33 crc kubenswrapper[4943]: I0307 15:13:33.285770 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3247c7e7-b192-4f48-9f6c-494dd33578c6-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:33 crc kubenswrapper[4943]: I0307 15:13:33.712463 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b162c332b5df6770a527174d69858d160efb9f4d78c38dff9cbe2c231591b15d" Mar 07 15:13:33 crc kubenswrapper[4943]: I0307 15:13:33.712545 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-76qn9" Mar 07 15:13:34 crc kubenswrapper[4943]: I0307 15:13:34.248452 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-rnhnd"] Mar 07 15:13:34 crc kubenswrapper[4943]: E0307 15:13:34.248923 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3247c7e7-b192-4f48-9f6c-494dd33578c6" containerName="swift-ring-rebalance" Mar 07 15:13:34 crc kubenswrapper[4943]: I0307 15:13:34.248974 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="3247c7e7-b192-4f48-9f6c-494dd33578c6" containerName="swift-ring-rebalance" Mar 07 15:13:34 crc kubenswrapper[4943]: I0307 15:13:34.249294 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="3247c7e7-b192-4f48-9f6c-494dd33578c6" containerName="swift-ring-rebalance" Mar 07 15:13:34 crc kubenswrapper[4943]: I0307 15:13:34.250064 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-rnhnd" Mar 07 15:13:34 crc kubenswrapper[4943]: I0307 15:13:34.253286 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:13:34 crc kubenswrapper[4943]: I0307 15:13:34.260800 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:13:34 crc kubenswrapper[4943]: I0307 15:13:34.264255 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-rnhnd"] Mar 07 15:13:34 crc kubenswrapper[4943]: I0307 15:13:34.416296 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f8cd3dfb-7ce4-4c22-9044-e483f38131b9-scripts\") pod \"swift-ring-rebalance-debug-rnhnd\" (UID: \"f8cd3dfb-7ce4-4c22-9044-e483f38131b9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rnhnd" Mar 07 15:13:34 crc kubenswrapper[4943]: I0307 15:13:34.416510 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mkkr7\" (UniqueName: \"kubernetes.io/projected/f8cd3dfb-7ce4-4c22-9044-e483f38131b9-kube-api-access-mkkr7\") pod \"swift-ring-rebalance-debug-rnhnd\" (UID: \"f8cd3dfb-7ce4-4c22-9044-e483f38131b9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rnhnd" Mar 07 15:13:34 crc kubenswrapper[4943]: I0307 15:13:34.416625 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f8cd3dfb-7ce4-4c22-9044-e483f38131b9-dispersionconf\") pod \"swift-ring-rebalance-debug-rnhnd\" (UID: \"f8cd3dfb-7ce4-4c22-9044-e483f38131b9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rnhnd" Mar 07 15:13:34 crc kubenswrapper[4943]: I0307 15:13:34.416724 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f8cd3dfb-7ce4-4c22-9044-e483f38131b9-etc-swift\") pod \"swift-ring-rebalance-debug-rnhnd\" (UID: \"f8cd3dfb-7ce4-4c22-9044-e483f38131b9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rnhnd" Mar 07 15:13:34 crc kubenswrapper[4943]: I0307 15:13:34.416821 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f8cd3dfb-7ce4-4c22-9044-e483f38131b9-swiftconf\") pod \"swift-ring-rebalance-debug-rnhnd\" (UID: \"f8cd3dfb-7ce4-4c22-9044-e483f38131b9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rnhnd" Mar 07 15:13:34 crc kubenswrapper[4943]: I0307 15:13:34.417563 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f8cd3dfb-7ce4-4c22-9044-e483f38131b9-ring-data-devices\") pod \"swift-ring-rebalance-debug-rnhnd\" (UID: \"f8cd3dfb-7ce4-4c22-9044-e483f38131b9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rnhnd" Mar 07 15:13:34 crc kubenswrapper[4943]: I0307 15:13:34.519780 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mkkr7\" (UniqueName: \"kubernetes.io/projected/f8cd3dfb-7ce4-4c22-9044-e483f38131b9-kube-api-access-mkkr7\") pod \"swift-ring-rebalance-debug-rnhnd\" (UID: \"f8cd3dfb-7ce4-4c22-9044-e483f38131b9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rnhnd" Mar 07 15:13:34 crc kubenswrapper[4943]: I0307 15:13:34.519863 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f8cd3dfb-7ce4-4c22-9044-e483f38131b9-dispersionconf\") pod \"swift-ring-rebalance-debug-rnhnd\" (UID: \"f8cd3dfb-7ce4-4c22-9044-e483f38131b9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rnhnd" Mar 07 15:13:34 crc kubenswrapper[4943]: I0307 15:13:34.519912 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f8cd3dfb-7ce4-4c22-9044-e483f38131b9-etc-swift\") pod \"swift-ring-rebalance-debug-rnhnd\" (UID: \"f8cd3dfb-7ce4-4c22-9044-e483f38131b9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rnhnd" Mar 07 15:13:34 crc kubenswrapper[4943]: I0307 15:13:34.520012 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f8cd3dfb-7ce4-4c22-9044-e483f38131b9-swiftconf\") pod \"swift-ring-rebalance-debug-rnhnd\" (UID: \"f8cd3dfb-7ce4-4c22-9044-e483f38131b9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rnhnd" Mar 07 15:13:34 crc kubenswrapper[4943]: I0307 15:13:34.520100 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f8cd3dfb-7ce4-4c22-9044-e483f38131b9-ring-data-devices\") pod \"swift-ring-rebalance-debug-rnhnd\" (UID: \"f8cd3dfb-7ce4-4c22-9044-e483f38131b9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rnhnd" Mar 07 15:13:34 crc kubenswrapper[4943]: I0307 15:13:34.520200 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f8cd3dfb-7ce4-4c22-9044-e483f38131b9-scripts\") pod \"swift-ring-rebalance-debug-rnhnd\" (UID: \"f8cd3dfb-7ce4-4c22-9044-e483f38131b9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rnhnd" Mar 07 15:13:34 crc kubenswrapper[4943]: I0307 15:13:34.520805 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f8cd3dfb-7ce4-4c22-9044-e483f38131b9-etc-swift\") pod \"swift-ring-rebalance-debug-rnhnd\" (UID: \"f8cd3dfb-7ce4-4c22-9044-e483f38131b9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rnhnd" Mar 07 15:13:34 crc kubenswrapper[4943]: I0307 15:13:34.521365 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f8cd3dfb-7ce4-4c22-9044-e483f38131b9-ring-data-devices\") pod \"swift-ring-rebalance-debug-rnhnd\" (UID: \"f8cd3dfb-7ce4-4c22-9044-e483f38131b9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rnhnd" Mar 07 15:13:34 crc kubenswrapper[4943]: I0307 15:13:34.521581 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f8cd3dfb-7ce4-4c22-9044-e483f38131b9-scripts\") pod \"swift-ring-rebalance-debug-rnhnd\" (UID: \"f8cd3dfb-7ce4-4c22-9044-e483f38131b9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rnhnd" Mar 07 15:13:34 crc kubenswrapper[4943]: I0307 15:13:34.528401 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f8cd3dfb-7ce4-4c22-9044-e483f38131b9-dispersionconf\") pod \"swift-ring-rebalance-debug-rnhnd\" (UID: \"f8cd3dfb-7ce4-4c22-9044-e483f38131b9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rnhnd" Mar 07 15:13:34 crc kubenswrapper[4943]: I0307 15:13:34.533718 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f8cd3dfb-7ce4-4c22-9044-e483f38131b9-swiftconf\") pod \"swift-ring-rebalance-debug-rnhnd\" (UID: \"f8cd3dfb-7ce4-4c22-9044-e483f38131b9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rnhnd" Mar 07 15:13:34 crc kubenswrapper[4943]: I0307 15:13:34.547855 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mkkr7\" (UniqueName: \"kubernetes.io/projected/f8cd3dfb-7ce4-4c22-9044-e483f38131b9-kube-api-access-mkkr7\") pod \"swift-ring-rebalance-debug-rnhnd\" (UID: \"f8cd3dfb-7ce4-4c22-9044-e483f38131b9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rnhnd" Mar 07 15:13:34 crc kubenswrapper[4943]: I0307 15:13:34.581583 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-rnhnd" Mar 07 15:13:34 crc kubenswrapper[4943]: I0307 15:13:34.766315 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3247c7e7-b192-4f48-9f6c-494dd33578c6" path="/var/lib/kubelet/pods/3247c7e7-b192-4f48-9f6c-494dd33578c6/volumes" Mar 07 15:13:34 crc kubenswrapper[4943]: I0307 15:13:34.874611 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-rnhnd"] Mar 07 15:13:34 crc kubenswrapper[4943]: W0307 15:13:34.886118 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf8cd3dfb_7ce4_4c22_9044_e483f38131b9.slice/crio-a77b51e8a9ef837a7fb680192feed052645afb6bea486028b287663e56a15f19 WatchSource:0}: Error finding container a77b51e8a9ef837a7fb680192feed052645afb6bea486028b287663e56a15f19: Status 404 returned error can't find the container with id a77b51e8a9ef837a7fb680192feed052645afb6bea486028b287663e56a15f19 Mar 07 15:13:35 crc kubenswrapper[4943]: I0307 15:13:35.740530 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-rnhnd" event={"ID":"f8cd3dfb-7ce4-4c22-9044-e483f38131b9","Type":"ContainerStarted","Data":"591eb6dd9de2b9f9bb51c5010145c2bbc85e0db9ba0d105ec7c263b7cf1eea2c"} Mar 07 15:13:35 crc kubenswrapper[4943]: I0307 15:13:35.740946 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-rnhnd" event={"ID":"f8cd3dfb-7ce4-4c22-9044-e483f38131b9","Type":"ContainerStarted","Data":"a77b51e8a9ef837a7fb680192feed052645afb6bea486028b287663e56a15f19"} Mar 07 15:13:35 crc kubenswrapper[4943]: I0307 15:13:35.767670 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-rnhnd" podStartSLOduration=1.767643032 podStartE2EDuration="1.767643032s" podCreationTimestamp="2026-03-07 15:13:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:13:35.766802522 +0000 UTC m=+2057.718939060" watchObservedRunningTime="2026-03-07 15:13:35.767643032 +0000 UTC m=+2057.719779570" Mar 07 15:13:36 crc kubenswrapper[4943]: I0307 15:13:36.073773 4943 patch_prober.go:28] interesting pod/machine-config-daemon-cpdmm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 07 15:13:36 crc kubenswrapper[4943]: I0307 15:13:36.073839 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 07 15:13:36 crc kubenswrapper[4943]: I0307 15:13:36.073917 4943 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" Mar 07 15:13:36 crc kubenswrapper[4943]: I0307 15:13:36.074612 4943 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"796971a8f0687fda2fdc2038d9ae78273781f30a32f7c084798a95ee29d28df7"} pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 07 15:13:36 crc kubenswrapper[4943]: I0307 15:13:36.074683 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerName="machine-config-daemon" containerID="cri-o://796971a8f0687fda2fdc2038d9ae78273781f30a32f7c084798a95ee29d28df7" gracePeriod=600 Mar 07 15:13:36 crc kubenswrapper[4943]: I0307 15:13:36.751555 4943 generic.go:334] "Generic (PLEG): container finished" podID="f8cd3dfb-7ce4-4c22-9044-e483f38131b9" containerID="591eb6dd9de2b9f9bb51c5010145c2bbc85e0db9ba0d105ec7c263b7cf1eea2c" exitCode=0 Mar 07 15:13:36 crc kubenswrapper[4943]: I0307 15:13:36.751740 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-rnhnd" event={"ID":"f8cd3dfb-7ce4-4c22-9044-e483f38131b9","Type":"ContainerDied","Data":"591eb6dd9de2b9f9bb51c5010145c2bbc85e0db9ba0d105ec7c263b7cf1eea2c"} Mar 07 15:13:36 crc kubenswrapper[4943]: I0307 15:13:36.759994 4943 generic.go:334] "Generic (PLEG): container finished" podID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerID="796971a8f0687fda2fdc2038d9ae78273781f30a32f7c084798a95ee29d28df7" exitCode=0 Mar 07 15:13:36 crc kubenswrapper[4943]: I0307 15:13:36.763630 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" event={"ID":"4c1f18ea-e557-41fc-95dd-ce4e1722a61e","Type":"ContainerDied","Data":"796971a8f0687fda2fdc2038d9ae78273781f30a32f7c084798a95ee29d28df7"} Mar 07 15:13:36 crc kubenswrapper[4943]: I0307 15:13:36.763685 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" event={"ID":"4c1f18ea-e557-41fc-95dd-ce4e1722a61e","Type":"ContainerStarted","Data":"448308ad6e34ac922d66b3c90e66d0927407a94dd28633ec350502d69e4ca8aa"} Mar 07 15:13:36 crc kubenswrapper[4943]: I0307 15:13:36.763706 4943 scope.go:117] "RemoveContainer" containerID="26f35e1f941cca0c34e3650d7aa047834e769dd76d2a61c3a39cad5bb2d090ae" Mar 07 15:13:38 crc kubenswrapper[4943]: I0307 15:13:38.121105 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-rnhnd" Mar 07 15:13:38 crc kubenswrapper[4943]: I0307 15:13:38.165401 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-rnhnd"] Mar 07 15:13:38 crc kubenswrapper[4943]: I0307 15:13:38.174865 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-rnhnd"] Mar 07 15:13:38 crc kubenswrapper[4943]: I0307 15:13:38.278137 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f8cd3dfb-7ce4-4c22-9044-e483f38131b9-ring-data-devices\") pod \"f8cd3dfb-7ce4-4c22-9044-e483f38131b9\" (UID: \"f8cd3dfb-7ce4-4c22-9044-e483f38131b9\") " Mar 07 15:13:38 crc kubenswrapper[4943]: I0307 15:13:38.278210 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f8cd3dfb-7ce4-4c22-9044-e483f38131b9-dispersionconf\") pod \"f8cd3dfb-7ce4-4c22-9044-e483f38131b9\" (UID: \"f8cd3dfb-7ce4-4c22-9044-e483f38131b9\") " Mar 07 15:13:38 crc kubenswrapper[4943]: I0307 15:13:38.278261 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f8cd3dfb-7ce4-4c22-9044-e483f38131b9-scripts\") pod \"f8cd3dfb-7ce4-4c22-9044-e483f38131b9\" (UID: \"f8cd3dfb-7ce4-4c22-9044-e483f38131b9\") " Mar 07 15:13:38 crc kubenswrapper[4943]: I0307 15:13:38.279261 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f8cd3dfb-7ce4-4c22-9044-e483f38131b9-swiftconf\") pod \"f8cd3dfb-7ce4-4c22-9044-e483f38131b9\" (UID: \"f8cd3dfb-7ce4-4c22-9044-e483f38131b9\") " Mar 07 15:13:38 crc kubenswrapper[4943]: I0307 15:13:38.279306 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f8cd3dfb-7ce4-4c22-9044-e483f38131b9-etc-swift\") pod \"f8cd3dfb-7ce4-4c22-9044-e483f38131b9\" (UID: \"f8cd3dfb-7ce4-4c22-9044-e483f38131b9\") " Mar 07 15:13:38 crc kubenswrapper[4943]: I0307 15:13:38.279359 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mkkr7\" (UniqueName: \"kubernetes.io/projected/f8cd3dfb-7ce4-4c22-9044-e483f38131b9-kube-api-access-mkkr7\") pod \"f8cd3dfb-7ce4-4c22-9044-e483f38131b9\" (UID: \"f8cd3dfb-7ce4-4c22-9044-e483f38131b9\") " Mar 07 15:13:38 crc kubenswrapper[4943]: I0307 15:13:38.279596 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f8cd3dfb-7ce4-4c22-9044-e483f38131b9-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "f8cd3dfb-7ce4-4c22-9044-e483f38131b9" (UID: "f8cd3dfb-7ce4-4c22-9044-e483f38131b9"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:13:38 crc kubenswrapper[4943]: I0307 15:13:38.279919 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f8cd3dfb-7ce4-4c22-9044-e483f38131b9-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:38 crc kubenswrapper[4943]: I0307 15:13:38.280461 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8cd3dfb-7ce4-4c22-9044-e483f38131b9-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "f8cd3dfb-7ce4-4c22-9044-e483f38131b9" (UID: "f8cd3dfb-7ce4-4c22-9044-e483f38131b9"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:13:38 crc kubenswrapper[4943]: I0307 15:13:38.284953 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8cd3dfb-7ce4-4c22-9044-e483f38131b9-kube-api-access-mkkr7" (OuterVolumeSpecName: "kube-api-access-mkkr7") pod "f8cd3dfb-7ce4-4c22-9044-e483f38131b9" (UID: "f8cd3dfb-7ce4-4c22-9044-e483f38131b9"). InnerVolumeSpecName "kube-api-access-mkkr7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:13:38 crc kubenswrapper[4943]: I0307 15:13:38.310514 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8cd3dfb-7ce4-4c22-9044-e483f38131b9-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "f8cd3dfb-7ce4-4c22-9044-e483f38131b9" (UID: "f8cd3dfb-7ce4-4c22-9044-e483f38131b9"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:13:38 crc kubenswrapper[4943]: I0307 15:13:38.318741 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8cd3dfb-7ce4-4c22-9044-e483f38131b9-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "f8cd3dfb-7ce4-4c22-9044-e483f38131b9" (UID: "f8cd3dfb-7ce4-4c22-9044-e483f38131b9"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:13:38 crc kubenswrapper[4943]: I0307 15:13:38.323325 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f8cd3dfb-7ce4-4c22-9044-e483f38131b9-scripts" (OuterVolumeSpecName: "scripts") pod "f8cd3dfb-7ce4-4c22-9044-e483f38131b9" (UID: "f8cd3dfb-7ce4-4c22-9044-e483f38131b9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:13:38 crc kubenswrapper[4943]: I0307 15:13:38.381783 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f8cd3dfb-7ce4-4c22-9044-e483f38131b9-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:38 crc kubenswrapper[4943]: I0307 15:13:38.381829 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f8cd3dfb-7ce4-4c22-9044-e483f38131b9-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:38 crc kubenswrapper[4943]: I0307 15:13:38.381847 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f8cd3dfb-7ce4-4c22-9044-e483f38131b9-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:38 crc kubenswrapper[4943]: I0307 15:13:38.381864 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f8cd3dfb-7ce4-4c22-9044-e483f38131b9-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:38 crc kubenswrapper[4943]: I0307 15:13:38.381881 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mkkr7\" (UniqueName: \"kubernetes.io/projected/f8cd3dfb-7ce4-4c22-9044-e483f38131b9-kube-api-access-mkkr7\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:38 crc kubenswrapper[4943]: I0307 15:13:38.787604 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8cd3dfb-7ce4-4c22-9044-e483f38131b9" path="/var/lib/kubelet/pods/f8cd3dfb-7ce4-4c22-9044-e483f38131b9/volumes" Mar 07 15:13:38 crc kubenswrapper[4943]: I0307 15:13:38.802359 4943 scope.go:117] "RemoveContainer" containerID="591eb6dd9de2b9f9bb51c5010145c2bbc85e0db9ba0d105ec7c263b7cf1eea2c" Mar 07 15:13:38 crc kubenswrapper[4943]: I0307 15:13:38.802525 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-rnhnd" Mar 07 15:13:39 crc kubenswrapper[4943]: I0307 15:13:39.385327 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-fwzm9"] Mar 07 15:13:39 crc kubenswrapper[4943]: E0307 15:13:39.385856 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8cd3dfb-7ce4-4c22-9044-e483f38131b9" containerName="swift-ring-rebalance" Mar 07 15:13:39 crc kubenswrapper[4943]: I0307 15:13:39.385876 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8cd3dfb-7ce4-4c22-9044-e483f38131b9" containerName="swift-ring-rebalance" Mar 07 15:13:39 crc kubenswrapper[4943]: I0307 15:13:39.386129 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8cd3dfb-7ce4-4c22-9044-e483f38131b9" containerName="swift-ring-rebalance" Mar 07 15:13:39 crc kubenswrapper[4943]: I0307 15:13:39.386871 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-fwzm9" Mar 07 15:13:39 crc kubenswrapper[4943]: I0307 15:13:39.390843 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:13:39 crc kubenswrapper[4943]: I0307 15:13:39.391378 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:13:39 crc kubenswrapper[4943]: I0307 15:13:39.412272 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-fwzm9"] Mar 07 15:13:39 crc kubenswrapper[4943]: I0307 15:13:39.503885 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/dffa497b-0eb1-46a8-ba2a-e2511cd59c6f-ring-data-devices\") pod \"swift-ring-rebalance-debug-fwzm9\" (UID: \"dffa497b-0eb1-46a8-ba2a-e2511cd59c6f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fwzm9" Mar 07 15:13:39 crc kubenswrapper[4943]: I0307 15:13:39.504021 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6dgk\" (UniqueName: \"kubernetes.io/projected/dffa497b-0eb1-46a8-ba2a-e2511cd59c6f-kube-api-access-t6dgk\") pod \"swift-ring-rebalance-debug-fwzm9\" (UID: \"dffa497b-0eb1-46a8-ba2a-e2511cd59c6f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fwzm9" Mar 07 15:13:39 crc kubenswrapper[4943]: I0307 15:13:39.504123 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dffa497b-0eb1-46a8-ba2a-e2511cd59c6f-scripts\") pod \"swift-ring-rebalance-debug-fwzm9\" (UID: \"dffa497b-0eb1-46a8-ba2a-e2511cd59c6f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fwzm9" Mar 07 15:13:39 crc kubenswrapper[4943]: I0307 15:13:39.504182 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/dffa497b-0eb1-46a8-ba2a-e2511cd59c6f-swiftconf\") pod \"swift-ring-rebalance-debug-fwzm9\" (UID: \"dffa497b-0eb1-46a8-ba2a-e2511cd59c6f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fwzm9" Mar 07 15:13:39 crc kubenswrapper[4943]: I0307 15:13:39.504256 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/dffa497b-0eb1-46a8-ba2a-e2511cd59c6f-etc-swift\") pod \"swift-ring-rebalance-debug-fwzm9\" (UID: \"dffa497b-0eb1-46a8-ba2a-e2511cd59c6f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fwzm9" Mar 07 15:13:39 crc kubenswrapper[4943]: I0307 15:13:39.504290 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/dffa497b-0eb1-46a8-ba2a-e2511cd59c6f-dispersionconf\") pod \"swift-ring-rebalance-debug-fwzm9\" (UID: \"dffa497b-0eb1-46a8-ba2a-e2511cd59c6f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fwzm9" Mar 07 15:13:39 crc kubenswrapper[4943]: I0307 15:13:39.605718 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dffa497b-0eb1-46a8-ba2a-e2511cd59c6f-scripts\") pod \"swift-ring-rebalance-debug-fwzm9\" (UID: \"dffa497b-0eb1-46a8-ba2a-e2511cd59c6f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fwzm9" Mar 07 15:13:39 crc kubenswrapper[4943]: I0307 15:13:39.605798 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/dffa497b-0eb1-46a8-ba2a-e2511cd59c6f-swiftconf\") pod \"swift-ring-rebalance-debug-fwzm9\" (UID: \"dffa497b-0eb1-46a8-ba2a-e2511cd59c6f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fwzm9" Mar 07 15:13:39 crc kubenswrapper[4943]: I0307 15:13:39.605880 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/dffa497b-0eb1-46a8-ba2a-e2511cd59c6f-etc-swift\") pod \"swift-ring-rebalance-debug-fwzm9\" (UID: \"dffa497b-0eb1-46a8-ba2a-e2511cd59c6f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fwzm9" Mar 07 15:13:39 crc kubenswrapper[4943]: I0307 15:13:39.606004 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/dffa497b-0eb1-46a8-ba2a-e2511cd59c6f-dispersionconf\") pod \"swift-ring-rebalance-debug-fwzm9\" (UID: \"dffa497b-0eb1-46a8-ba2a-e2511cd59c6f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fwzm9" Mar 07 15:13:39 crc kubenswrapper[4943]: I0307 15:13:39.606138 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/dffa497b-0eb1-46a8-ba2a-e2511cd59c6f-ring-data-devices\") pod \"swift-ring-rebalance-debug-fwzm9\" (UID: \"dffa497b-0eb1-46a8-ba2a-e2511cd59c6f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fwzm9" Mar 07 15:13:39 crc kubenswrapper[4943]: I0307 15:13:39.606212 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6dgk\" (UniqueName: \"kubernetes.io/projected/dffa497b-0eb1-46a8-ba2a-e2511cd59c6f-kube-api-access-t6dgk\") pod \"swift-ring-rebalance-debug-fwzm9\" (UID: \"dffa497b-0eb1-46a8-ba2a-e2511cd59c6f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fwzm9" Mar 07 15:13:39 crc kubenswrapper[4943]: I0307 15:13:39.606706 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/dffa497b-0eb1-46a8-ba2a-e2511cd59c6f-etc-swift\") pod \"swift-ring-rebalance-debug-fwzm9\" (UID: \"dffa497b-0eb1-46a8-ba2a-e2511cd59c6f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fwzm9" Mar 07 15:13:39 crc kubenswrapper[4943]: I0307 15:13:39.607027 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dffa497b-0eb1-46a8-ba2a-e2511cd59c6f-scripts\") pod \"swift-ring-rebalance-debug-fwzm9\" (UID: \"dffa497b-0eb1-46a8-ba2a-e2511cd59c6f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fwzm9" Mar 07 15:13:39 crc kubenswrapper[4943]: I0307 15:13:39.607654 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/dffa497b-0eb1-46a8-ba2a-e2511cd59c6f-ring-data-devices\") pod \"swift-ring-rebalance-debug-fwzm9\" (UID: \"dffa497b-0eb1-46a8-ba2a-e2511cd59c6f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fwzm9" Mar 07 15:13:39 crc kubenswrapper[4943]: I0307 15:13:39.613386 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/dffa497b-0eb1-46a8-ba2a-e2511cd59c6f-dispersionconf\") pod \"swift-ring-rebalance-debug-fwzm9\" (UID: \"dffa497b-0eb1-46a8-ba2a-e2511cd59c6f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fwzm9" Mar 07 15:13:39 crc kubenswrapper[4943]: I0307 15:13:39.615808 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/dffa497b-0eb1-46a8-ba2a-e2511cd59c6f-swiftconf\") pod \"swift-ring-rebalance-debug-fwzm9\" (UID: \"dffa497b-0eb1-46a8-ba2a-e2511cd59c6f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fwzm9" Mar 07 15:13:39 crc kubenswrapper[4943]: I0307 15:13:39.639801 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6dgk\" (UniqueName: \"kubernetes.io/projected/dffa497b-0eb1-46a8-ba2a-e2511cd59c6f-kube-api-access-t6dgk\") pod \"swift-ring-rebalance-debug-fwzm9\" (UID: \"dffa497b-0eb1-46a8-ba2a-e2511cd59c6f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fwzm9" Mar 07 15:13:39 crc kubenswrapper[4943]: I0307 15:13:39.709564 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-fwzm9" Mar 07 15:13:40 crc kubenswrapper[4943]: I0307 15:13:40.186555 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-fwzm9"] Mar 07 15:13:40 crc kubenswrapper[4943]: I0307 15:13:40.831294 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-fwzm9" event={"ID":"dffa497b-0eb1-46a8-ba2a-e2511cd59c6f","Type":"ContainerStarted","Data":"2800f30d0d51f86586364cf6aa464f7fe6d5bca06dd80bd35c8d82cccb70b9a3"} Mar 07 15:13:40 crc kubenswrapper[4943]: I0307 15:13:40.831819 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-fwzm9" event={"ID":"dffa497b-0eb1-46a8-ba2a-e2511cd59c6f","Type":"ContainerStarted","Data":"a66b3dad5c0751eb1c169a5b3b2c4a4c829d46c31e5470e7e44923a07c4f8f56"} Mar 07 15:13:40 crc kubenswrapper[4943]: I0307 15:13:40.858141 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-fwzm9" podStartSLOduration=1.858113178 podStartE2EDuration="1.858113178s" podCreationTimestamp="2026-03-07 15:13:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:13:40.847432236 +0000 UTC m=+2062.799568764" watchObservedRunningTime="2026-03-07 15:13:40.858113178 +0000 UTC m=+2062.810249706" Mar 07 15:13:41 crc kubenswrapper[4943]: I0307 15:13:41.859385 4943 generic.go:334] "Generic (PLEG): container finished" podID="dffa497b-0eb1-46a8-ba2a-e2511cd59c6f" containerID="2800f30d0d51f86586364cf6aa464f7fe6d5bca06dd80bd35c8d82cccb70b9a3" exitCode=0 Mar 07 15:13:41 crc kubenswrapper[4943]: I0307 15:13:41.859454 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-fwzm9" event={"ID":"dffa497b-0eb1-46a8-ba2a-e2511cd59c6f","Type":"ContainerDied","Data":"2800f30d0d51f86586364cf6aa464f7fe6d5bca06dd80bd35c8d82cccb70b9a3"} Mar 07 15:13:43 crc kubenswrapper[4943]: I0307 15:13:43.200978 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-fwzm9" Mar 07 15:13:43 crc kubenswrapper[4943]: I0307 15:13:43.244310 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-fwzm9"] Mar 07 15:13:43 crc kubenswrapper[4943]: I0307 15:13:43.253514 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-fwzm9"] Mar 07 15:13:43 crc kubenswrapper[4943]: I0307 15:13:43.371823 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/dffa497b-0eb1-46a8-ba2a-e2511cd59c6f-swiftconf\") pod \"dffa497b-0eb1-46a8-ba2a-e2511cd59c6f\" (UID: \"dffa497b-0eb1-46a8-ba2a-e2511cd59c6f\") " Mar 07 15:13:43 crc kubenswrapper[4943]: I0307 15:13:43.371901 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/dffa497b-0eb1-46a8-ba2a-e2511cd59c6f-dispersionconf\") pod \"dffa497b-0eb1-46a8-ba2a-e2511cd59c6f\" (UID: \"dffa497b-0eb1-46a8-ba2a-e2511cd59c6f\") " Mar 07 15:13:43 crc kubenswrapper[4943]: I0307 15:13:43.371976 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t6dgk\" (UniqueName: \"kubernetes.io/projected/dffa497b-0eb1-46a8-ba2a-e2511cd59c6f-kube-api-access-t6dgk\") pod \"dffa497b-0eb1-46a8-ba2a-e2511cd59c6f\" (UID: \"dffa497b-0eb1-46a8-ba2a-e2511cd59c6f\") " Mar 07 15:13:43 crc kubenswrapper[4943]: I0307 15:13:43.372052 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/dffa497b-0eb1-46a8-ba2a-e2511cd59c6f-etc-swift\") pod \"dffa497b-0eb1-46a8-ba2a-e2511cd59c6f\" (UID: \"dffa497b-0eb1-46a8-ba2a-e2511cd59c6f\") " Mar 07 15:13:43 crc kubenswrapper[4943]: I0307 15:13:43.372955 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/dffa497b-0eb1-46a8-ba2a-e2511cd59c6f-ring-data-devices\") pod \"dffa497b-0eb1-46a8-ba2a-e2511cd59c6f\" (UID: \"dffa497b-0eb1-46a8-ba2a-e2511cd59c6f\") " Mar 07 15:13:43 crc kubenswrapper[4943]: I0307 15:13:43.373008 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dffa497b-0eb1-46a8-ba2a-e2511cd59c6f-scripts\") pod \"dffa497b-0eb1-46a8-ba2a-e2511cd59c6f\" (UID: \"dffa497b-0eb1-46a8-ba2a-e2511cd59c6f\") " Mar 07 15:13:43 crc kubenswrapper[4943]: I0307 15:13:43.374027 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dffa497b-0eb1-46a8-ba2a-e2511cd59c6f-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "dffa497b-0eb1-46a8-ba2a-e2511cd59c6f" (UID: "dffa497b-0eb1-46a8-ba2a-e2511cd59c6f"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:13:43 crc kubenswrapper[4943]: I0307 15:13:43.374072 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dffa497b-0eb1-46a8-ba2a-e2511cd59c6f-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "dffa497b-0eb1-46a8-ba2a-e2511cd59c6f" (UID: "dffa497b-0eb1-46a8-ba2a-e2511cd59c6f"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:13:43 crc kubenswrapper[4943]: I0307 15:13:43.379108 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dffa497b-0eb1-46a8-ba2a-e2511cd59c6f-kube-api-access-t6dgk" (OuterVolumeSpecName: "kube-api-access-t6dgk") pod "dffa497b-0eb1-46a8-ba2a-e2511cd59c6f" (UID: "dffa497b-0eb1-46a8-ba2a-e2511cd59c6f"). InnerVolumeSpecName "kube-api-access-t6dgk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:13:43 crc kubenswrapper[4943]: I0307 15:13:43.409914 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dffa497b-0eb1-46a8-ba2a-e2511cd59c6f-scripts" (OuterVolumeSpecName: "scripts") pod "dffa497b-0eb1-46a8-ba2a-e2511cd59c6f" (UID: "dffa497b-0eb1-46a8-ba2a-e2511cd59c6f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:13:43 crc kubenswrapper[4943]: I0307 15:13:43.410118 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dffa497b-0eb1-46a8-ba2a-e2511cd59c6f-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "dffa497b-0eb1-46a8-ba2a-e2511cd59c6f" (UID: "dffa497b-0eb1-46a8-ba2a-e2511cd59c6f"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:13:43 crc kubenswrapper[4943]: I0307 15:13:43.415830 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dffa497b-0eb1-46a8-ba2a-e2511cd59c6f-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "dffa497b-0eb1-46a8-ba2a-e2511cd59c6f" (UID: "dffa497b-0eb1-46a8-ba2a-e2511cd59c6f"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:13:43 crc kubenswrapper[4943]: I0307 15:13:43.475003 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t6dgk\" (UniqueName: \"kubernetes.io/projected/dffa497b-0eb1-46a8-ba2a-e2511cd59c6f-kube-api-access-t6dgk\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:43 crc kubenswrapper[4943]: I0307 15:13:43.475038 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/dffa497b-0eb1-46a8-ba2a-e2511cd59c6f-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:43 crc kubenswrapper[4943]: I0307 15:13:43.475051 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/dffa497b-0eb1-46a8-ba2a-e2511cd59c6f-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:43 crc kubenswrapper[4943]: I0307 15:13:43.475064 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dffa497b-0eb1-46a8-ba2a-e2511cd59c6f-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:43 crc kubenswrapper[4943]: I0307 15:13:43.475076 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/dffa497b-0eb1-46a8-ba2a-e2511cd59c6f-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:43 crc kubenswrapper[4943]: I0307 15:13:43.475087 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/dffa497b-0eb1-46a8-ba2a-e2511cd59c6f-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:43 crc kubenswrapper[4943]: I0307 15:13:43.880519 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a66b3dad5c0751eb1c169a5b3b2c4a4c829d46c31e5470e7e44923a07c4f8f56" Mar 07 15:13:43 crc kubenswrapper[4943]: I0307 15:13:43.880588 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-fwzm9" Mar 07 15:13:44 crc kubenswrapper[4943]: I0307 15:13:44.375765 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-flsvk"] Mar 07 15:13:44 crc kubenswrapper[4943]: E0307 15:13:44.376156 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dffa497b-0eb1-46a8-ba2a-e2511cd59c6f" containerName="swift-ring-rebalance" Mar 07 15:13:44 crc kubenswrapper[4943]: I0307 15:13:44.376187 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="dffa497b-0eb1-46a8-ba2a-e2511cd59c6f" containerName="swift-ring-rebalance" Mar 07 15:13:44 crc kubenswrapper[4943]: I0307 15:13:44.376380 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="dffa497b-0eb1-46a8-ba2a-e2511cd59c6f" containerName="swift-ring-rebalance" Mar 07 15:13:44 crc kubenswrapper[4943]: I0307 15:13:44.377044 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-flsvk" Mar 07 15:13:44 crc kubenswrapper[4943]: I0307 15:13:44.381234 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:13:44 crc kubenswrapper[4943]: I0307 15:13:44.393082 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:13:44 crc kubenswrapper[4943]: I0307 15:13:44.396671 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-flsvk"] Mar 07 15:13:44 crc kubenswrapper[4943]: I0307 15:13:44.493098 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11-ring-data-devices\") pod \"swift-ring-rebalance-debug-flsvk\" (UID: \"60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-flsvk" Mar 07 15:13:44 crc kubenswrapper[4943]: I0307 15:13:44.493216 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11-scripts\") pod \"swift-ring-rebalance-debug-flsvk\" (UID: \"60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-flsvk" Mar 07 15:13:44 crc kubenswrapper[4943]: I0307 15:13:44.493285 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6bpl\" (UniqueName: \"kubernetes.io/projected/60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11-kube-api-access-x6bpl\") pod \"swift-ring-rebalance-debug-flsvk\" (UID: \"60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-flsvk" Mar 07 15:13:44 crc kubenswrapper[4943]: I0307 15:13:44.493413 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11-etc-swift\") pod \"swift-ring-rebalance-debug-flsvk\" (UID: \"60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-flsvk" Mar 07 15:13:44 crc kubenswrapper[4943]: I0307 15:13:44.493520 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11-swiftconf\") pod \"swift-ring-rebalance-debug-flsvk\" (UID: \"60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-flsvk" Mar 07 15:13:44 crc kubenswrapper[4943]: I0307 15:13:44.493571 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11-dispersionconf\") pod \"swift-ring-rebalance-debug-flsvk\" (UID: \"60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-flsvk" Mar 07 15:13:44 crc kubenswrapper[4943]: I0307 15:13:44.595522 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11-ring-data-devices\") pod \"swift-ring-rebalance-debug-flsvk\" (UID: \"60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-flsvk" Mar 07 15:13:44 crc kubenswrapper[4943]: I0307 15:13:44.595626 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11-scripts\") pod \"swift-ring-rebalance-debug-flsvk\" (UID: \"60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-flsvk" Mar 07 15:13:44 crc kubenswrapper[4943]: I0307 15:13:44.595673 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6bpl\" (UniqueName: \"kubernetes.io/projected/60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11-kube-api-access-x6bpl\") pod \"swift-ring-rebalance-debug-flsvk\" (UID: \"60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-flsvk" Mar 07 15:13:44 crc kubenswrapper[4943]: I0307 15:13:44.595780 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11-etc-swift\") pod \"swift-ring-rebalance-debug-flsvk\" (UID: \"60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-flsvk" Mar 07 15:13:44 crc kubenswrapper[4943]: I0307 15:13:44.595843 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11-swiftconf\") pod \"swift-ring-rebalance-debug-flsvk\" (UID: \"60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-flsvk" Mar 07 15:13:44 crc kubenswrapper[4943]: I0307 15:13:44.595890 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11-dispersionconf\") pod \"swift-ring-rebalance-debug-flsvk\" (UID: \"60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-flsvk" Mar 07 15:13:44 crc kubenswrapper[4943]: I0307 15:13:44.596697 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11-scripts\") pod \"swift-ring-rebalance-debug-flsvk\" (UID: \"60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-flsvk" Mar 07 15:13:44 crc kubenswrapper[4943]: I0307 15:13:44.596792 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11-etc-swift\") pod \"swift-ring-rebalance-debug-flsvk\" (UID: \"60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-flsvk" Mar 07 15:13:44 crc kubenswrapper[4943]: I0307 15:13:44.597299 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11-ring-data-devices\") pod \"swift-ring-rebalance-debug-flsvk\" (UID: \"60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-flsvk" Mar 07 15:13:44 crc kubenswrapper[4943]: I0307 15:13:44.600101 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11-dispersionconf\") pod \"swift-ring-rebalance-debug-flsvk\" (UID: \"60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-flsvk" Mar 07 15:13:44 crc kubenswrapper[4943]: I0307 15:13:44.602742 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11-swiftconf\") pod \"swift-ring-rebalance-debug-flsvk\" (UID: \"60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-flsvk" Mar 07 15:13:44 crc kubenswrapper[4943]: I0307 15:13:44.630232 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6bpl\" (UniqueName: \"kubernetes.io/projected/60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11-kube-api-access-x6bpl\") pod \"swift-ring-rebalance-debug-flsvk\" (UID: \"60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-flsvk" Mar 07 15:13:44 crc kubenswrapper[4943]: I0307 15:13:44.702058 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-flsvk" Mar 07 15:13:44 crc kubenswrapper[4943]: I0307 15:13:44.771112 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dffa497b-0eb1-46a8-ba2a-e2511cd59c6f" path="/var/lib/kubelet/pods/dffa497b-0eb1-46a8-ba2a-e2511cd59c6f/volumes" Mar 07 15:13:45 crc kubenswrapper[4943]: I0307 15:13:45.208745 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-flsvk"] Mar 07 15:13:45 crc kubenswrapper[4943]: I0307 15:13:45.899897 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-flsvk" event={"ID":"60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11","Type":"ContainerStarted","Data":"993062ef4dbda6a60b17c0beb5ea3eb87f823ec5204863e96ab428657fe89fe1"} Mar 07 15:13:45 crc kubenswrapper[4943]: I0307 15:13:45.900191 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-flsvk" event={"ID":"60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11","Type":"ContainerStarted","Data":"5bd832da18b469da4ec1831b8fe0e6a9046a69663acbababf3699bbe754d34f0"} Mar 07 15:13:45 crc kubenswrapper[4943]: I0307 15:13:45.921255 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-flsvk" podStartSLOduration=1.9212396630000002 podStartE2EDuration="1.921239663s" podCreationTimestamp="2026-03-07 15:13:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:13:45.919131932 +0000 UTC m=+2067.871268430" watchObservedRunningTime="2026-03-07 15:13:45.921239663 +0000 UTC m=+2067.873376161" Mar 07 15:13:46 crc kubenswrapper[4943]: I0307 15:13:46.912038 4943 generic.go:334] "Generic (PLEG): container finished" podID="60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11" containerID="993062ef4dbda6a60b17c0beb5ea3eb87f823ec5204863e96ab428657fe89fe1" exitCode=0 Mar 07 15:13:46 crc kubenswrapper[4943]: I0307 15:13:46.912153 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-flsvk" event={"ID":"60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11","Type":"ContainerDied","Data":"993062ef4dbda6a60b17c0beb5ea3eb87f823ec5204863e96ab428657fe89fe1"} Mar 07 15:13:48 crc kubenswrapper[4943]: I0307 15:13:48.271335 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-flsvk" Mar 07 15:13:48 crc kubenswrapper[4943]: I0307 15:13:48.310937 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-flsvk"] Mar 07 15:13:48 crc kubenswrapper[4943]: I0307 15:13:48.318089 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-flsvk"] Mar 07 15:13:48 crc kubenswrapper[4943]: I0307 15:13:48.458050 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11-ring-data-devices\") pod \"60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11\" (UID: \"60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11\") " Mar 07 15:13:48 crc kubenswrapper[4943]: I0307 15:13:48.458148 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11-dispersionconf\") pod \"60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11\" (UID: \"60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11\") " Mar 07 15:13:48 crc kubenswrapper[4943]: I0307 15:13:48.458179 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6bpl\" (UniqueName: \"kubernetes.io/projected/60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11-kube-api-access-x6bpl\") pod \"60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11\" (UID: \"60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11\") " Mar 07 15:13:48 crc kubenswrapper[4943]: I0307 15:13:48.458209 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11-swiftconf\") pod \"60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11\" (UID: \"60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11\") " Mar 07 15:13:48 crc kubenswrapper[4943]: I0307 15:13:48.458253 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11-etc-swift\") pod \"60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11\" (UID: \"60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11\") " Mar 07 15:13:48 crc kubenswrapper[4943]: I0307 15:13:48.458283 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11-scripts\") pod \"60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11\" (UID: \"60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11\") " Mar 07 15:13:48 crc kubenswrapper[4943]: I0307 15:13:48.458896 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11" (UID: "60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:13:48 crc kubenswrapper[4943]: I0307 15:13:48.459018 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11" (UID: "60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:13:48 crc kubenswrapper[4943]: I0307 15:13:48.466871 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11-kube-api-access-x6bpl" (OuterVolumeSpecName: "kube-api-access-x6bpl") pod "60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11" (UID: "60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11"). InnerVolumeSpecName "kube-api-access-x6bpl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:13:48 crc kubenswrapper[4943]: I0307 15:13:48.485985 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11" (UID: "60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:13:48 crc kubenswrapper[4943]: I0307 15:13:48.491441 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11-scripts" (OuterVolumeSpecName: "scripts") pod "60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11" (UID: "60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:13:48 crc kubenswrapper[4943]: I0307 15:13:48.493056 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11" (UID: "60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:13:48 crc kubenswrapper[4943]: I0307 15:13:48.560215 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:48 crc kubenswrapper[4943]: I0307 15:13:48.560260 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6bpl\" (UniqueName: \"kubernetes.io/projected/60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11-kube-api-access-x6bpl\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:48 crc kubenswrapper[4943]: I0307 15:13:48.560281 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:48 crc kubenswrapper[4943]: I0307 15:13:48.560298 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:48 crc kubenswrapper[4943]: I0307 15:13:48.560315 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:48 crc kubenswrapper[4943]: I0307 15:13:48.560332 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:48 crc kubenswrapper[4943]: I0307 15:13:48.765556 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11" path="/var/lib/kubelet/pods/60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11/volumes" Mar 07 15:13:48 crc kubenswrapper[4943]: I0307 15:13:48.938230 4943 scope.go:117] "RemoveContainer" containerID="993062ef4dbda6a60b17c0beb5ea3eb87f823ec5204863e96ab428657fe89fe1" Mar 07 15:13:48 crc kubenswrapper[4943]: I0307 15:13:48.938330 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-flsvk" Mar 07 15:13:49 crc kubenswrapper[4943]: I0307 15:13:49.528388 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-jst4g"] Mar 07 15:13:49 crc kubenswrapper[4943]: E0307 15:13:49.530047 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11" containerName="swift-ring-rebalance" Mar 07 15:13:49 crc kubenswrapper[4943]: I0307 15:13:49.530210 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11" containerName="swift-ring-rebalance" Mar 07 15:13:49 crc kubenswrapper[4943]: I0307 15:13:49.530636 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="60d6ec23-efa9-45a8-b0b2-d96cfdbc4e11" containerName="swift-ring-rebalance" Mar 07 15:13:49 crc kubenswrapper[4943]: I0307 15:13:49.531549 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-jst4g" Mar 07 15:13:49 crc kubenswrapper[4943]: I0307 15:13:49.535641 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:13:49 crc kubenswrapper[4943]: I0307 15:13:49.535987 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:13:49 crc kubenswrapper[4943]: I0307 15:13:49.549269 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-jst4g"] Mar 07 15:13:49 crc kubenswrapper[4943]: I0307 15:13:49.678778 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/4b659c4f-8b3c-4caf-9dae-c0b150ac3480-swiftconf\") pod \"swift-ring-rebalance-debug-jst4g\" (UID: \"4b659c4f-8b3c-4caf-9dae-c0b150ac3480\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jst4g" Mar 07 15:13:49 crc kubenswrapper[4943]: I0307 15:13:49.678897 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ms9vt\" (UniqueName: \"kubernetes.io/projected/4b659c4f-8b3c-4caf-9dae-c0b150ac3480-kube-api-access-ms9vt\") pod \"swift-ring-rebalance-debug-jst4g\" (UID: \"4b659c4f-8b3c-4caf-9dae-c0b150ac3480\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jst4g" Mar 07 15:13:49 crc kubenswrapper[4943]: I0307 15:13:49.679032 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4b659c4f-8b3c-4caf-9dae-c0b150ac3480-scripts\") pod \"swift-ring-rebalance-debug-jst4g\" (UID: \"4b659c4f-8b3c-4caf-9dae-c0b150ac3480\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jst4g" Mar 07 15:13:49 crc kubenswrapper[4943]: I0307 15:13:49.679095 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/4b659c4f-8b3c-4caf-9dae-c0b150ac3480-etc-swift\") pod \"swift-ring-rebalance-debug-jst4g\" (UID: \"4b659c4f-8b3c-4caf-9dae-c0b150ac3480\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jst4g" Mar 07 15:13:49 crc kubenswrapper[4943]: I0307 15:13:49.679125 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/4b659c4f-8b3c-4caf-9dae-c0b150ac3480-dispersionconf\") pod \"swift-ring-rebalance-debug-jst4g\" (UID: \"4b659c4f-8b3c-4caf-9dae-c0b150ac3480\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jst4g" Mar 07 15:13:49 crc kubenswrapper[4943]: I0307 15:13:49.679409 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/4b659c4f-8b3c-4caf-9dae-c0b150ac3480-ring-data-devices\") pod \"swift-ring-rebalance-debug-jst4g\" (UID: \"4b659c4f-8b3c-4caf-9dae-c0b150ac3480\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jst4g" Mar 07 15:13:49 crc kubenswrapper[4943]: I0307 15:13:49.780857 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/4b659c4f-8b3c-4caf-9dae-c0b150ac3480-ring-data-devices\") pod \"swift-ring-rebalance-debug-jst4g\" (UID: \"4b659c4f-8b3c-4caf-9dae-c0b150ac3480\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jst4g" Mar 07 15:13:49 crc kubenswrapper[4943]: I0307 15:13:49.780968 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/4b659c4f-8b3c-4caf-9dae-c0b150ac3480-swiftconf\") pod \"swift-ring-rebalance-debug-jst4g\" (UID: \"4b659c4f-8b3c-4caf-9dae-c0b150ac3480\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jst4g" Mar 07 15:13:49 crc kubenswrapper[4943]: I0307 15:13:49.781036 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ms9vt\" (UniqueName: \"kubernetes.io/projected/4b659c4f-8b3c-4caf-9dae-c0b150ac3480-kube-api-access-ms9vt\") pod \"swift-ring-rebalance-debug-jst4g\" (UID: \"4b659c4f-8b3c-4caf-9dae-c0b150ac3480\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jst4g" Mar 07 15:13:49 crc kubenswrapper[4943]: I0307 15:13:49.781127 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4b659c4f-8b3c-4caf-9dae-c0b150ac3480-scripts\") pod \"swift-ring-rebalance-debug-jst4g\" (UID: \"4b659c4f-8b3c-4caf-9dae-c0b150ac3480\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jst4g" Mar 07 15:13:49 crc kubenswrapper[4943]: I0307 15:13:49.781211 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/4b659c4f-8b3c-4caf-9dae-c0b150ac3480-etc-swift\") pod \"swift-ring-rebalance-debug-jst4g\" (UID: \"4b659c4f-8b3c-4caf-9dae-c0b150ac3480\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jst4g" Mar 07 15:13:49 crc kubenswrapper[4943]: I0307 15:13:49.781255 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/4b659c4f-8b3c-4caf-9dae-c0b150ac3480-dispersionconf\") pod \"swift-ring-rebalance-debug-jst4g\" (UID: \"4b659c4f-8b3c-4caf-9dae-c0b150ac3480\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jst4g" Mar 07 15:13:49 crc kubenswrapper[4943]: I0307 15:13:49.781785 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/4b659c4f-8b3c-4caf-9dae-c0b150ac3480-etc-swift\") pod \"swift-ring-rebalance-debug-jst4g\" (UID: \"4b659c4f-8b3c-4caf-9dae-c0b150ac3480\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jst4g" Mar 07 15:13:49 crc kubenswrapper[4943]: I0307 15:13:49.782253 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4b659c4f-8b3c-4caf-9dae-c0b150ac3480-scripts\") pod \"swift-ring-rebalance-debug-jst4g\" (UID: \"4b659c4f-8b3c-4caf-9dae-c0b150ac3480\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jst4g" Mar 07 15:13:49 crc kubenswrapper[4943]: I0307 15:13:49.782472 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/4b659c4f-8b3c-4caf-9dae-c0b150ac3480-ring-data-devices\") pod \"swift-ring-rebalance-debug-jst4g\" (UID: \"4b659c4f-8b3c-4caf-9dae-c0b150ac3480\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jst4g" Mar 07 15:13:49 crc kubenswrapper[4943]: I0307 15:13:49.789403 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/4b659c4f-8b3c-4caf-9dae-c0b150ac3480-dispersionconf\") pod \"swift-ring-rebalance-debug-jst4g\" (UID: \"4b659c4f-8b3c-4caf-9dae-c0b150ac3480\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jst4g" Mar 07 15:13:49 crc kubenswrapper[4943]: I0307 15:13:49.794392 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/4b659c4f-8b3c-4caf-9dae-c0b150ac3480-swiftconf\") pod \"swift-ring-rebalance-debug-jst4g\" (UID: \"4b659c4f-8b3c-4caf-9dae-c0b150ac3480\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jst4g" Mar 07 15:13:49 crc kubenswrapper[4943]: I0307 15:13:49.817458 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ms9vt\" (UniqueName: \"kubernetes.io/projected/4b659c4f-8b3c-4caf-9dae-c0b150ac3480-kube-api-access-ms9vt\") pod \"swift-ring-rebalance-debug-jst4g\" (UID: \"4b659c4f-8b3c-4caf-9dae-c0b150ac3480\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jst4g" Mar 07 15:13:49 crc kubenswrapper[4943]: I0307 15:13:49.848494 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-jst4g" Mar 07 15:13:50 crc kubenswrapper[4943]: I0307 15:13:50.308778 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-jst4g"] Mar 07 15:13:50 crc kubenswrapper[4943]: I0307 15:13:50.962041 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-jst4g" event={"ID":"4b659c4f-8b3c-4caf-9dae-c0b150ac3480","Type":"ContainerStarted","Data":"b301320a41d48cca02200a810c67ad9d75cc4daaa821bf5cd1ad024d1933c225"} Mar 07 15:13:50 crc kubenswrapper[4943]: I0307 15:13:50.962091 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-jst4g" event={"ID":"4b659c4f-8b3c-4caf-9dae-c0b150ac3480","Type":"ContainerStarted","Data":"e78367e7acff4010e4b7dfbf4e33de34365d7057524d4a2826c18258e99dd6b0"} Mar 07 15:13:50 crc kubenswrapper[4943]: I0307 15:13:50.995369 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-jst4g" podStartSLOduration=1.995350267 podStartE2EDuration="1.995350267s" podCreationTimestamp="2026-03-07 15:13:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:13:50.989565165 +0000 UTC m=+2072.941701703" watchObservedRunningTime="2026-03-07 15:13:50.995350267 +0000 UTC m=+2072.947486775" Mar 07 15:13:51 crc kubenswrapper[4943]: I0307 15:13:51.976500 4943 generic.go:334] "Generic (PLEG): container finished" podID="4b659c4f-8b3c-4caf-9dae-c0b150ac3480" containerID="b301320a41d48cca02200a810c67ad9d75cc4daaa821bf5cd1ad024d1933c225" exitCode=0 Mar 07 15:13:51 crc kubenswrapper[4943]: I0307 15:13:51.976584 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-jst4g" event={"ID":"4b659c4f-8b3c-4caf-9dae-c0b150ac3480","Type":"ContainerDied","Data":"b301320a41d48cca02200a810c67ad9d75cc4daaa821bf5cd1ad024d1933c225"} Mar 07 15:13:53 crc kubenswrapper[4943]: I0307 15:13:53.301649 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-jst4g" Mar 07 15:13:53 crc kubenswrapper[4943]: I0307 15:13:53.342077 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-jst4g"] Mar 07 15:13:53 crc kubenswrapper[4943]: I0307 15:13:53.349592 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-jst4g"] Mar 07 15:13:53 crc kubenswrapper[4943]: I0307 15:13:53.433254 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4b659c4f-8b3c-4caf-9dae-c0b150ac3480-scripts\") pod \"4b659c4f-8b3c-4caf-9dae-c0b150ac3480\" (UID: \"4b659c4f-8b3c-4caf-9dae-c0b150ac3480\") " Mar 07 15:13:53 crc kubenswrapper[4943]: I0307 15:13:53.433373 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/4b659c4f-8b3c-4caf-9dae-c0b150ac3480-ring-data-devices\") pod \"4b659c4f-8b3c-4caf-9dae-c0b150ac3480\" (UID: \"4b659c4f-8b3c-4caf-9dae-c0b150ac3480\") " Mar 07 15:13:53 crc kubenswrapper[4943]: I0307 15:13:53.433603 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/4b659c4f-8b3c-4caf-9dae-c0b150ac3480-swiftconf\") pod \"4b659c4f-8b3c-4caf-9dae-c0b150ac3480\" (UID: \"4b659c4f-8b3c-4caf-9dae-c0b150ac3480\") " Mar 07 15:13:53 crc kubenswrapper[4943]: I0307 15:13:53.433728 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ms9vt\" (UniqueName: \"kubernetes.io/projected/4b659c4f-8b3c-4caf-9dae-c0b150ac3480-kube-api-access-ms9vt\") pod \"4b659c4f-8b3c-4caf-9dae-c0b150ac3480\" (UID: \"4b659c4f-8b3c-4caf-9dae-c0b150ac3480\") " Mar 07 15:13:53 crc kubenswrapper[4943]: I0307 15:13:53.434389 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/4b659c4f-8b3c-4caf-9dae-c0b150ac3480-etc-swift\") pod \"4b659c4f-8b3c-4caf-9dae-c0b150ac3480\" (UID: \"4b659c4f-8b3c-4caf-9dae-c0b150ac3480\") " Mar 07 15:13:53 crc kubenswrapper[4943]: I0307 15:13:53.434574 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b659c4f-8b3c-4caf-9dae-c0b150ac3480-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "4b659c4f-8b3c-4caf-9dae-c0b150ac3480" (UID: "4b659c4f-8b3c-4caf-9dae-c0b150ac3480"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:13:53 crc kubenswrapper[4943]: I0307 15:13:53.434634 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/4b659c4f-8b3c-4caf-9dae-c0b150ac3480-dispersionconf\") pod \"4b659c4f-8b3c-4caf-9dae-c0b150ac3480\" (UID: \"4b659c4f-8b3c-4caf-9dae-c0b150ac3480\") " Mar 07 15:13:53 crc kubenswrapper[4943]: I0307 15:13:53.435662 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/4b659c4f-8b3c-4caf-9dae-c0b150ac3480-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:53 crc kubenswrapper[4943]: I0307 15:13:53.437272 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b659c4f-8b3c-4caf-9dae-c0b150ac3480-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "4b659c4f-8b3c-4caf-9dae-c0b150ac3480" (UID: "4b659c4f-8b3c-4caf-9dae-c0b150ac3480"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:13:53 crc kubenswrapper[4943]: I0307 15:13:53.439245 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b659c4f-8b3c-4caf-9dae-c0b150ac3480-kube-api-access-ms9vt" (OuterVolumeSpecName: "kube-api-access-ms9vt") pod "4b659c4f-8b3c-4caf-9dae-c0b150ac3480" (UID: "4b659c4f-8b3c-4caf-9dae-c0b150ac3480"). InnerVolumeSpecName "kube-api-access-ms9vt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:13:53 crc kubenswrapper[4943]: I0307 15:13:53.459011 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b659c4f-8b3c-4caf-9dae-c0b150ac3480-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "4b659c4f-8b3c-4caf-9dae-c0b150ac3480" (UID: "4b659c4f-8b3c-4caf-9dae-c0b150ac3480"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:13:53 crc kubenswrapper[4943]: I0307 15:13:53.461700 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b659c4f-8b3c-4caf-9dae-c0b150ac3480-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "4b659c4f-8b3c-4caf-9dae-c0b150ac3480" (UID: "4b659c4f-8b3c-4caf-9dae-c0b150ac3480"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:13:53 crc kubenswrapper[4943]: I0307 15:13:53.464618 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b659c4f-8b3c-4caf-9dae-c0b150ac3480-scripts" (OuterVolumeSpecName: "scripts") pod "4b659c4f-8b3c-4caf-9dae-c0b150ac3480" (UID: "4b659c4f-8b3c-4caf-9dae-c0b150ac3480"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:13:53 crc kubenswrapper[4943]: I0307 15:13:53.536786 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/4b659c4f-8b3c-4caf-9dae-c0b150ac3480-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:53 crc kubenswrapper[4943]: I0307 15:13:53.536826 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ms9vt\" (UniqueName: \"kubernetes.io/projected/4b659c4f-8b3c-4caf-9dae-c0b150ac3480-kube-api-access-ms9vt\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:53 crc kubenswrapper[4943]: I0307 15:13:53.536836 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/4b659c4f-8b3c-4caf-9dae-c0b150ac3480-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:53 crc kubenswrapper[4943]: I0307 15:13:53.536844 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/4b659c4f-8b3c-4caf-9dae-c0b150ac3480-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:53 crc kubenswrapper[4943]: I0307 15:13:53.536854 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4b659c4f-8b3c-4caf-9dae-c0b150ac3480-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:53 crc kubenswrapper[4943]: I0307 15:13:53.999538 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e78367e7acff4010e4b7dfbf4e33de34365d7057524d4a2826c18258e99dd6b0" Mar 07 15:13:53 crc kubenswrapper[4943]: I0307 15:13:53.999622 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-jst4g" Mar 07 15:13:54 crc kubenswrapper[4943]: I0307 15:13:54.522911 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-m82qh"] Mar 07 15:13:54 crc kubenswrapper[4943]: E0307 15:13:54.523599 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b659c4f-8b3c-4caf-9dae-c0b150ac3480" containerName="swift-ring-rebalance" Mar 07 15:13:54 crc kubenswrapper[4943]: I0307 15:13:54.523624 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b659c4f-8b3c-4caf-9dae-c0b150ac3480" containerName="swift-ring-rebalance" Mar 07 15:13:54 crc kubenswrapper[4943]: I0307 15:13:54.523966 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b659c4f-8b3c-4caf-9dae-c0b150ac3480" containerName="swift-ring-rebalance" Mar 07 15:13:54 crc kubenswrapper[4943]: I0307 15:13:54.524834 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-m82qh" Mar 07 15:13:54 crc kubenswrapper[4943]: I0307 15:13:54.528692 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:13:54 crc kubenswrapper[4943]: I0307 15:13:54.529068 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:13:54 crc kubenswrapper[4943]: I0307 15:13:54.535494 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-m82qh"] Mar 07 15:13:54 crc kubenswrapper[4943]: I0307 15:13:54.656377 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f42dba37-e7f7-4657-8341-e5afa2ad256a-swiftconf\") pod \"swift-ring-rebalance-debug-m82qh\" (UID: \"f42dba37-e7f7-4657-8341-e5afa2ad256a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m82qh" Mar 07 15:13:54 crc kubenswrapper[4943]: I0307 15:13:54.656456 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f42dba37-e7f7-4657-8341-e5afa2ad256a-ring-data-devices\") pod \"swift-ring-rebalance-debug-m82qh\" (UID: \"f42dba37-e7f7-4657-8341-e5afa2ad256a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m82qh" Mar 07 15:13:54 crc kubenswrapper[4943]: I0307 15:13:54.656572 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f42dba37-e7f7-4657-8341-e5afa2ad256a-etc-swift\") pod \"swift-ring-rebalance-debug-m82qh\" (UID: \"f42dba37-e7f7-4657-8341-e5afa2ad256a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m82qh" Mar 07 15:13:54 crc kubenswrapper[4943]: I0307 15:13:54.656634 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f42dba37-e7f7-4657-8341-e5afa2ad256a-scripts\") pod \"swift-ring-rebalance-debug-m82qh\" (UID: \"f42dba37-e7f7-4657-8341-e5afa2ad256a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m82qh" Mar 07 15:13:54 crc kubenswrapper[4943]: I0307 15:13:54.656668 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f42dba37-e7f7-4657-8341-e5afa2ad256a-dispersionconf\") pod \"swift-ring-rebalance-debug-m82qh\" (UID: \"f42dba37-e7f7-4657-8341-e5afa2ad256a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m82qh" Mar 07 15:13:54 crc kubenswrapper[4943]: I0307 15:13:54.656703 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wwgd\" (UniqueName: \"kubernetes.io/projected/f42dba37-e7f7-4657-8341-e5afa2ad256a-kube-api-access-9wwgd\") pod \"swift-ring-rebalance-debug-m82qh\" (UID: \"f42dba37-e7f7-4657-8341-e5afa2ad256a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m82qh" Mar 07 15:13:54 crc kubenswrapper[4943]: I0307 15:13:54.758014 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f42dba37-e7f7-4657-8341-e5afa2ad256a-swiftconf\") pod \"swift-ring-rebalance-debug-m82qh\" (UID: \"f42dba37-e7f7-4657-8341-e5afa2ad256a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m82qh" Mar 07 15:13:54 crc kubenswrapper[4943]: I0307 15:13:54.758065 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f42dba37-e7f7-4657-8341-e5afa2ad256a-ring-data-devices\") pod \"swift-ring-rebalance-debug-m82qh\" (UID: \"f42dba37-e7f7-4657-8341-e5afa2ad256a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m82qh" Mar 07 15:13:54 crc kubenswrapper[4943]: I0307 15:13:54.758121 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f42dba37-e7f7-4657-8341-e5afa2ad256a-etc-swift\") pod \"swift-ring-rebalance-debug-m82qh\" (UID: \"f42dba37-e7f7-4657-8341-e5afa2ad256a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m82qh" Mar 07 15:13:54 crc kubenswrapper[4943]: I0307 15:13:54.758152 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f42dba37-e7f7-4657-8341-e5afa2ad256a-scripts\") pod \"swift-ring-rebalance-debug-m82qh\" (UID: \"f42dba37-e7f7-4657-8341-e5afa2ad256a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m82qh" Mar 07 15:13:54 crc kubenswrapper[4943]: I0307 15:13:54.758170 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f42dba37-e7f7-4657-8341-e5afa2ad256a-dispersionconf\") pod \"swift-ring-rebalance-debug-m82qh\" (UID: \"f42dba37-e7f7-4657-8341-e5afa2ad256a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m82qh" Mar 07 15:13:54 crc kubenswrapper[4943]: I0307 15:13:54.758186 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wwgd\" (UniqueName: \"kubernetes.io/projected/f42dba37-e7f7-4657-8341-e5afa2ad256a-kube-api-access-9wwgd\") pod \"swift-ring-rebalance-debug-m82qh\" (UID: \"f42dba37-e7f7-4657-8341-e5afa2ad256a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m82qh" Mar 07 15:13:54 crc kubenswrapper[4943]: I0307 15:13:54.759149 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f42dba37-e7f7-4657-8341-e5afa2ad256a-etc-swift\") pod \"swift-ring-rebalance-debug-m82qh\" (UID: \"f42dba37-e7f7-4657-8341-e5afa2ad256a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m82qh" Mar 07 15:13:54 crc kubenswrapper[4943]: I0307 15:13:54.759552 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f42dba37-e7f7-4657-8341-e5afa2ad256a-ring-data-devices\") pod \"swift-ring-rebalance-debug-m82qh\" (UID: \"f42dba37-e7f7-4657-8341-e5afa2ad256a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m82qh" Mar 07 15:13:54 crc kubenswrapper[4943]: I0307 15:13:54.760205 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f42dba37-e7f7-4657-8341-e5afa2ad256a-scripts\") pod \"swift-ring-rebalance-debug-m82qh\" (UID: \"f42dba37-e7f7-4657-8341-e5afa2ad256a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m82qh" Mar 07 15:13:54 crc kubenswrapper[4943]: I0307 15:13:54.765590 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f42dba37-e7f7-4657-8341-e5afa2ad256a-swiftconf\") pod \"swift-ring-rebalance-debug-m82qh\" (UID: \"f42dba37-e7f7-4657-8341-e5afa2ad256a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m82qh" Mar 07 15:13:54 crc kubenswrapper[4943]: I0307 15:13:54.766307 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f42dba37-e7f7-4657-8341-e5afa2ad256a-dispersionconf\") pod \"swift-ring-rebalance-debug-m82qh\" (UID: \"f42dba37-e7f7-4657-8341-e5afa2ad256a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m82qh" Mar 07 15:13:54 crc kubenswrapper[4943]: I0307 15:13:54.768775 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b659c4f-8b3c-4caf-9dae-c0b150ac3480" path="/var/lib/kubelet/pods/4b659c4f-8b3c-4caf-9dae-c0b150ac3480/volumes" Mar 07 15:13:54 crc kubenswrapper[4943]: I0307 15:13:54.792720 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wwgd\" (UniqueName: \"kubernetes.io/projected/f42dba37-e7f7-4657-8341-e5afa2ad256a-kube-api-access-9wwgd\") pod \"swift-ring-rebalance-debug-m82qh\" (UID: \"f42dba37-e7f7-4657-8341-e5afa2ad256a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m82qh" Mar 07 15:13:54 crc kubenswrapper[4943]: I0307 15:13:54.860360 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-m82qh" Mar 07 15:13:55 crc kubenswrapper[4943]: I0307 15:13:55.097838 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-m82qh"] Mar 07 15:13:55 crc kubenswrapper[4943]: W0307 15:13:55.106057 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf42dba37_e7f7_4657_8341_e5afa2ad256a.slice/crio-cacecc0c3a2baff6f28bf37fe7d5c13480f98528337cf5c33d287037f659bdb0 WatchSource:0}: Error finding container cacecc0c3a2baff6f28bf37fe7d5c13480f98528337cf5c33d287037f659bdb0: Status 404 returned error can't find the container with id cacecc0c3a2baff6f28bf37fe7d5c13480f98528337cf5c33d287037f659bdb0 Mar 07 15:13:56 crc kubenswrapper[4943]: I0307 15:13:56.024134 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-m82qh" event={"ID":"f42dba37-e7f7-4657-8341-e5afa2ad256a","Type":"ContainerStarted","Data":"bc86645c1d58a5d22060406238daa6c32169dc99dab31decce3b44d595bbbb29"} Mar 07 15:13:56 crc kubenswrapper[4943]: I0307 15:13:56.024526 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-m82qh" event={"ID":"f42dba37-e7f7-4657-8341-e5afa2ad256a","Type":"ContainerStarted","Data":"cacecc0c3a2baff6f28bf37fe7d5c13480f98528337cf5c33d287037f659bdb0"} Mar 07 15:13:56 crc kubenswrapper[4943]: I0307 15:13:56.059586 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-m82qh" podStartSLOduration=2.059570309 podStartE2EDuration="2.059570309s" podCreationTimestamp="2026-03-07 15:13:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:13:56.04860881 +0000 UTC m=+2078.000745348" watchObservedRunningTime="2026-03-07 15:13:56.059570309 +0000 UTC m=+2078.011706807" Mar 07 15:13:57 crc kubenswrapper[4943]: I0307 15:13:57.051580 4943 generic.go:334] "Generic (PLEG): container finished" podID="f42dba37-e7f7-4657-8341-e5afa2ad256a" containerID="bc86645c1d58a5d22060406238daa6c32169dc99dab31decce3b44d595bbbb29" exitCode=0 Mar 07 15:13:57 crc kubenswrapper[4943]: I0307 15:13:57.052236 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-m82qh" event={"ID":"f42dba37-e7f7-4657-8341-e5afa2ad256a","Type":"ContainerDied","Data":"bc86645c1d58a5d22060406238daa6c32169dc99dab31decce3b44d595bbbb29"} Mar 07 15:13:58 crc kubenswrapper[4943]: I0307 15:13:58.433005 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-m82qh" Mar 07 15:13:58 crc kubenswrapper[4943]: I0307 15:13:58.478651 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-m82qh"] Mar 07 15:13:58 crc kubenswrapper[4943]: I0307 15:13:58.489095 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-m82qh"] Mar 07 15:13:58 crc kubenswrapper[4943]: I0307 15:13:58.622694 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f42dba37-e7f7-4657-8341-e5afa2ad256a-etc-swift\") pod \"f42dba37-e7f7-4657-8341-e5afa2ad256a\" (UID: \"f42dba37-e7f7-4657-8341-e5afa2ad256a\") " Mar 07 15:13:58 crc kubenswrapper[4943]: I0307 15:13:58.622779 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9wwgd\" (UniqueName: \"kubernetes.io/projected/f42dba37-e7f7-4657-8341-e5afa2ad256a-kube-api-access-9wwgd\") pod \"f42dba37-e7f7-4657-8341-e5afa2ad256a\" (UID: \"f42dba37-e7f7-4657-8341-e5afa2ad256a\") " Mar 07 15:13:58 crc kubenswrapper[4943]: I0307 15:13:58.622846 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f42dba37-e7f7-4657-8341-e5afa2ad256a-scripts\") pod \"f42dba37-e7f7-4657-8341-e5afa2ad256a\" (UID: \"f42dba37-e7f7-4657-8341-e5afa2ad256a\") " Mar 07 15:13:58 crc kubenswrapper[4943]: I0307 15:13:58.622890 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f42dba37-e7f7-4657-8341-e5afa2ad256a-dispersionconf\") pod \"f42dba37-e7f7-4657-8341-e5afa2ad256a\" (UID: \"f42dba37-e7f7-4657-8341-e5afa2ad256a\") " Mar 07 15:13:58 crc kubenswrapper[4943]: I0307 15:13:58.622968 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f42dba37-e7f7-4657-8341-e5afa2ad256a-swiftconf\") pod \"f42dba37-e7f7-4657-8341-e5afa2ad256a\" (UID: \"f42dba37-e7f7-4657-8341-e5afa2ad256a\") " Mar 07 15:13:58 crc kubenswrapper[4943]: I0307 15:13:58.623022 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f42dba37-e7f7-4657-8341-e5afa2ad256a-ring-data-devices\") pod \"f42dba37-e7f7-4657-8341-e5afa2ad256a\" (UID: \"f42dba37-e7f7-4657-8341-e5afa2ad256a\") " Mar 07 15:13:58 crc kubenswrapper[4943]: I0307 15:13:58.624087 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f42dba37-e7f7-4657-8341-e5afa2ad256a-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "f42dba37-e7f7-4657-8341-e5afa2ad256a" (UID: "f42dba37-e7f7-4657-8341-e5afa2ad256a"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:13:58 crc kubenswrapper[4943]: I0307 15:13:58.624217 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f42dba37-e7f7-4657-8341-e5afa2ad256a-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "f42dba37-e7f7-4657-8341-e5afa2ad256a" (UID: "f42dba37-e7f7-4657-8341-e5afa2ad256a"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:13:58 crc kubenswrapper[4943]: I0307 15:13:58.636186 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f42dba37-e7f7-4657-8341-e5afa2ad256a-kube-api-access-9wwgd" (OuterVolumeSpecName: "kube-api-access-9wwgd") pod "f42dba37-e7f7-4657-8341-e5afa2ad256a" (UID: "f42dba37-e7f7-4657-8341-e5afa2ad256a"). InnerVolumeSpecName "kube-api-access-9wwgd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:13:58 crc kubenswrapper[4943]: I0307 15:13:58.647054 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f42dba37-e7f7-4657-8341-e5afa2ad256a-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "f42dba37-e7f7-4657-8341-e5afa2ad256a" (UID: "f42dba37-e7f7-4657-8341-e5afa2ad256a"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:13:58 crc kubenswrapper[4943]: I0307 15:13:58.663512 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f42dba37-e7f7-4657-8341-e5afa2ad256a-scripts" (OuterVolumeSpecName: "scripts") pod "f42dba37-e7f7-4657-8341-e5afa2ad256a" (UID: "f42dba37-e7f7-4657-8341-e5afa2ad256a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:13:58 crc kubenswrapper[4943]: I0307 15:13:58.664770 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f42dba37-e7f7-4657-8341-e5afa2ad256a-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "f42dba37-e7f7-4657-8341-e5afa2ad256a" (UID: "f42dba37-e7f7-4657-8341-e5afa2ad256a"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:13:58 crc kubenswrapper[4943]: I0307 15:13:58.724743 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9wwgd\" (UniqueName: \"kubernetes.io/projected/f42dba37-e7f7-4657-8341-e5afa2ad256a-kube-api-access-9wwgd\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:58 crc kubenswrapper[4943]: I0307 15:13:58.724793 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f42dba37-e7f7-4657-8341-e5afa2ad256a-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:58 crc kubenswrapper[4943]: I0307 15:13:58.724811 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f42dba37-e7f7-4657-8341-e5afa2ad256a-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:58 crc kubenswrapper[4943]: I0307 15:13:58.724831 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f42dba37-e7f7-4657-8341-e5afa2ad256a-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:58 crc kubenswrapper[4943]: I0307 15:13:58.724848 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f42dba37-e7f7-4657-8341-e5afa2ad256a-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:58 crc kubenswrapper[4943]: I0307 15:13:58.724864 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f42dba37-e7f7-4657-8341-e5afa2ad256a-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:13:58 crc kubenswrapper[4943]: I0307 15:13:58.765375 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f42dba37-e7f7-4657-8341-e5afa2ad256a" path="/var/lib/kubelet/pods/f42dba37-e7f7-4657-8341-e5afa2ad256a/volumes" Mar 07 15:13:59 crc kubenswrapper[4943]: I0307 15:13:59.075106 4943 scope.go:117] "RemoveContainer" containerID="bc86645c1d58a5d22060406238daa6c32169dc99dab31decce3b44d595bbbb29" Mar 07 15:13:59 crc kubenswrapper[4943]: I0307 15:13:59.075245 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-m82qh" Mar 07 15:13:59 crc kubenswrapper[4943]: I0307 15:13:59.675489 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-vnhdq"] Mar 07 15:13:59 crc kubenswrapper[4943]: E0307 15:13:59.676361 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f42dba37-e7f7-4657-8341-e5afa2ad256a" containerName="swift-ring-rebalance" Mar 07 15:13:59 crc kubenswrapper[4943]: I0307 15:13:59.676385 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="f42dba37-e7f7-4657-8341-e5afa2ad256a" containerName="swift-ring-rebalance" Mar 07 15:13:59 crc kubenswrapper[4943]: I0307 15:13:59.676621 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="f42dba37-e7f7-4657-8341-e5afa2ad256a" containerName="swift-ring-rebalance" Mar 07 15:13:59 crc kubenswrapper[4943]: I0307 15:13:59.677324 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-vnhdq" Mar 07 15:13:59 crc kubenswrapper[4943]: I0307 15:13:59.680870 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:13:59 crc kubenswrapper[4943]: I0307 15:13:59.681109 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:13:59 crc kubenswrapper[4943]: I0307 15:13:59.691729 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-vnhdq"] Mar 07 15:13:59 crc kubenswrapper[4943]: I0307 15:13:59.745571 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a97f54a9-b37e-42e3-b524-4ec6315e48e9-etc-swift\") pod \"swift-ring-rebalance-debug-vnhdq\" (UID: \"a97f54a9-b37e-42e3-b524-4ec6315e48e9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vnhdq" Mar 07 15:13:59 crc kubenswrapper[4943]: I0307 15:13:59.745731 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a97f54a9-b37e-42e3-b524-4ec6315e48e9-ring-data-devices\") pod \"swift-ring-rebalance-debug-vnhdq\" (UID: \"a97f54a9-b37e-42e3-b524-4ec6315e48e9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vnhdq" Mar 07 15:13:59 crc kubenswrapper[4943]: I0307 15:13:59.746018 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a97f54a9-b37e-42e3-b524-4ec6315e48e9-scripts\") pod \"swift-ring-rebalance-debug-vnhdq\" (UID: \"a97f54a9-b37e-42e3-b524-4ec6315e48e9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vnhdq" Mar 07 15:13:59 crc kubenswrapper[4943]: I0307 15:13:59.746092 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qqxx\" (UniqueName: \"kubernetes.io/projected/a97f54a9-b37e-42e3-b524-4ec6315e48e9-kube-api-access-2qqxx\") pod \"swift-ring-rebalance-debug-vnhdq\" (UID: \"a97f54a9-b37e-42e3-b524-4ec6315e48e9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vnhdq" Mar 07 15:13:59 crc kubenswrapper[4943]: I0307 15:13:59.746175 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a97f54a9-b37e-42e3-b524-4ec6315e48e9-swiftconf\") pod \"swift-ring-rebalance-debug-vnhdq\" (UID: \"a97f54a9-b37e-42e3-b524-4ec6315e48e9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vnhdq" Mar 07 15:13:59 crc kubenswrapper[4943]: I0307 15:13:59.746223 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a97f54a9-b37e-42e3-b524-4ec6315e48e9-dispersionconf\") pod \"swift-ring-rebalance-debug-vnhdq\" (UID: \"a97f54a9-b37e-42e3-b524-4ec6315e48e9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vnhdq" Mar 07 15:13:59 crc kubenswrapper[4943]: I0307 15:13:59.847171 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a97f54a9-b37e-42e3-b524-4ec6315e48e9-scripts\") pod \"swift-ring-rebalance-debug-vnhdq\" (UID: \"a97f54a9-b37e-42e3-b524-4ec6315e48e9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vnhdq" Mar 07 15:13:59 crc kubenswrapper[4943]: I0307 15:13:59.847228 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qqxx\" (UniqueName: \"kubernetes.io/projected/a97f54a9-b37e-42e3-b524-4ec6315e48e9-kube-api-access-2qqxx\") pod \"swift-ring-rebalance-debug-vnhdq\" (UID: \"a97f54a9-b37e-42e3-b524-4ec6315e48e9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vnhdq" Mar 07 15:13:59 crc kubenswrapper[4943]: I0307 15:13:59.847278 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a97f54a9-b37e-42e3-b524-4ec6315e48e9-swiftconf\") pod \"swift-ring-rebalance-debug-vnhdq\" (UID: \"a97f54a9-b37e-42e3-b524-4ec6315e48e9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vnhdq" Mar 07 15:13:59 crc kubenswrapper[4943]: I0307 15:13:59.847308 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a97f54a9-b37e-42e3-b524-4ec6315e48e9-dispersionconf\") pod \"swift-ring-rebalance-debug-vnhdq\" (UID: \"a97f54a9-b37e-42e3-b524-4ec6315e48e9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vnhdq" Mar 07 15:13:59 crc kubenswrapper[4943]: I0307 15:13:59.847367 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a97f54a9-b37e-42e3-b524-4ec6315e48e9-etc-swift\") pod \"swift-ring-rebalance-debug-vnhdq\" (UID: \"a97f54a9-b37e-42e3-b524-4ec6315e48e9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vnhdq" Mar 07 15:13:59 crc kubenswrapper[4943]: I0307 15:13:59.847390 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a97f54a9-b37e-42e3-b524-4ec6315e48e9-ring-data-devices\") pod \"swift-ring-rebalance-debug-vnhdq\" (UID: \"a97f54a9-b37e-42e3-b524-4ec6315e48e9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vnhdq" Mar 07 15:13:59 crc kubenswrapper[4943]: I0307 15:13:59.848076 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a97f54a9-b37e-42e3-b524-4ec6315e48e9-scripts\") pod \"swift-ring-rebalance-debug-vnhdq\" (UID: \"a97f54a9-b37e-42e3-b524-4ec6315e48e9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vnhdq" Mar 07 15:13:59 crc kubenswrapper[4943]: I0307 15:13:59.848202 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a97f54a9-b37e-42e3-b524-4ec6315e48e9-etc-swift\") pod \"swift-ring-rebalance-debug-vnhdq\" (UID: \"a97f54a9-b37e-42e3-b524-4ec6315e48e9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vnhdq" Mar 07 15:13:59 crc kubenswrapper[4943]: I0307 15:13:59.848312 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a97f54a9-b37e-42e3-b524-4ec6315e48e9-ring-data-devices\") pod \"swift-ring-rebalance-debug-vnhdq\" (UID: \"a97f54a9-b37e-42e3-b524-4ec6315e48e9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vnhdq" Mar 07 15:13:59 crc kubenswrapper[4943]: I0307 15:13:59.856562 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a97f54a9-b37e-42e3-b524-4ec6315e48e9-dispersionconf\") pod \"swift-ring-rebalance-debug-vnhdq\" (UID: \"a97f54a9-b37e-42e3-b524-4ec6315e48e9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vnhdq" Mar 07 15:13:59 crc kubenswrapper[4943]: I0307 15:13:59.856611 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a97f54a9-b37e-42e3-b524-4ec6315e48e9-swiftconf\") pod \"swift-ring-rebalance-debug-vnhdq\" (UID: \"a97f54a9-b37e-42e3-b524-4ec6315e48e9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vnhdq" Mar 07 15:13:59 crc kubenswrapper[4943]: I0307 15:13:59.871208 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qqxx\" (UniqueName: \"kubernetes.io/projected/a97f54a9-b37e-42e3-b524-4ec6315e48e9-kube-api-access-2qqxx\") pod \"swift-ring-rebalance-debug-vnhdq\" (UID: \"a97f54a9-b37e-42e3-b524-4ec6315e48e9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vnhdq" Mar 07 15:14:00 crc kubenswrapper[4943]: I0307 15:14:00.010081 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-vnhdq" Mar 07 15:14:00 crc kubenswrapper[4943]: I0307 15:14:00.135084 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29548274-9dcnx"] Mar 07 15:14:00 crc kubenswrapper[4943]: I0307 15:14:00.136589 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548274-9dcnx" Mar 07 15:14:00 crc kubenswrapper[4943]: I0307 15:14:00.138484 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 07 15:14:00 crc kubenswrapper[4943]: I0307 15:14:00.138567 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 07 15:14:00 crc kubenswrapper[4943]: I0307 15:14:00.143944 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-qqlk5" Mar 07 15:14:00 crc kubenswrapper[4943]: I0307 15:14:00.153898 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29548274-9dcnx"] Mar 07 15:14:00 crc kubenswrapper[4943]: I0307 15:14:00.154414 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8w5zj\" (UniqueName: \"kubernetes.io/projected/9b797be7-6317-4ddc-95dd-652f99ba6a32-kube-api-access-8w5zj\") pod \"auto-csr-approver-29548274-9dcnx\" (UID: \"9b797be7-6317-4ddc-95dd-652f99ba6a32\") " pod="openshift-infra/auto-csr-approver-29548274-9dcnx" Mar 07 15:14:00 crc kubenswrapper[4943]: I0307 15:14:00.255865 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8w5zj\" (UniqueName: \"kubernetes.io/projected/9b797be7-6317-4ddc-95dd-652f99ba6a32-kube-api-access-8w5zj\") pod \"auto-csr-approver-29548274-9dcnx\" (UID: \"9b797be7-6317-4ddc-95dd-652f99ba6a32\") " pod="openshift-infra/auto-csr-approver-29548274-9dcnx" Mar 07 15:14:00 crc kubenswrapper[4943]: I0307 15:14:00.275025 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8w5zj\" (UniqueName: \"kubernetes.io/projected/9b797be7-6317-4ddc-95dd-652f99ba6a32-kube-api-access-8w5zj\") pod \"auto-csr-approver-29548274-9dcnx\" (UID: \"9b797be7-6317-4ddc-95dd-652f99ba6a32\") " pod="openshift-infra/auto-csr-approver-29548274-9dcnx" Mar 07 15:14:00 crc kubenswrapper[4943]: I0307 15:14:00.440272 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-vnhdq"] Mar 07 15:14:00 crc kubenswrapper[4943]: I0307 15:14:00.488919 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548274-9dcnx" Mar 07 15:14:00 crc kubenswrapper[4943]: I0307 15:14:00.778147 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29548274-9dcnx"] Mar 07 15:14:00 crc kubenswrapper[4943]: W0307 15:14:00.779043 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9b797be7_6317_4ddc_95dd_652f99ba6a32.slice/crio-2020731faa0e40a0ff8876725c36e943d28caca92fe2ba54626f97899dbc535c WatchSource:0}: Error finding container 2020731faa0e40a0ff8876725c36e943d28caca92fe2ba54626f97899dbc535c: Status 404 returned error can't find the container with id 2020731faa0e40a0ff8876725c36e943d28caca92fe2ba54626f97899dbc535c Mar 07 15:14:01 crc kubenswrapper[4943]: I0307 15:14:01.117793 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-vnhdq" event={"ID":"a97f54a9-b37e-42e3-b524-4ec6315e48e9","Type":"ContainerStarted","Data":"f5f943654c0799359019ede37edc2fc94b884ba56d5f2f4d65ae0f659a310e73"} Mar 07 15:14:01 crc kubenswrapper[4943]: I0307 15:14:01.118273 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-vnhdq" event={"ID":"a97f54a9-b37e-42e3-b524-4ec6315e48e9","Type":"ContainerStarted","Data":"ba724188daa880d3d29ff9b47bc06e13e3feefeba197a63529e7515cef21bd40"} Mar 07 15:14:01 crc kubenswrapper[4943]: I0307 15:14:01.119191 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548274-9dcnx" event={"ID":"9b797be7-6317-4ddc-95dd-652f99ba6a32","Type":"ContainerStarted","Data":"2020731faa0e40a0ff8876725c36e943d28caca92fe2ba54626f97899dbc535c"} Mar 07 15:14:01 crc kubenswrapper[4943]: I0307 15:14:01.156646 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-vnhdq" podStartSLOduration=2.156622396 podStartE2EDuration="2.156622396s" podCreationTimestamp="2026-03-07 15:13:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:14:01.146350454 +0000 UTC m=+2083.098486962" watchObservedRunningTime="2026-03-07 15:14:01.156622396 +0000 UTC m=+2083.108758904" Mar 07 15:14:02 crc kubenswrapper[4943]: I0307 15:14:02.128870 4943 generic.go:334] "Generic (PLEG): container finished" podID="a97f54a9-b37e-42e3-b524-4ec6315e48e9" containerID="f5f943654c0799359019ede37edc2fc94b884ba56d5f2f4d65ae0f659a310e73" exitCode=0 Mar 07 15:14:02 crc kubenswrapper[4943]: I0307 15:14:02.128917 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-vnhdq" event={"ID":"a97f54a9-b37e-42e3-b524-4ec6315e48e9","Type":"ContainerDied","Data":"f5f943654c0799359019ede37edc2fc94b884ba56d5f2f4d65ae0f659a310e73"} Mar 07 15:14:02 crc kubenswrapper[4943]: I0307 15:14:02.130969 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548274-9dcnx" event={"ID":"9b797be7-6317-4ddc-95dd-652f99ba6a32","Type":"ContainerStarted","Data":"7091e809442e2336edbc5d300cf4a15937fc81d3519536fbf5a5d0cbbc5bb824"} Mar 07 15:14:02 crc kubenswrapper[4943]: I0307 15:14:02.173307 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29548274-9dcnx" podStartSLOduration=1.211209455 podStartE2EDuration="2.173285344s" podCreationTimestamp="2026-03-07 15:14:00 +0000 UTC" firstStartedPulling="2026-03-07 15:14:00.783287688 +0000 UTC m=+2082.735424186" lastFinishedPulling="2026-03-07 15:14:01.745363577 +0000 UTC m=+2083.697500075" observedRunningTime="2026-03-07 15:14:02.171829568 +0000 UTC m=+2084.123966066" watchObservedRunningTime="2026-03-07 15:14:02.173285344 +0000 UTC m=+2084.125421882" Mar 07 15:14:03 crc kubenswrapper[4943]: I0307 15:14:03.143393 4943 generic.go:334] "Generic (PLEG): container finished" podID="9b797be7-6317-4ddc-95dd-652f99ba6a32" containerID="7091e809442e2336edbc5d300cf4a15937fc81d3519536fbf5a5d0cbbc5bb824" exitCode=0 Mar 07 15:14:03 crc kubenswrapper[4943]: I0307 15:14:03.143536 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548274-9dcnx" event={"ID":"9b797be7-6317-4ddc-95dd-652f99ba6a32","Type":"ContainerDied","Data":"7091e809442e2336edbc5d300cf4a15937fc81d3519536fbf5a5d0cbbc5bb824"} Mar 07 15:14:03 crc kubenswrapper[4943]: I0307 15:14:03.456175 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-vnhdq" Mar 07 15:14:03 crc kubenswrapper[4943]: I0307 15:14:03.503971 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a97f54a9-b37e-42e3-b524-4ec6315e48e9-ring-data-devices\") pod \"a97f54a9-b37e-42e3-b524-4ec6315e48e9\" (UID: \"a97f54a9-b37e-42e3-b524-4ec6315e48e9\") " Mar 07 15:14:03 crc kubenswrapper[4943]: I0307 15:14:03.504030 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a97f54a9-b37e-42e3-b524-4ec6315e48e9-scripts\") pod \"a97f54a9-b37e-42e3-b524-4ec6315e48e9\" (UID: \"a97f54a9-b37e-42e3-b524-4ec6315e48e9\") " Mar 07 15:14:03 crc kubenswrapper[4943]: I0307 15:14:03.504082 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2qqxx\" (UniqueName: \"kubernetes.io/projected/a97f54a9-b37e-42e3-b524-4ec6315e48e9-kube-api-access-2qqxx\") pod \"a97f54a9-b37e-42e3-b524-4ec6315e48e9\" (UID: \"a97f54a9-b37e-42e3-b524-4ec6315e48e9\") " Mar 07 15:14:03 crc kubenswrapper[4943]: I0307 15:14:03.504187 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a97f54a9-b37e-42e3-b524-4ec6315e48e9-swiftconf\") pod \"a97f54a9-b37e-42e3-b524-4ec6315e48e9\" (UID: \"a97f54a9-b37e-42e3-b524-4ec6315e48e9\") " Mar 07 15:14:03 crc kubenswrapper[4943]: I0307 15:14:03.504218 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a97f54a9-b37e-42e3-b524-4ec6315e48e9-dispersionconf\") pod \"a97f54a9-b37e-42e3-b524-4ec6315e48e9\" (UID: \"a97f54a9-b37e-42e3-b524-4ec6315e48e9\") " Mar 07 15:14:03 crc kubenswrapper[4943]: I0307 15:14:03.504275 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a97f54a9-b37e-42e3-b524-4ec6315e48e9-etc-swift\") pod \"a97f54a9-b37e-42e3-b524-4ec6315e48e9\" (UID: \"a97f54a9-b37e-42e3-b524-4ec6315e48e9\") " Mar 07 15:14:03 crc kubenswrapper[4943]: I0307 15:14:03.505535 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a97f54a9-b37e-42e3-b524-4ec6315e48e9-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "a97f54a9-b37e-42e3-b524-4ec6315e48e9" (UID: "a97f54a9-b37e-42e3-b524-4ec6315e48e9"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:14:03 crc kubenswrapper[4943]: I0307 15:14:03.506365 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a97f54a9-b37e-42e3-b524-4ec6315e48e9-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "a97f54a9-b37e-42e3-b524-4ec6315e48e9" (UID: "a97f54a9-b37e-42e3-b524-4ec6315e48e9"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:14:03 crc kubenswrapper[4943]: I0307 15:14:03.512588 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a97f54a9-b37e-42e3-b524-4ec6315e48e9-kube-api-access-2qqxx" (OuterVolumeSpecName: "kube-api-access-2qqxx") pod "a97f54a9-b37e-42e3-b524-4ec6315e48e9" (UID: "a97f54a9-b37e-42e3-b524-4ec6315e48e9"). InnerVolumeSpecName "kube-api-access-2qqxx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:14:03 crc kubenswrapper[4943]: I0307 15:14:03.515097 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-vnhdq"] Mar 07 15:14:03 crc kubenswrapper[4943]: I0307 15:14:03.522787 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-vnhdq"] Mar 07 15:14:03 crc kubenswrapper[4943]: I0307 15:14:03.527060 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a97f54a9-b37e-42e3-b524-4ec6315e48e9-scripts" (OuterVolumeSpecName: "scripts") pod "a97f54a9-b37e-42e3-b524-4ec6315e48e9" (UID: "a97f54a9-b37e-42e3-b524-4ec6315e48e9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:14:03 crc kubenswrapper[4943]: I0307 15:14:03.543893 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a97f54a9-b37e-42e3-b524-4ec6315e48e9-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "a97f54a9-b37e-42e3-b524-4ec6315e48e9" (UID: "a97f54a9-b37e-42e3-b524-4ec6315e48e9"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:14:03 crc kubenswrapper[4943]: I0307 15:14:03.545680 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a97f54a9-b37e-42e3-b524-4ec6315e48e9-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "a97f54a9-b37e-42e3-b524-4ec6315e48e9" (UID: "a97f54a9-b37e-42e3-b524-4ec6315e48e9"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:14:03 crc kubenswrapper[4943]: I0307 15:14:03.605798 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a97f54a9-b37e-42e3-b524-4ec6315e48e9-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:03 crc kubenswrapper[4943]: I0307 15:14:03.605835 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a97f54a9-b37e-42e3-b524-4ec6315e48e9-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:03 crc kubenswrapper[4943]: I0307 15:14:03.605850 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a97f54a9-b37e-42e3-b524-4ec6315e48e9-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:03 crc kubenswrapper[4943]: I0307 15:14:03.605862 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a97f54a9-b37e-42e3-b524-4ec6315e48e9-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:03 crc kubenswrapper[4943]: I0307 15:14:03.605874 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a97f54a9-b37e-42e3-b524-4ec6315e48e9-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:03 crc kubenswrapper[4943]: I0307 15:14:03.605886 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2qqxx\" (UniqueName: \"kubernetes.io/projected/a97f54a9-b37e-42e3-b524-4ec6315e48e9-kube-api-access-2qqxx\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:04 crc kubenswrapper[4943]: I0307 15:14:04.157430 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ba724188daa880d3d29ff9b47bc06e13e3feefeba197a63529e7515cef21bd40" Mar 07 15:14:04 crc kubenswrapper[4943]: I0307 15:14:04.157478 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-vnhdq" Mar 07 15:14:04 crc kubenswrapper[4943]: I0307 15:14:04.546609 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548274-9dcnx" Mar 07 15:14:04 crc kubenswrapper[4943]: I0307 15:14:04.632464 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8w5zj\" (UniqueName: \"kubernetes.io/projected/9b797be7-6317-4ddc-95dd-652f99ba6a32-kube-api-access-8w5zj\") pod \"9b797be7-6317-4ddc-95dd-652f99ba6a32\" (UID: \"9b797be7-6317-4ddc-95dd-652f99ba6a32\") " Mar 07 15:14:04 crc kubenswrapper[4943]: I0307 15:14:04.638786 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b797be7-6317-4ddc-95dd-652f99ba6a32-kube-api-access-8w5zj" (OuterVolumeSpecName: "kube-api-access-8w5zj") pod "9b797be7-6317-4ddc-95dd-652f99ba6a32" (UID: "9b797be7-6317-4ddc-95dd-652f99ba6a32"). InnerVolumeSpecName "kube-api-access-8w5zj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:14:04 crc kubenswrapper[4943]: I0307 15:14:04.671018 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-56gh4"] Mar 07 15:14:04 crc kubenswrapper[4943]: E0307 15:14:04.671448 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b797be7-6317-4ddc-95dd-652f99ba6a32" containerName="oc" Mar 07 15:14:04 crc kubenswrapper[4943]: I0307 15:14:04.671486 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b797be7-6317-4ddc-95dd-652f99ba6a32" containerName="oc" Mar 07 15:14:04 crc kubenswrapper[4943]: E0307 15:14:04.671498 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a97f54a9-b37e-42e3-b524-4ec6315e48e9" containerName="swift-ring-rebalance" Mar 07 15:14:04 crc kubenswrapper[4943]: I0307 15:14:04.671506 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="a97f54a9-b37e-42e3-b524-4ec6315e48e9" containerName="swift-ring-rebalance" Mar 07 15:14:04 crc kubenswrapper[4943]: I0307 15:14:04.671767 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="a97f54a9-b37e-42e3-b524-4ec6315e48e9" containerName="swift-ring-rebalance" Mar 07 15:14:04 crc kubenswrapper[4943]: I0307 15:14:04.671808 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b797be7-6317-4ddc-95dd-652f99ba6a32" containerName="oc" Mar 07 15:14:04 crc kubenswrapper[4943]: I0307 15:14:04.672463 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-56gh4" Mar 07 15:14:04 crc kubenswrapper[4943]: I0307 15:14:04.673968 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:14:04 crc kubenswrapper[4943]: I0307 15:14:04.677157 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:14:04 crc kubenswrapper[4943]: I0307 15:14:04.682298 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-56gh4"] Mar 07 15:14:04 crc kubenswrapper[4943]: I0307 15:14:04.736016 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8w5zj\" (UniqueName: \"kubernetes.io/projected/9b797be7-6317-4ddc-95dd-652f99ba6a32-kube-api-access-8w5zj\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:04 crc kubenswrapper[4943]: I0307 15:14:04.763590 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a97f54a9-b37e-42e3-b524-4ec6315e48e9" path="/var/lib/kubelet/pods/a97f54a9-b37e-42e3-b524-4ec6315e48e9/volumes" Mar 07 15:14:04 crc kubenswrapper[4943]: I0307 15:14:04.837299 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6pmjr\" (UniqueName: \"kubernetes.io/projected/95e89075-16f2-4cc2-8f97-095e13422bc9-kube-api-access-6pmjr\") pod \"swift-ring-rebalance-debug-56gh4\" (UID: \"95e89075-16f2-4cc2-8f97-095e13422bc9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-56gh4" Mar 07 15:14:04 crc kubenswrapper[4943]: I0307 15:14:04.837630 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/95e89075-16f2-4cc2-8f97-095e13422bc9-dispersionconf\") pod \"swift-ring-rebalance-debug-56gh4\" (UID: \"95e89075-16f2-4cc2-8f97-095e13422bc9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-56gh4" Mar 07 15:14:04 crc kubenswrapper[4943]: I0307 15:14:04.837740 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/95e89075-16f2-4cc2-8f97-095e13422bc9-swiftconf\") pod \"swift-ring-rebalance-debug-56gh4\" (UID: \"95e89075-16f2-4cc2-8f97-095e13422bc9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-56gh4" Mar 07 15:14:04 crc kubenswrapper[4943]: I0307 15:14:04.837846 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/95e89075-16f2-4cc2-8f97-095e13422bc9-scripts\") pod \"swift-ring-rebalance-debug-56gh4\" (UID: \"95e89075-16f2-4cc2-8f97-095e13422bc9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-56gh4" Mar 07 15:14:04 crc kubenswrapper[4943]: I0307 15:14:04.837965 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/95e89075-16f2-4cc2-8f97-095e13422bc9-ring-data-devices\") pod \"swift-ring-rebalance-debug-56gh4\" (UID: \"95e89075-16f2-4cc2-8f97-095e13422bc9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-56gh4" Mar 07 15:14:04 crc kubenswrapper[4943]: I0307 15:14:04.838080 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/95e89075-16f2-4cc2-8f97-095e13422bc9-etc-swift\") pod \"swift-ring-rebalance-debug-56gh4\" (UID: \"95e89075-16f2-4cc2-8f97-095e13422bc9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-56gh4" Mar 07 15:14:04 crc kubenswrapper[4943]: I0307 15:14:04.939581 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/95e89075-16f2-4cc2-8f97-095e13422bc9-swiftconf\") pod \"swift-ring-rebalance-debug-56gh4\" (UID: \"95e89075-16f2-4cc2-8f97-095e13422bc9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-56gh4" Mar 07 15:14:04 crc kubenswrapper[4943]: I0307 15:14:04.939854 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/95e89075-16f2-4cc2-8f97-095e13422bc9-scripts\") pod \"swift-ring-rebalance-debug-56gh4\" (UID: \"95e89075-16f2-4cc2-8f97-095e13422bc9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-56gh4" Mar 07 15:14:04 crc kubenswrapper[4943]: I0307 15:14:04.939968 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/95e89075-16f2-4cc2-8f97-095e13422bc9-ring-data-devices\") pod \"swift-ring-rebalance-debug-56gh4\" (UID: \"95e89075-16f2-4cc2-8f97-095e13422bc9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-56gh4" Mar 07 15:14:04 crc kubenswrapper[4943]: I0307 15:14:04.940080 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/95e89075-16f2-4cc2-8f97-095e13422bc9-etc-swift\") pod \"swift-ring-rebalance-debug-56gh4\" (UID: \"95e89075-16f2-4cc2-8f97-095e13422bc9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-56gh4" Mar 07 15:14:04 crc kubenswrapper[4943]: I0307 15:14:04.940289 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6pmjr\" (UniqueName: \"kubernetes.io/projected/95e89075-16f2-4cc2-8f97-095e13422bc9-kube-api-access-6pmjr\") pod \"swift-ring-rebalance-debug-56gh4\" (UID: \"95e89075-16f2-4cc2-8f97-095e13422bc9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-56gh4" Mar 07 15:14:04 crc kubenswrapper[4943]: I0307 15:14:04.940426 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/95e89075-16f2-4cc2-8f97-095e13422bc9-dispersionconf\") pod \"swift-ring-rebalance-debug-56gh4\" (UID: \"95e89075-16f2-4cc2-8f97-095e13422bc9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-56gh4" Mar 07 15:14:04 crc kubenswrapper[4943]: I0307 15:14:04.941580 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/95e89075-16f2-4cc2-8f97-095e13422bc9-etc-swift\") pod \"swift-ring-rebalance-debug-56gh4\" (UID: \"95e89075-16f2-4cc2-8f97-095e13422bc9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-56gh4" Mar 07 15:14:04 crc kubenswrapper[4943]: I0307 15:14:04.941668 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/95e89075-16f2-4cc2-8f97-095e13422bc9-scripts\") pod \"swift-ring-rebalance-debug-56gh4\" (UID: \"95e89075-16f2-4cc2-8f97-095e13422bc9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-56gh4" Mar 07 15:14:04 crc kubenswrapper[4943]: I0307 15:14:04.941771 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/95e89075-16f2-4cc2-8f97-095e13422bc9-ring-data-devices\") pod \"swift-ring-rebalance-debug-56gh4\" (UID: \"95e89075-16f2-4cc2-8f97-095e13422bc9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-56gh4" Mar 07 15:14:04 crc kubenswrapper[4943]: I0307 15:14:04.949430 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/95e89075-16f2-4cc2-8f97-095e13422bc9-swiftconf\") pod \"swift-ring-rebalance-debug-56gh4\" (UID: \"95e89075-16f2-4cc2-8f97-095e13422bc9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-56gh4" Mar 07 15:14:04 crc kubenswrapper[4943]: I0307 15:14:04.952666 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/95e89075-16f2-4cc2-8f97-095e13422bc9-dispersionconf\") pod \"swift-ring-rebalance-debug-56gh4\" (UID: \"95e89075-16f2-4cc2-8f97-095e13422bc9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-56gh4" Mar 07 15:14:04 crc kubenswrapper[4943]: I0307 15:14:04.956540 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6pmjr\" (UniqueName: \"kubernetes.io/projected/95e89075-16f2-4cc2-8f97-095e13422bc9-kube-api-access-6pmjr\") pod \"swift-ring-rebalance-debug-56gh4\" (UID: \"95e89075-16f2-4cc2-8f97-095e13422bc9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-56gh4" Mar 07 15:14:04 crc kubenswrapper[4943]: I0307 15:14:04.996507 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-56gh4" Mar 07 15:14:05 crc kubenswrapper[4943]: I0307 15:14:05.168445 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548274-9dcnx" event={"ID":"9b797be7-6317-4ddc-95dd-652f99ba6a32","Type":"ContainerDied","Data":"2020731faa0e40a0ff8876725c36e943d28caca92fe2ba54626f97899dbc535c"} Mar 07 15:14:05 crc kubenswrapper[4943]: I0307 15:14:05.168490 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2020731faa0e40a0ff8876725c36e943d28caca92fe2ba54626f97899dbc535c" Mar 07 15:14:05 crc kubenswrapper[4943]: I0307 15:14:05.168548 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548274-9dcnx" Mar 07 15:14:05 crc kubenswrapper[4943]: I0307 15:14:05.479335 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-56gh4"] Mar 07 15:14:05 crc kubenswrapper[4943]: W0307 15:14:05.493791 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod95e89075_16f2_4cc2_8f97_095e13422bc9.slice/crio-2d19df085699ce389840d00c2c20d850ecfbe45e5820673ee7eff403956072d3 WatchSource:0}: Error finding container 2d19df085699ce389840d00c2c20d850ecfbe45e5820673ee7eff403956072d3: Status 404 returned error can't find the container with id 2d19df085699ce389840d00c2c20d850ecfbe45e5820673ee7eff403956072d3 Mar 07 15:14:05 crc kubenswrapper[4943]: I0307 15:14:05.613503 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29548268-nlpcf"] Mar 07 15:14:05 crc kubenswrapper[4943]: I0307 15:14:05.620333 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29548268-nlpcf"] Mar 07 15:14:06 crc kubenswrapper[4943]: I0307 15:14:06.177993 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-56gh4" event={"ID":"95e89075-16f2-4cc2-8f97-095e13422bc9","Type":"ContainerStarted","Data":"6dce9a5b5b27e19bbf1e1be4e2cf256568efbbf4caad1b6ce9db6885dcfd2866"} Mar 07 15:14:06 crc kubenswrapper[4943]: I0307 15:14:06.178049 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-56gh4" event={"ID":"95e89075-16f2-4cc2-8f97-095e13422bc9","Type":"ContainerStarted","Data":"2d19df085699ce389840d00c2c20d850ecfbe45e5820673ee7eff403956072d3"} Mar 07 15:14:06 crc kubenswrapper[4943]: I0307 15:14:06.224596 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-56gh4" podStartSLOduration=2.224564869 podStartE2EDuration="2.224564869s" podCreationTimestamp="2026-03-07 15:14:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:14:06.219566107 +0000 UTC m=+2088.171702655" watchObservedRunningTime="2026-03-07 15:14:06.224564869 +0000 UTC m=+2088.176701417" Mar 07 15:14:06 crc kubenswrapper[4943]: I0307 15:14:06.765911 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb3b8ef8-e951-4f8c-8579-c9515a42b092" path="/var/lib/kubelet/pods/eb3b8ef8-e951-4f8c-8579-c9515a42b092/volumes" Mar 07 15:14:07 crc kubenswrapper[4943]: I0307 15:14:07.185505 4943 generic.go:334] "Generic (PLEG): container finished" podID="95e89075-16f2-4cc2-8f97-095e13422bc9" containerID="6dce9a5b5b27e19bbf1e1be4e2cf256568efbbf4caad1b6ce9db6885dcfd2866" exitCode=0 Mar 07 15:14:07 crc kubenswrapper[4943]: I0307 15:14:07.185540 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-56gh4" event={"ID":"95e89075-16f2-4cc2-8f97-095e13422bc9","Type":"ContainerDied","Data":"6dce9a5b5b27e19bbf1e1be4e2cf256568efbbf4caad1b6ce9db6885dcfd2866"} Mar 07 15:14:08 crc kubenswrapper[4943]: I0307 15:14:08.527057 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-56gh4" Mar 07 15:14:08 crc kubenswrapper[4943]: I0307 15:14:08.567448 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-56gh4"] Mar 07 15:14:08 crc kubenswrapper[4943]: I0307 15:14:08.573534 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-56gh4"] Mar 07 15:14:08 crc kubenswrapper[4943]: I0307 15:14:08.701437 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/95e89075-16f2-4cc2-8f97-095e13422bc9-dispersionconf\") pod \"95e89075-16f2-4cc2-8f97-095e13422bc9\" (UID: \"95e89075-16f2-4cc2-8f97-095e13422bc9\") " Mar 07 15:14:08 crc kubenswrapper[4943]: I0307 15:14:08.701526 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/95e89075-16f2-4cc2-8f97-095e13422bc9-scripts\") pod \"95e89075-16f2-4cc2-8f97-095e13422bc9\" (UID: \"95e89075-16f2-4cc2-8f97-095e13422bc9\") " Mar 07 15:14:08 crc kubenswrapper[4943]: I0307 15:14:08.701586 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6pmjr\" (UniqueName: \"kubernetes.io/projected/95e89075-16f2-4cc2-8f97-095e13422bc9-kube-api-access-6pmjr\") pod \"95e89075-16f2-4cc2-8f97-095e13422bc9\" (UID: \"95e89075-16f2-4cc2-8f97-095e13422bc9\") " Mar 07 15:14:08 crc kubenswrapper[4943]: I0307 15:14:08.701629 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/95e89075-16f2-4cc2-8f97-095e13422bc9-swiftconf\") pod \"95e89075-16f2-4cc2-8f97-095e13422bc9\" (UID: \"95e89075-16f2-4cc2-8f97-095e13422bc9\") " Mar 07 15:14:08 crc kubenswrapper[4943]: I0307 15:14:08.702457 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/95e89075-16f2-4cc2-8f97-095e13422bc9-etc-swift\") pod \"95e89075-16f2-4cc2-8f97-095e13422bc9\" (UID: \"95e89075-16f2-4cc2-8f97-095e13422bc9\") " Mar 07 15:14:08 crc kubenswrapper[4943]: I0307 15:14:08.703078 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/95e89075-16f2-4cc2-8f97-095e13422bc9-ring-data-devices\") pod \"95e89075-16f2-4cc2-8f97-095e13422bc9\" (UID: \"95e89075-16f2-4cc2-8f97-095e13422bc9\") " Mar 07 15:14:08 crc kubenswrapper[4943]: I0307 15:14:08.703949 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95e89075-16f2-4cc2-8f97-095e13422bc9-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "95e89075-16f2-4cc2-8f97-095e13422bc9" (UID: "95e89075-16f2-4cc2-8f97-095e13422bc9"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:14:08 crc kubenswrapper[4943]: I0307 15:14:08.704063 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95e89075-16f2-4cc2-8f97-095e13422bc9-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "95e89075-16f2-4cc2-8f97-095e13422bc9" (UID: "95e89075-16f2-4cc2-8f97-095e13422bc9"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:14:08 crc kubenswrapper[4943]: I0307 15:14:08.707213 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95e89075-16f2-4cc2-8f97-095e13422bc9-kube-api-access-6pmjr" (OuterVolumeSpecName: "kube-api-access-6pmjr") pod "95e89075-16f2-4cc2-8f97-095e13422bc9" (UID: "95e89075-16f2-4cc2-8f97-095e13422bc9"). InnerVolumeSpecName "kube-api-access-6pmjr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:14:08 crc kubenswrapper[4943]: I0307 15:14:08.732577 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95e89075-16f2-4cc2-8f97-095e13422bc9-scripts" (OuterVolumeSpecName: "scripts") pod "95e89075-16f2-4cc2-8f97-095e13422bc9" (UID: "95e89075-16f2-4cc2-8f97-095e13422bc9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:14:08 crc kubenswrapper[4943]: I0307 15:14:08.740417 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95e89075-16f2-4cc2-8f97-095e13422bc9-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "95e89075-16f2-4cc2-8f97-095e13422bc9" (UID: "95e89075-16f2-4cc2-8f97-095e13422bc9"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:14:08 crc kubenswrapper[4943]: I0307 15:14:08.741653 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95e89075-16f2-4cc2-8f97-095e13422bc9-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "95e89075-16f2-4cc2-8f97-095e13422bc9" (UID: "95e89075-16f2-4cc2-8f97-095e13422bc9"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:14:08 crc kubenswrapper[4943]: I0307 15:14:08.769525 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95e89075-16f2-4cc2-8f97-095e13422bc9" path="/var/lib/kubelet/pods/95e89075-16f2-4cc2-8f97-095e13422bc9/volumes" Mar 07 15:14:08 crc kubenswrapper[4943]: I0307 15:14:08.804469 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/95e89075-16f2-4cc2-8f97-095e13422bc9-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:08 crc kubenswrapper[4943]: I0307 15:14:08.804499 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/95e89075-16f2-4cc2-8f97-095e13422bc9-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:08 crc kubenswrapper[4943]: I0307 15:14:08.804508 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6pmjr\" (UniqueName: \"kubernetes.io/projected/95e89075-16f2-4cc2-8f97-095e13422bc9-kube-api-access-6pmjr\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:08 crc kubenswrapper[4943]: I0307 15:14:08.804519 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/95e89075-16f2-4cc2-8f97-095e13422bc9-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:08 crc kubenswrapper[4943]: I0307 15:14:08.804527 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/95e89075-16f2-4cc2-8f97-095e13422bc9-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:08 crc kubenswrapper[4943]: I0307 15:14:08.804535 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/95e89075-16f2-4cc2-8f97-095e13422bc9-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:09 crc kubenswrapper[4943]: I0307 15:14:09.212547 4943 scope.go:117] "RemoveContainer" containerID="6dce9a5b5b27e19bbf1e1be4e2cf256568efbbf4caad1b6ce9db6885dcfd2866" Mar 07 15:14:09 crc kubenswrapper[4943]: I0307 15:14:09.213168 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-56gh4" Mar 07 15:14:09 crc kubenswrapper[4943]: I0307 15:14:09.728641 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-pg7qh"] Mar 07 15:14:09 crc kubenswrapper[4943]: E0307 15:14:09.729109 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95e89075-16f2-4cc2-8f97-095e13422bc9" containerName="swift-ring-rebalance" Mar 07 15:14:09 crc kubenswrapper[4943]: I0307 15:14:09.729132 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="95e89075-16f2-4cc2-8f97-095e13422bc9" containerName="swift-ring-rebalance" Mar 07 15:14:09 crc kubenswrapper[4943]: I0307 15:14:09.729456 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="95e89075-16f2-4cc2-8f97-095e13422bc9" containerName="swift-ring-rebalance" Mar 07 15:14:09 crc kubenswrapper[4943]: I0307 15:14:09.730242 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pg7qh" Mar 07 15:14:09 crc kubenswrapper[4943]: I0307 15:14:09.733364 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:14:09 crc kubenswrapper[4943]: I0307 15:14:09.734150 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:14:09 crc kubenswrapper[4943]: I0307 15:14:09.744890 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-pg7qh"] Mar 07 15:14:09 crc kubenswrapper[4943]: I0307 15:14:09.921611 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/306bbfe1-fb3f-484d-a2f6-c2d7400beb1d-etc-swift\") pod \"swift-ring-rebalance-debug-pg7qh\" (UID: \"306bbfe1-fb3f-484d-a2f6-c2d7400beb1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pg7qh" Mar 07 15:14:09 crc kubenswrapper[4943]: I0307 15:14:09.921676 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/306bbfe1-fb3f-484d-a2f6-c2d7400beb1d-swiftconf\") pod \"swift-ring-rebalance-debug-pg7qh\" (UID: \"306bbfe1-fb3f-484d-a2f6-c2d7400beb1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pg7qh" Mar 07 15:14:09 crc kubenswrapper[4943]: I0307 15:14:09.921728 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plp58\" (UniqueName: \"kubernetes.io/projected/306bbfe1-fb3f-484d-a2f6-c2d7400beb1d-kube-api-access-plp58\") pod \"swift-ring-rebalance-debug-pg7qh\" (UID: \"306bbfe1-fb3f-484d-a2f6-c2d7400beb1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pg7qh" Mar 07 15:14:09 crc kubenswrapper[4943]: I0307 15:14:09.921861 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/306bbfe1-fb3f-484d-a2f6-c2d7400beb1d-scripts\") pod \"swift-ring-rebalance-debug-pg7qh\" (UID: \"306bbfe1-fb3f-484d-a2f6-c2d7400beb1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pg7qh" Mar 07 15:14:09 crc kubenswrapper[4943]: I0307 15:14:09.921918 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/306bbfe1-fb3f-484d-a2f6-c2d7400beb1d-ring-data-devices\") pod \"swift-ring-rebalance-debug-pg7qh\" (UID: \"306bbfe1-fb3f-484d-a2f6-c2d7400beb1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pg7qh" Mar 07 15:14:09 crc kubenswrapper[4943]: I0307 15:14:09.922009 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/306bbfe1-fb3f-484d-a2f6-c2d7400beb1d-dispersionconf\") pod \"swift-ring-rebalance-debug-pg7qh\" (UID: \"306bbfe1-fb3f-484d-a2f6-c2d7400beb1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pg7qh" Mar 07 15:14:10 crc kubenswrapper[4943]: I0307 15:14:10.024137 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/306bbfe1-fb3f-484d-a2f6-c2d7400beb1d-scripts\") pod \"swift-ring-rebalance-debug-pg7qh\" (UID: \"306bbfe1-fb3f-484d-a2f6-c2d7400beb1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pg7qh" Mar 07 15:14:10 crc kubenswrapper[4943]: I0307 15:14:10.024256 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/306bbfe1-fb3f-484d-a2f6-c2d7400beb1d-ring-data-devices\") pod \"swift-ring-rebalance-debug-pg7qh\" (UID: \"306bbfe1-fb3f-484d-a2f6-c2d7400beb1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pg7qh" Mar 07 15:14:10 crc kubenswrapper[4943]: I0307 15:14:10.024332 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/306bbfe1-fb3f-484d-a2f6-c2d7400beb1d-dispersionconf\") pod \"swift-ring-rebalance-debug-pg7qh\" (UID: \"306bbfe1-fb3f-484d-a2f6-c2d7400beb1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pg7qh" Mar 07 15:14:10 crc kubenswrapper[4943]: I0307 15:14:10.024447 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/306bbfe1-fb3f-484d-a2f6-c2d7400beb1d-etc-swift\") pod \"swift-ring-rebalance-debug-pg7qh\" (UID: \"306bbfe1-fb3f-484d-a2f6-c2d7400beb1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pg7qh" Mar 07 15:14:10 crc kubenswrapper[4943]: I0307 15:14:10.024515 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/306bbfe1-fb3f-484d-a2f6-c2d7400beb1d-swiftconf\") pod \"swift-ring-rebalance-debug-pg7qh\" (UID: \"306bbfe1-fb3f-484d-a2f6-c2d7400beb1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pg7qh" Mar 07 15:14:10 crc kubenswrapper[4943]: I0307 15:14:10.024582 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plp58\" (UniqueName: \"kubernetes.io/projected/306bbfe1-fb3f-484d-a2f6-c2d7400beb1d-kube-api-access-plp58\") pod \"swift-ring-rebalance-debug-pg7qh\" (UID: \"306bbfe1-fb3f-484d-a2f6-c2d7400beb1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pg7qh" Mar 07 15:14:10 crc kubenswrapper[4943]: I0307 15:14:10.025112 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/306bbfe1-fb3f-484d-a2f6-c2d7400beb1d-etc-swift\") pod \"swift-ring-rebalance-debug-pg7qh\" (UID: \"306bbfe1-fb3f-484d-a2f6-c2d7400beb1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pg7qh" Mar 07 15:14:10 crc kubenswrapper[4943]: I0307 15:14:10.025146 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/306bbfe1-fb3f-484d-a2f6-c2d7400beb1d-ring-data-devices\") pod \"swift-ring-rebalance-debug-pg7qh\" (UID: \"306bbfe1-fb3f-484d-a2f6-c2d7400beb1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pg7qh" Mar 07 15:14:10 crc kubenswrapper[4943]: I0307 15:14:10.025153 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/306bbfe1-fb3f-484d-a2f6-c2d7400beb1d-scripts\") pod \"swift-ring-rebalance-debug-pg7qh\" (UID: \"306bbfe1-fb3f-484d-a2f6-c2d7400beb1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pg7qh" Mar 07 15:14:10 crc kubenswrapper[4943]: I0307 15:14:10.028679 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/306bbfe1-fb3f-484d-a2f6-c2d7400beb1d-dispersionconf\") pod \"swift-ring-rebalance-debug-pg7qh\" (UID: \"306bbfe1-fb3f-484d-a2f6-c2d7400beb1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pg7qh" Mar 07 15:14:10 crc kubenswrapper[4943]: I0307 15:14:10.029239 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/306bbfe1-fb3f-484d-a2f6-c2d7400beb1d-swiftconf\") pod \"swift-ring-rebalance-debug-pg7qh\" (UID: \"306bbfe1-fb3f-484d-a2f6-c2d7400beb1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pg7qh" Mar 07 15:14:10 crc kubenswrapper[4943]: I0307 15:14:10.053042 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plp58\" (UniqueName: \"kubernetes.io/projected/306bbfe1-fb3f-484d-a2f6-c2d7400beb1d-kube-api-access-plp58\") pod \"swift-ring-rebalance-debug-pg7qh\" (UID: \"306bbfe1-fb3f-484d-a2f6-c2d7400beb1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pg7qh" Mar 07 15:14:10 crc kubenswrapper[4943]: I0307 15:14:10.061706 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pg7qh" Mar 07 15:14:10 crc kubenswrapper[4943]: I0307 15:14:10.515844 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-pg7qh"] Mar 07 15:14:10 crc kubenswrapper[4943]: W0307 15:14:10.520109 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod306bbfe1_fb3f_484d_a2f6_c2d7400beb1d.slice/crio-c6936b8519488f7174182e247d12c44be767d06fb24a243d1ba3afd7e7ae30e7 WatchSource:0}: Error finding container c6936b8519488f7174182e247d12c44be767d06fb24a243d1ba3afd7e7ae30e7: Status 404 returned error can't find the container with id c6936b8519488f7174182e247d12c44be767d06fb24a243d1ba3afd7e7ae30e7 Mar 07 15:14:11 crc kubenswrapper[4943]: I0307 15:14:11.240313 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pg7qh" event={"ID":"306bbfe1-fb3f-484d-a2f6-c2d7400beb1d","Type":"ContainerStarted","Data":"5e35fd56c2804ea758b6d8b42abb3f040da0696fcaec51f9e79d144dacd0b9c9"} Mar 07 15:14:11 crc kubenswrapper[4943]: I0307 15:14:11.240618 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pg7qh" event={"ID":"306bbfe1-fb3f-484d-a2f6-c2d7400beb1d","Type":"ContainerStarted","Data":"c6936b8519488f7174182e247d12c44be767d06fb24a243d1ba3afd7e7ae30e7"} Mar 07 15:14:11 crc kubenswrapper[4943]: I0307 15:14:11.286858 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pg7qh" podStartSLOduration=2.286842354 podStartE2EDuration="2.286842354s" podCreationTimestamp="2026-03-07 15:14:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:14:11.261063372 +0000 UTC m=+2093.213199870" watchObservedRunningTime="2026-03-07 15:14:11.286842354 +0000 UTC m=+2093.238978852" Mar 07 15:14:12 crc kubenswrapper[4943]: I0307 15:14:12.256191 4943 generic.go:334] "Generic (PLEG): container finished" podID="306bbfe1-fb3f-484d-a2f6-c2d7400beb1d" containerID="5e35fd56c2804ea758b6d8b42abb3f040da0696fcaec51f9e79d144dacd0b9c9" exitCode=0 Mar 07 15:14:12 crc kubenswrapper[4943]: I0307 15:14:12.256261 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pg7qh" event={"ID":"306bbfe1-fb3f-484d-a2f6-c2d7400beb1d","Type":"ContainerDied","Data":"5e35fd56c2804ea758b6d8b42abb3f040da0696fcaec51f9e79d144dacd0b9c9"} Mar 07 15:14:13 crc kubenswrapper[4943]: I0307 15:14:13.696140 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pg7qh" Mar 07 15:14:13 crc kubenswrapper[4943]: I0307 15:14:13.738353 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-pg7qh"] Mar 07 15:14:13 crc kubenswrapper[4943]: I0307 15:14:13.746226 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-pg7qh"] Mar 07 15:14:13 crc kubenswrapper[4943]: I0307 15:14:13.893586 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/306bbfe1-fb3f-484d-a2f6-c2d7400beb1d-swiftconf\") pod \"306bbfe1-fb3f-484d-a2f6-c2d7400beb1d\" (UID: \"306bbfe1-fb3f-484d-a2f6-c2d7400beb1d\") " Mar 07 15:14:13 crc kubenswrapper[4943]: I0307 15:14:13.893633 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/306bbfe1-fb3f-484d-a2f6-c2d7400beb1d-ring-data-devices\") pod \"306bbfe1-fb3f-484d-a2f6-c2d7400beb1d\" (UID: \"306bbfe1-fb3f-484d-a2f6-c2d7400beb1d\") " Mar 07 15:14:13 crc kubenswrapper[4943]: I0307 15:14:13.893666 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/306bbfe1-fb3f-484d-a2f6-c2d7400beb1d-scripts\") pod \"306bbfe1-fb3f-484d-a2f6-c2d7400beb1d\" (UID: \"306bbfe1-fb3f-484d-a2f6-c2d7400beb1d\") " Mar 07 15:14:13 crc kubenswrapper[4943]: I0307 15:14:13.893690 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/306bbfe1-fb3f-484d-a2f6-c2d7400beb1d-etc-swift\") pod \"306bbfe1-fb3f-484d-a2f6-c2d7400beb1d\" (UID: \"306bbfe1-fb3f-484d-a2f6-c2d7400beb1d\") " Mar 07 15:14:13 crc kubenswrapper[4943]: I0307 15:14:13.893728 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/306bbfe1-fb3f-484d-a2f6-c2d7400beb1d-dispersionconf\") pod \"306bbfe1-fb3f-484d-a2f6-c2d7400beb1d\" (UID: \"306bbfe1-fb3f-484d-a2f6-c2d7400beb1d\") " Mar 07 15:14:13 crc kubenswrapper[4943]: I0307 15:14:13.893853 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-plp58\" (UniqueName: \"kubernetes.io/projected/306bbfe1-fb3f-484d-a2f6-c2d7400beb1d-kube-api-access-plp58\") pod \"306bbfe1-fb3f-484d-a2f6-c2d7400beb1d\" (UID: \"306bbfe1-fb3f-484d-a2f6-c2d7400beb1d\") " Mar 07 15:14:13 crc kubenswrapper[4943]: I0307 15:14:13.894836 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/306bbfe1-fb3f-484d-a2f6-c2d7400beb1d-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "306bbfe1-fb3f-484d-a2f6-c2d7400beb1d" (UID: "306bbfe1-fb3f-484d-a2f6-c2d7400beb1d"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:14:13 crc kubenswrapper[4943]: I0307 15:14:13.895428 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/306bbfe1-fb3f-484d-a2f6-c2d7400beb1d-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "306bbfe1-fb3f-484d-a2f6-c2d7400beb1d" (UID: "306bbfe1-fb3f-484d-a2f6-c2d7400beb1d"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:14:13 crc kubenswrapper[4943]: I0307 15:14:13.902307 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/306bbfe1-fb3f-484d-a2f6-c2d7400beb1d-kube-api-access-plp58" (OuterVolumeSpecName: "kube-api-access-plp58") pod "306bbfe1-fb3f-484d-a2f6-c2d7400beb1d" (UID: "306bbfe1-fb3f-484d-a2f6-c2d7400beb1d"). InnerVolumeSpecName "kube-api-access-plp58". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:14:13 crc kubenswrapper[4943]: I0307 15:14:13.912081 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/306bbfe1-fb3f-484d-a2f6-c2d7400beb1d-scripts" (OuterVolumeSpecName: "scripts") pod "306bbfe1-fb3f-484d-a2f6-c2d7400beb1d" (UID: "306bbfe1-fb3f-484d-a2f6-c2d7400beb1d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:14:13 crc kubenswrapper[4943]: I0307 15:14:13.932270 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/306bbfe1-fb3f-484d-a2f6-c2d7400beb1d-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "306bbfe1-fb3f-484d-a2f6-c2d7400beb1d" (UID: "306bbfe1-fb3f-484d-a2f6-c2d7400beb1d"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:14:13 crc kubenswrapper[4943]: I0307 15:14:13.937477 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/306bbfe1-fb3f-484d-a2f6-c2d7400beb1d-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "306bbfe1-fb3f-484d-a2f6-c2d7400beb1d" (UID: "306bbfe1-fb3f-484d-a2f6-c2d7400beb1d"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:14:13 crc kubenswrapper[4943]: I0307 15:14:13.996616 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/306bbfe1-fb3f-484d-a2f6-c2d7400beb1d-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:13 crc kubenswrapper[4943]: I0307 15:14:13.996676 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/306bbfe1-fb3f-484d-a2f6-c2d7400beb1d-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:13 crc kubenswrapper[4943]: I0307 15:14:13.996704 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/306bbfe1-fb3f-484d-a2f6-c2d7400beb1d-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:13 crc kubenswrapper[4943]: I0307 15:14:13.996729 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-plp58\" (UniqueName: \"kubernetes.io/projected/306bbfe1-fb3f-484d-a2f6-c2d7400beb1d-kube-api-access-plp58\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:13 crc kubenswrapper[4943]: I0307 15:14:13.996754 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/306bbfe1-fb3f-484d-a2f6-c2d7400beb1d-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:13 crc kubenswrapper[4943]: I0307 15:14:13.996776 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/306bbfe1-fb3f-484d-a2f6-c2d7400beb1d-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:14 crc kubenswrapper[4943]: I0307 15:14:14.280313 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c6936b8519488f7174182e247d12c44be767d06fb24a243d1ba3afd7e7ae30e7" Mar 07 15:14:14 crc kubenswrapper[4943]: I0307 15:14:14.280405 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pg7qh" Mar 07 15:14:14 crc kubenswrapper[4943]: I0307 15:14:14.770359 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="306bbfe1-fb3f-484d-a2f6-c2d7400beb1d" path="/var/lib/kubelet/pods/306bbfe1-fb3f-484d-a2f6-c2d7400beb1d/volumes" Mar 07 15:14:14 crc kubenswrapper[4943]: I0307 15:14:14.943281 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-wkmvm"] Mar 07 15:14:14 crc kubenswrapper[4943]: E0307 15:14:14.943878 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="306bbfe1-fb3f-484d-a2f6-c2d7400beb1d" containerName="swift-ring-rebalance" Mar 07 15:14:14 crc kubenswrapper[4943]: I0307 15:14:14.943892 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="306bbfe1-fb3f-484d-a2f6-c2d7400beb1d" containerName="swift-ring-rebalance" Mar 07 15:14:14 crc kubenswrapper[4943]: I0307 15:14:14.944070 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="306bbfe1-fb3f-484d-a2f6-c2d7400beb1d" containerName="swift-ring-rebalance" Mar 07 15:14:14 crc kubenswrapper[4943]: I0307 15:14:14.944546 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wkmvm" Mar 07 15:14:14 crc kubenswrapper[4943]: I0307 15:14:14.947248 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:14:14 crc kubenswrapper[4943]: I0307 15:14:14.947731 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:14:14 crc kubenswrapper[4943]: I0307 15:14:14.957293 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-wkmvm"] Mar 07 15:14:15 crc kubenswrapper[4943]: I0307 15:14:15.113795 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/7701a814-c454-44af-b2df-3256f4a0df75-dispersionconf\") pod \"swift-ring-rebalance-debug-wkmvm\" (UID: \"7701a814-c454-44af-b2df-3256f4a0df75\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wkmvm" Mar 07 15:14:15 crc kubenswrapper[4943]: I0307 15:14:15.114162 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/7701a814-c454-44af-b2df-3256f4a0df75-swiftconf\") pod \"swift-ring-rebalance-debug-wkmvm\" (UID: \"7701a814-c454-44af-b2df-3256f4a0df75\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wkmvm" Mar 07 15:14:15 crc kubenswrapper[4943]: I0307 15:14:15.114371 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/7701a814-c454-44af-b2df-3256f4a0df75-etc-swift\") pod \"swift-ring-rebalance-debug-wkmvm\" (UID: \"7701a814-c454-44af-b2df-3256f4a0df75\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wkmvm" Mar 07 15:14:15 crc kubenswrapper[4943]: I0307 15:14:15.114556 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8z7x2\" (UniqueName: \"kubernetes.io/projected/7701a814-c454-44af-b2df-3256f4a0df75-kube-api-access-8z7x2\") pod \"swift-ring-rebalance-debug-wkmvm\" (UID: \"7701a814-c454-44af-b2df-3256f4a0df75\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wkmvm" Mar 07 15:14:15 crc kubenswrapper[4943]: I0307 15:14:15.114755 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/7701a814-c454-44af-b2df-3256f4a0df75-ring-data-devices\") pod \"swift-ring-rebalance-debug-wkmvm\" (UID: \"7701a814-c454-44af-b2df-3256f4a0df75\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wkmvm" Mar 07 15:14:15 crc kubenswrapper[4943]: I0307 15:14:15.114839 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7701a814-c454-44af-b2df-3256f4a0df75-scripts\") pod \"swift-ring-rebalance-debug-wkmvm\" (UID: \"7701a814-c454-44af-b2df-3256f4a0df75\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wkmvm" Mar 07 15:14:15 crc kubenswrapper[4943]: I0307 15:14:15.216153 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8z7x2\" (UniqueName: \"kubernetes.io/projected/7701a814-c454-44af-b2df-3256f4a0df75-kube-api-access-8z7x2\") pod \"swift-ring-rebalance-debug-wkmvm\" (UID: \"7701a814-c454-44af-b2df-3256f4a0df75\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wkmvm" Mar 07 15:14:15 crc kubenswrapper[4943]: I0307 15:14:15.216229 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/7701a814-c454-44af-b2df-3256f4a0df75-ring-data-devices\") pod \"swift-ring-rebalance-debug-wkmvm\" (UID: \"7701a814-c454-44af-b2df-3256f4a0df75\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wkmvm" Mar 07 15:14:15 crc kubenswrapper[4943]: I0307 15:14:15.216263 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7701a814-c454-44af-b2df-3256f4a0df75-scripts\") pod \"swift-ring-rebalance-debug-wkmvm\" (UID: \"7701a814-c454-44af-b2df-3256f4a0df75\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wkmvm" Mar 07 15:14:15 crc kubenswrapper[4943]: I0307 15:14:15.216305 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/7701a814-c454-44af-b2df-3256f4a0df75-dispersionconf\") pod \"swift-ring-rebalance-debug-wkmvm\" (UID: \"7701a814-c454-44af-b2df-3256f4a0df75\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wkmvm" Mar 07 15:14:15 crc kubenswrapper[4943]: I0307 15:14:15.216381 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/7701a814-c454-44af-b2df-3256f4a0df75-swiftconf\") pod \"swift-ring-rebalance-debug-wkmvm\" (UID: \"7701a814-c454-44af-b2df-3256f4a0df75\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wkmvm" Mar 07 15:14:15 crc kubenswrapper[4943]: I0307 15:14:15.216405 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/7701a814-c454-44af-b2df-3256f4a0df75-etc-swift\") pod \"swift-ring-rebalance-debug-wkmvm\" (UID: \"7701a814-c454-44af-b2df-3256f4a0df75\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wkmvm" Mar 07 15:14:15 crc kubenswrapper[4943]: I0307 15:14:15.217151 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/7701a814-c454-44af-b2df-3256f4a0df75-etc-swift\") pod \"swift-ring-rebalance-debug-wkmvm\" (UID: \"7701a814-c454-44af-b2df-3256f4a0df75\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wkmvm" Mar 07 15:14:15 crc kubenswrapper[4943]: I0307 15:14:15.217320 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/7701a814-c454-44af-b2df-3256f4a0df75-ring-data-devices\") pod \"swift-ring-rebalance-debug-wkmvm\" (UID: \"7701a814-c454-44af-b2df-3256f4a0df75\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wkmvm" Mar 07 15:14:15 crc kubenswrapper[4943]: I0307 15:14:15.217559 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7701a814-c454-44af-b2df-3256f4a0df75-scripts\") pod \"swift-ring-rebalance-debug-wkmvm\" (UID: \"7701a814-c454-44af-b2df-3256f4a0df75\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wkmvm" Mar 07 15:14:15 crc kubenswrapper[4943]: I0307 15:14:15.222565 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/7701a814-c454-44af-b2df-3256f4a0df75-dispersionconf\") pod \"swift-ring-rebalance-debug-wkmvm\" (UID: \"7701a814-c454-44af-b2df-3256f4a0df75\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wkmvm" Mar 07 15:14:15 crc kubenswrapper[4943]: I0307 15:14:15.225606 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/7701a814-c454-44af-b2df-3256f4a0df75-swiftconf\") pod \"swift-ring-rebalance-debug-wkmvm\" (UID: \"7701a814-c454-44af-b2df-3256f4a0df75\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wkmvm" Mar 07 15:14:15 crc kubenswrapper[4943]: I0307 15:14:15.245084 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8z7x2\" (UniqueName: \"kubernetes.io/projected/7701a814-c454-44af-b2df-3256f4a0df75-kube-api-access-8z7x2\") pod \"swift-ring-rebalance-debug-wkmvm\" (UID: \"7701a814-c454-44af-b2df-3256f4a0df75\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wkmvm" Mar 07 15:14:15 crc kubenswrapper[4943]: I0307 15:14:15.259773 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wkmvm" Mar 07 15:14:15 crc kubenswrapper[4943]: W0307 15:14:15.756250 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7701a814_c454_44af_b2df_3256f4a0df75.slice/crio-70aee9132b3d9239a8dd973f224e5c56c2c4fa2cc3dd82607e35a2f0af1163f0 WatchSource:0}: Error finding container 70aee9132b3d9239a8dd973f224e5c56c2c4fa2cc3dd82607e35a2f0af1163f0: Status 404 returned error can't find the container with id 70aee9132b3d9239a8dd973f224e5c56c2c4fa2cc3dd82607e35a2f0af1163f0 Mar 07 15:14:15 crc kubenswrapper[4943]: I0307 15:14:15.757219 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-wkmvm"] Mar 07 15:14:16 crc kubenswrapper[4943]: I0307 15:14:16.298993 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wkmvm" event={"ID":"7701a814-c454-44af-b2df-3256f4a0df75","Type":"ContainerStarted","Data":"6f6cccd11766b0092fa2277ecc5517a1de2dab8f790ecc19b36566c9df30c99d"} Mar 07 15:14:16 crc kubenswrapper[4943]: I0307 15:14:16.299307 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wkmvm" event={"ID":"7701a814-c454-44af-b2df-3256f4a0df75","Type":"ContainerStarted","Data":"70aee9132b3d9239a8dd973f224e5c56c2c4fa2cc3dd82607e35a2f0af1163f0"} Mar 07 15:14:16 crc kubenswrapper[4943]: I0307 15:14:16.322194 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wkmvm" podStartSLOduration=2.322174687 podStartE2EDuration="2.322174687s" podCreationTimestamp="2026-03-07 15:14:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:14:16.318138508 +0000 UTC m=+2098.270275016" watchObservedRunningTime="2026-03-07 15:14:16.322174687 +0000 UTC m=+2098.274311185" Mar 07 15:14:17 crc kubenswrapper[4943]: I0307 15:14:17.311698 4943 generic.go:334] "Generic (PLEG): container finished" podID="7701a814-c454-44af-b2df-3256f4a0df75" containerID="6f6cccd11766b0092fa2277ecc5517a1de2dab8f790ecc19b36566c9df30c99d" exitCode=0 Mar 07 15:14:17 crc kubenswrapper[4943]: I0307 15:14:17.311834 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wkmvm" event={"ID":"7701a814-c454-44af-b2df-3256f4a0df75","Type":"ContainerDied","Data":"6f6cccd11766b0092fa2277ecc5517a1de2dab8f790ecc19b36566c9df30c99d"} Mar 07 15:14:18 crc kubenswrapper[4943]: I0307 15:14:18.674298 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wkmvm" Mar 07 15:14:18 crc kubenswrapper[4943]: I0307 15:14:18.719322 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-wkmvm"] Mar 07 15:14:18 crc kubenswrapper[4943]: I0307 15:14:18.728208 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-wkmvm"] Mar 07 15:14:18 crc kubenswrapper[4943]: I0307 15:14:18.875166 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/7701a814-c454-44af-b2df-3256f4a0df75-etc-swift\") pod \"7701a814-c454-44af-b2df-3256f4a0df75\" (UID: \"7701a814-c454-44af-b2df-3256f4a0df75\") " Mar 07 15:14:18 crc kubenswrapper[4943]: I0307 15:14:18.875207 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/7701a814-c454-44af-b2df-3256f4a0df75-dispersionconf\") pod \"7701a814-c454-44af-b2df-3256f4a0df75\" (UID: \"7701a814-c454-44af-b2df-3256f4a0df75\") " Mar 07 15:14:18 crc kubenswrapper[4943]: I0307 15:14:18.875244 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8z7x2\" (UniqueName: \"kubernetes.io/projected/7701a814-c454-44af-b2df-3256f4a0df75-kube-api-access-8z7x2\") pod \"7701a814-c454-44af-b2df-3256f4a0df75\" (UID: \"7701a814-c454-44af-b2df-3256f4a0df75\") " Mar 07 15:14:18 crc kubenswrapper[4943]: I0307 15:14:18.875266 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/7701a814-c454-44af-b2df-3256f4a0df75-swiftconf\") pod \"7701a814-c454-44af-b2df-3256f4a0df75\" (UID: \"7701a814-c454-44af-b2df-3256f4a0df75\") " Mar 07 15:14:18 crc kubenswrapper[4943]: I0307 15:14:18.875322 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7701a814-c454-44af-b2df-3256f4a0df75-scripts\") pod \"7701a814-c454-44af-b2df-3256f4a0df75\" (UID: \"7701a814-c454-44af-b2df-3256f4a0df75\") " Mar 07 15:14:18 crc kubenswrapper[4943]: I0307 15:14:18.875442 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/7701a814-c454-44af-b2df-3256f4a0df75-ring-data-devices\") pod \"7701a814-c454-44af-b2df-3256f4a0df75\" (UID: \"7701a814-c454-44af-b2df-3256f4a0df75\") " Mar 07 15:14:18 crc kubenswrapper[4943]: I0307 15:14:18.876160 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7701a814-c454-44af-b2df-3256f4a0df75-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "7701a814-c454-44af-b2df-3256f4a0df75" (UID: "7701a814-c454-44af-b2df-3256f4a0df75"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:14:18 crc kubenswrapper[4943]: I0307 15:14:18.876325 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7701a814-c454-44af-b2df-3256f4a0df75-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "7701a814-c454-44af-b2df-3256f4a0df75" (UID: "7701a814-c454-44af-b2df-3256f4a0df75"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:14:18 crc kubenswrapper[4943]: I0307 15:14:18.880264 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7701a814-c454-44af-b2df-3256f4a0df75-kube-api-access-8z7x2" (OuterVolumeSpecName: "kube-api-access-8z7x2") pod "7701a814-c454-44af-b2df-3256f4a0df75" (UID: "7701a814-c454-44af-b2df-3256f4a0df75"). InnerVolumeSpecName "kube-api-access-8z7x2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:14:18 crc kubenswrapper[4943]: I0307 15:14:18.894168 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7701a814-c454-44af-b2df-3256f4a0df75-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "7701a814-c454-44af-b2df-3256f4a0df75" (UID: "7701a814-c454-44af-b2df-3256f4a0df75"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:14:18 crc kubenswrapper[4943]: I0307 15:14:18.894590 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7701a814-c454-44af-b2df-3256f4a0df75-scripts" (OuterVolumeSpecName: "scripts") pod "7701a814-c454-44af-b2df-3256f4a0df75" (UID: "7701a814-c454-44af-b2df-3256f4a0df75"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:14:18 crc kubenswrapper[4943]: I0307 15:14:18.896125 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7701a814-c454-44af-b2df-3256f4a0df75-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "7701a814-c454-44af-b2df-3256f4a0df75" (UID: "7701a814-c454-44af-b2df-3256f4a0df75"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:14:18 crc kubenswrapper[4943]: I0307 15:14:18.977098 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/7701a814-c454-44af-b2df-3256f4a0df75-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:18 crc kubenswrapper[4943]: I0307 15:14:18.977134 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/7701a814-c454-44af-b2df-3256f4a0df75-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:18 crc kubenswrapper[4943]: I0307 15:14:18.977144 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8z7x2\" (UniqueName: \"kubernetes.io/projected/7701a814-c454-44af-b2df-3256f4a0df75-kube-api-access-8z7x2\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:18 crc kubenswrapper[4943]: I0307 15:14:18.977155 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/7701a814-c454-44af-b2df-3256f4a0df75-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:18 crc kubenswrapper[4943]: I0307 15:14:18.977164 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7701a814-c454-44af-b2df-3256f4a0df75-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:18 crc kubenswrapper[4943]: I0307 15:14:18.977172 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/7701a814-c454-44af-b2df-3256f4a0df75-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:19 crc kubenswrapper[4943]: I0307 15:14:19.334984 4943 scope.go:117] "RemoveContainer" containerID="6f6cccd11766b0092fa2277ecc5517a1de2dab8f790ecc19b36566c9df30c99d" Mar 07 15:14:19 crc kubenswrapper[4943]: I0307 15:14:19.335250 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wkmvm" Mar 07 15:14:19 crc kubenswrapper[4943]: I0307 15:14:19.896345 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-d9k5l"] Mar 07 15:14:19 crc kubenswrapper[4943]: E0307 15:14:19.896956 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7701a814-c454-44af-b2df-3256f4a0df75" containerName="swift-ring-rebalance" Mar 07 15:14:19 crc kubenswrapper[4943]: I0307 15:14:19.896969 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="7701a814-c454-44af-b2df-3256f4a0df75" containerName="swift-ring-rebalance" Mar 07 15:14:19 crc kubenswrapper[4943]: I0307 15:14:19.897116 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="7701a814-c454-44af-b2df-3256f4a0df75" containerName="swift-ring-rebalance" Mar 07 15:14:19 crc kubenswrapper[4943]: I0307 15:14:19.897591 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-d9k5l" Mar 07 15:14:19 crc kubenswrapper[4943]: I0307 15:14:19.900346 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:14:19 crc kubenswrapper[4943]: I0307 15:14:19.900395 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:14:19 crc kubenswrapper[4943]: I0307 15:14:19.915520 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-d9k5l"] Mar 07 15:14:19 crc kubenswrapper[4943]: I0307 15:14:19.991201 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6f1cf6d1-5e2d-4fac-b87d-b783b1421161-etc-swift\") pod \"swift-ring-rebalance-debug-d9k5l\" (UID: \"6f1cf6d1-5e2d-4fac-b87d-b783b1421161\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-d9k5l" Mar 07 15:14:19 crc kubenswrapper[4943]: I0307 15:14:19.991266 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6f1cf6d1-5e2d-4fac-b87d-b783b1421161-dispersionconf\") pod \"swift-ring-rebalance-debug-d9k5l\" (UID: \"6f1cf6d1-5e2d-4fac-b87d-b783b1421161\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-d9k5l" Mar 07 15:14:19 crc kubenswrapper[4943]: I0307 15:14:19.991345 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6f1cf6d1-5e2d-4fac-b87d-b783b1421161-scripts\") pod \"swift-ring-rebalance-debug-d9k5l\" (UID: \"6f1cf6d1-5e2d-4fac-b87d-b783b1421161\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-d9k5l" Mar 07 15:14:19 crc kubenswrapper[4943]: I0307 15:14:19.991511 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdb7n\" (UniqueName: \"kubernetes.io/projected/6f1cf6d1-5e2d-4fac-b87d-b783b1421161-kube-api-access-hdb7n\") pod \"swift-ring-rebalance-debug-d9k5l\" (UID: \"6f1cf6d1-5e2d-4fac-b87d-b783b1421161\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-d9k5l" Mar 07 15:14:19 crc kubenswrapper[4943]: I0307 15:14:19.991577 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6f1cf6d1-5e2d-4fac-b87d-b783b1421161-swiftconf\") pod \"swift-ring-rebalance-debug-d9k5l\" (UID: \"6f1cf6d1-5e2d-4fac-b87d-b783b1421161\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-d9k5l" Mar 07 15:14:19 crc kubenswrapper[4943]: I0307 15:14:19.991699 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6f1cf6d1-5e2d-4fac-b87d-b783b1421161-ring-data-devices\") pod \"swift-ring-rebalance-debug-d9k5l\" (UID: \"6f1cf6d1-5e2d-4fac-b87d-b783b1421161\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-d9k5l" Mar 07 15:14:20 crc kubenswrapper[4943]: I0307 15:14:20.092761 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdb7n\" (UniqueName: \"kubernetes.io/projected/6f1cf6d1-5e2d-4fac-b87d-b783b1421161-kube-api-access-hdb7n\") pod \"swift-ring-rebalance-debug-d9k5l\" (UID: \"6f1cf6d1-5e2d-4fac-b87d-b783b1421161\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-d9k5l" Mar 07 15:14:20 crc kubenswrapper[4943]: I0307 15:14:20.092849 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6f1cf6d1-5e2d-4fac-b87d-b783b1421161-swiftconf\") pod \"swift-ring-rebalance-debug-d9k5l\" (UID: \"6f1cf6d1-5e2d-4fac-b87d-b783b1421161\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-d9k5l" Mar 07 15:14:20 crc kubenswrapper[4943]: I0307 15:14:20.092920 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6f1cf6d1-5e2d-4fac-b87d-b783b1421161-ring-data-devices\") pod \"swift-ring-rebalance-debug-d9k5l\" (UID: \"6f1cf6d1-5e2d-4fac-b87d-b783b1421161\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-d9k5l" Mar 07 15:14:20 crc kubenswrapper[4943]: I0307 15:14:20.093046 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6f1cf6d1-5e2d-4fac-b87d-b783b1421161-etc-swift\") pod \"swift-ring-rebalance-debug-d9k5l\" (UID: \"6f1cf6d1-5e2d-4fac-b87d-b783b1421161\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-d9k5l" Mar 07 15:14:20 crc kubenswrapper[4943]: I0307 15:14:20.093076 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6f1cf6d1-5e2d-4fac-b87d-b783b1421161-dispersionconf\") pod \"swift-ring-rebalance-debug-d9k5l\" (UID: \"6f1cf6d1-5e2d-4fac-b87d-b783b1421161\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-d9k5l" Mar 07 15:14:20 crc kubenswrapper[4943]: I0307 15:14:20.093109 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6f1cf6d1-5e2d-4fac-b87d-b783b1421161-scripts\") pod \"swift-ring-rebalance-debug-d9k5l\" (UID: \"6f1cf6d1-5e2d-4fac-b87d-b783b1421161\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-d9k5l" Mar 07 15:14:20 crc kubenswrapper[4943]: I0307 15:14:20.093579 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6f1cf6d1-5e2d-4fac-b87d-b783b1421161-etc-swift\") pod \"swift-ring-rebalance-debug-d9k5l\" (UID: \"6f1cf6d1-5e2d-4fac-b87d-b783b1421161\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-d9k5l" Mar 07 15:14:20 crc kubenswrapper[4943]: I0307 15:14:20.094317 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6f1cf6d1-5e2d-4fac-b87d-b783b1421161-ring-data-devices\") pod \"swift-ring-rebalance-debug-d9k5l\" (UID: \"6f1cf6d1-5e2d-4fac-b87d-b783b1421161\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-d9k5l" Mar 07 15:14:20 crc kubenswrapper[4943]: I0307 15:14:20.094596 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6f1cf6d1-5e2d-4fac-b87d-b783b1421161-scripts\") pod \"swift-ring-rebalance-debug-d9k5l\" (UID: \"6f1cf6d1-5e2d-4fac-b87d-b783b1421161\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-d9k5l" Mar 07 15:14:20 crc kubenswrapper[4943]: I0307 15:14:20.097126 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6f1cf6d1-5e2d-4fac-b87d-b783b1421161-swiftconf\") pod \"swift-ring-rebalance-debug-d9k5l\" (UID: \"6f1cf6d1-5e2d-4fac-b87d-b783b1421161\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-d9k5l" Mar 07 15:14:20 crc kubenswrapper[4943]: I0307 15:14:20.097350 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6f1cf6d1-5e2d-4fac-b87d-b783b1421161-dispersionconf\") pod \"swift-ring-rebalance-debug-d9k5l\" (UID: \"6f1cf6d1-5e2d-4fac-b87d-b783b1421161\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-d9k5l" Mar 07 15:14:20 crc kubenswrapper[4943]: I0307 15:14:20.121603 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdb7n\" (UniqueName: \"kubernetes.io/projected/6f1cf6d1-5e2d-4fac-b87d-b783b1421161-kube-api-access-hdb7n\") pod \"swift-ring-rebalance-debug-d9k5l\" (UID: \"6f1cf6d1-5e2d-4fac-b87d-b783b1421161\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-d9k5l" Mar 07 15:14:20 crc kubenswrapper[4943]: I0307 15:14:20.221187 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-d9k5l" Mar 07 15:14:20 crc kubenswrapper[4943]: I0307 15:14:20.663725 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-d9k5l"] Mar 07 15:14:20 crc kubenswrapper[4943]: I0307 15:14:20.766027 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7701a814-c454-44af-b2df-3256f4a0df75" path="/var/lib/kubelet/pods/7701a814-c454-44af-b2df-3256f4a0df75/volumes" Mar 07 15:14:21 crc kubenswrapper[4943]: I0307 15:14:21.358264 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-d9k5l" event={"ID":"6f1cf6d1-5e2d-4fac-b87d-b783b1421161","Type":"ContainerStarted","Data":"15062d0afbc2b0fce1e0d349ad9933e0354c4de994c64b443f3dc6f13774de2e"} Mar 07 15:14:21 crc kubenswrapper[4943]: I0307 15:14:21.358633 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-d9k5l" event={"ID":"6f1cf6d1-5e2d-4fac-b87d-b783b1421161","Type":"ContainerStarted","Data":"4ffc7848ad9694dfee8c44c8084fd64aab44cf21d82a1f9220961e8598202896"} Mar 07 15:14:21 crc kubenswrapper[4943]: I0307 15:14:21.379498 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-d9k5l" podStartSLOduration=2.37947812 podStartE2EDuration="2.37947812s" podCreationTimestamp="2026-03-07 15:14:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:14:21.375109863 +0000 UTC m=+2103.327246441" watchObservedRunningTime="2026-03-07 15:14:21.37947812 +0000 UTC m=+2103.331614628" Mar 07 15:14:21 crc kubenswrapper[4943]: I0307 15:14:21.587412 4943 scope.go:117] "RemoveContainer" containerID="084497ed13c5cd353f7e40a35741d561595a66526e4545b3576e21cb941b9128" Mar 07 15:14:23 crc kubenswrapper[4943]: I0307 15:14:23.381094 4943 generic.go:334] "Generic (PLEG): container finished" podID="6f1cf6d1-5e2d-4fac-b87d-b783b1421161" containerID="15062d0afbc2b0fce1e0d349ad9933e0354c4de994c64b443f3dc6f13774de2e" exitCode=0 Mar 07 15:14:23 crc kubenswrapper[4943]: I0307 15:14:23.381177 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-d9k5l" event={"ID":"6f1cf6d1-5e2d-4fac-b87d-b783b1421161","Type":"ContainerDied","Data":"15062d0afbc2b0fce1e0d349ad9933e0354c4de994c64b443f3dc6f13774de2e"} Mar 07 15:14:24 crc kubenswrapper[4943]: I0307 15:14:24.810985 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-d9k5l" Mar 07 15:14:24 crc kubenswrapper[4943]: I0307 15:14:24.873204 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-d9k5l"] Mar 07 15:14:24 crc kubenswrapper[4943]: I0307 15:14:24.882198 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-d9k5l"] Mar 07 15:14:24 crc kubenswrapper[4943]: I0307 15:14:24.965532 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6f1cf6d1-5e2d-4fac-b87d-b783b1421161-etc-swift\") pod \"6f1cf6d1-5e2d-4fac-b87d-b783b1421161\" (UID: \"6f1cf6d1-5e2d-4fac-b87d-b783b1421161\") " Mar 07 15:14:24 crc kubenswrapper[4943]: I0307 15:14:24.965786 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6f1cf6d1-5e2d-4fac-b87d-b783b1421161-ring-data-devices\") pod \"6f1cf6d1-5e2d-4fac-b87d-b783b1421161\" (UID: \"6f1cf6d1-5e2d-4fac-b87d-b783b1421161\") " Mar 07 15:14:24 crc kubenswrapper[4943]: I0307 15:14:24.965882 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hdb7n\" (UniqueName: \"kubernetes.io/projected/6f1cf6d1-5e2d-4fac-b87d-b783b1421161-kube-api-access-hdb7n\") pod \"6f1cf6d1-5e2d-4fac-b87d-b783b1421161\" (UID: \"6f1cf6d1-5e2d-4fac-b87d-b783b1421161\") " Mar 07 15:14:24 crc kubenswrapper[4943]: I0307 15:14:24.965912 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6f1cf6d1-5e2d-4fac-b87d-b783b1421161-scripts\") pod \"6f1cf6d1-5e2d-4fac-b87d-b783b1421161\" (UID: \"6f1cf6d1-5e2d-4fac-b87d-b783b1421161\") " Mar 07 15:14:24 crc kubenswrapper[4943]: I0307 15:14:24.966038 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6f1cf6d1-5e2d-4fac-b87d-b783b1421161-dispersionconf\") pod \"6f1cf6d1-5e2d-4fac-b87d-b783b1421161\" (UID: \"6f1cf6d1-5e2d-4fac-b87d-b783b1421161\") " Mar 07 15:14:24 crc kubenswrapper[4943]: I0307 15:14:24.966154 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6f1cf6d1-5e2d-4fac-b87d-b783b1421161-swiftconf\") pod \"6f1cf6d1-5e2d-4fac-b87d-b783b1421161\" (UID: \"6f1cf6d1-5e2d-4fac-b87d-b783b1421161\") " Mar 07 15:14:24 crc kubenswrapper[4943]: I0307 15:14:24.967575 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f1cf6d1-5e2d-4fac-b87d-b783b1421161-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "6f1cf6d1-5e2d-4fac-b87d-b783b1421161" (UID: "6f1cf6d1-5e2d-4fac-b87d-b783b1421161"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:14:24 crc kubenswrapper[4943]: I0307 15:14:24.967966 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6f1cf6d1-5e2d-4fac-b87d-b783b1421161-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "6f1cf6d1-5e2d-4fac-b87d-b783b1421161" (UID: "6f1cf6d1-5e2d-4fac-b87d-b783b1421161"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:14:24 crc kubenswrapper[4943]: I0307 15:14:24.971114 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f1cf6d1-5e2d-4fac-b87d-b783b1421161-kube-api-access-hdb7n" (OuterVolumeSpecName: "kube-api-access-hdb7n") pod "6f1cf6d1-5e2d-4fac-b87d-b783b1421161" (UID: "6f1cf6d1-5e2d-4fac-b87d-b783b1421161"). InnerVolumeSpecName "kube-api-access-hdb7n". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:14:24 crc kubenswrapper[4943]: I0307 15:14:24.987002 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f1cf6d1-5e2d-4fac-b87d-b783b1421161-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "6f1cf6d1-5e2d-4fac-b87d-b783b1421161" (UID: "6f1cf6d1-5e2d-4fac-b87d-b783b1421161"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:14:24 crc kubenswrapper[4943]: I0307 15:14:24.990664 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f1cf6d1-5e2d-4fac-b87d-b783b1421161-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "6f1cf6d1-5e2d-4fac-b87d-b783b1421161" (UID: "6f1cf6d1-5e2d-4fac-b87d-b783b1421161"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:14:25 crc kubenswrapper[4943]: I0307 15:14:25.005087 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f1cf6d1-5e2d-4fac-b87d-b783b1421161-scripts" (OuterVolumeSpecName: "scripts") pod "6f1cf6d1-5e2d-4fac-b87d-b783b1421161" (UID: "6f1cf6d1-5e2d-4fac-b87d-b783b1421161"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:14:25 crc kubenswrapper[4943]: I0307 15:14:25.071793 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6f1cf6d1-5e2d-4fac-b87d-b783b1421161-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:25 crc kubenswrapper[4943]: I0307 15:14:25.071832 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6f1cf6d1-5e2d-4fac-b87d-b783b1421161-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:25 crc kubenswrapper[4943]: I0307 15:14:25.071841 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6f1cf6d1-5e2d-4fac-b87d-b783b1421161-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:25 crc kubenswrapper[4943]: I0307 15:14:25.071850 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6f1cf6d1-5e2d-4fac-b87d-b783b1421161-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:25 crc kubenswrapper[4943]: I0307 15:14:25.071861 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hdb7n\" (UniqueName: \"kubernetes.io/projected/6f1cf6d1-5e2d-4fac-b87d-b783b1421161-kube-api-access-hdb7n\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:25 crc kubenswrapper[4943]: I0307 15:14:25.071870 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6f1cf6d1-5e2d-4fac-b87d-b783b1421161-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:25 crc kubenswrapper[4943]: I0307 15:14:25.408273 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4ffc7848ad9694dfee8c44c8084fd64aab44cf21d82a1f9220961e8598202896" Mar 07 15:14:25 crc kubenswrapper[4943]: I0307 15:14:25.408561 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-d9k5l" Mar 07 15:14:26 crc kubenswrapper[4943]: I0307 15:14:26.021808 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-44b8v"] Mar 07 15:14:26 crc kubenswrapper[4943]: E0307 15:14:26.022275 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f1cf6d1-5e2d-4fac-b87d-b783b1421161" containerName="swift-ring-rebalance" Mar 07 15:14:26 crc kubenswrapper[4943]: I0307 15:14:26.022298 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f1cf6d1-5e2d-4fac-b87d-b783b1421161" containerName="swift-ring-rebalance" Mar 07 15:14:26 crc kubenswrapper[4943]: I0307 15:14:26.022555 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f1cf6d1-5e2d-4fac-b87d-b783b1421161" containerName="swift-ring-rebalance" Mar 07 15:14:26 crc kubenswrapper[4943]: I0307 15:14:26.023507 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-44b8v" Mar 07 15:14:26 crc kubenswrapper[4943]: I0307 15:14:26.026190 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:14:26 crc kubenswrapper[4943]: I0307 15:14:26.027042 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:14:26 crc kubenswrapper[4943]: I0307 15:14:26.036047 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-44b8v"] Mar 07 15:14:26 crc kubenswrapper[4943]: I0307 15:14:26.187103 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/24776482-420f-4a27-96d0-9a77a66d2c1f-scripts\") pod \"swift-ring-rebalance-debug-44b8v\" (UID: \"24776482-420f-4a27-96d0-9a77a66d2c1f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-44b8v" Mar 07 15:14:26 crc kubenswrapper[4943]: I0307 15:14:26.187712 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2ddh\" (UniqueName: \"kubernetes.io/projected/24776482-420f-4a27-96d0-9a77a66d2c1f-kube-api-access-j2ddh\") pod \"swift-ring-rebalance-debug-44b8v\" (UID: \"24776482-420f-4a27-96d0-9a77a66d2c1f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-44b8v" Mar 07 15:14:26 crc kubenswrapper[4943]: I0307 15:14:26.188032 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/24776482-420f-4a27-96d0-9a77a66d2c1f-dispersionconf\") pod \"swift-ring-rebalance-debug-44b8v\" (UID: \"24776482-420f-4a27-96d0-9a77a66d2c1f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-44b8v" Mar 07 15:14:26 crc kubenswrapper[4943]: I0307 15:14:26.188332 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/24776482-420f-4a27-96d0-9a77a66d2c1f-ring-data-devices\") pod \"swift-ring-rebalance-debug-44b8v\" (UID: \"24776482-420f-4a27-96d0-9a77a66d2c1f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-44b8v" Mar 07 15:14:26 crc kubenswrapper[4943]: I0307 15:14:26.188636 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/24776482-420f-4a27-96d0-9a77a66d2c1f-swiftconf\") pod \"swift-ring-rebalance-debug-44b8v\" (UID: \"24776482-420f-4a27-96d0-9a77a66d2c1f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-44b8v" Mar 07 15:14:26 crc kubenswrapper[4943]: I0307 15:14:26.188865 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/24776482-420f-4a27-96d0-9a77a66d2c1f-etc-swift\") pod \"swift-ring-rebalance-debug-44b8v\" (UID: \"24776482-420f-4a27-96d0-9a77a66d2c1f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-44b8v" Mar 07 15:14:26 crc kubenswrapper[4943]: I0307 15:14:26.291773 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/24776482-420f-4a27-96d0-9a77a66d2c1f-etc-swift\") pod \"swift-ring-rebalance-debug-44b8v\" (UID: \"24776482-420f-4a27-96d0-9a77a66d2c1f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-44b8v" Mar 07 15:14:26 crc kubenswrapper[4943]: I0307 15:14:26.291030 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/24776482-420f-4a27-96d0-9a77a66d2c1f-etc-swift\") pod \"swift-ring-rebalance-debug-44b8v\" (UID: \"24776482-420f-4a27-96d0-9a77a66d2c1f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-44b8v" Mar 07 15:14:26 crc kubenswrapper[4943]: I0307 15:14:26.293087 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/24776482-420f-4a27-96d0-9a77a66d2c1f-swiftconf\") pod \"swift-ring-rebalance-debug-44b8v\" (UID: \"24776482-420f-4a27-96d0-9a77a66d2c1f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-44b8v" Mar 07 15:14:26 crc kubenswrapper[4943]: I0307 15:14:26.294326 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/24776482-420f-4a27-96d0-9a77a66d2c1f-scripts\") pod \"swift-ring-rebalance-debug-44b8v\" (UID: \"24776482-420f-4a27-96d0-9a77a66d2c1f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-44b8v" Mar 07 15:14:26 crc kubenswrapper[4943]: I0307 15:14:26.295542 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/24776482-420f-4a27-96d0-9a77a66d2c1f-scripts\") pod \"swift-ring-rebalance-debug-44b8v\" (UID: \"24776482-420f-4a27-96d0-9a77a66d2c1f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-44b8v" Mar 07 15:14:26 crc kubenswrapper[4943]: I0307 15:14:26.296192 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2ddh\" (UniqueName: \"kubernetes.io/projected/24776482-420f-4a27-96d0-9a77a66d2c1f-kube-api-access-j2ddh\") pod \"swift-ring-rebalance-debug-44b8v\" (UID: \"24776482-420f-4a27-96d0-9a77a66d2c1f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-44b8v" Mar 07 15:14:26 crc kubenswrapper[4943]: I0307 15:14:26.296865 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/24776482-420f-4a27-96d0-9a77a66d2c1f-dispersionconf\") pod \"swift-ring-rebalance-debug-44b8v\" (UID: \"24776482-420f-4a27-96d0-9a77a66d2c1f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-44b8v" Mar 07 15:14:26 crc kubenswrapper[4943]: I0307 15:14:26.299702 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/24776482-420f-4a27-96d0-9a77a66d2c1f-ring-data-devices\") pod \"swift-ring-rebalance-debug-44b8v\" (UID: \"24776482-420f-4a27-96d0-9a77a66d2c1f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-44b8v" Mar 07 15:14:26 crc kubenswrapper[4943]: I0307 15:14:26.300884 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/24776482-420f-4a27-96d0-9a77a66d2c1f-ring-data-devices\") pod \"swift-ring-rebalance-debug-44b8v\" (UID: \"24776482-420f-4a27-96d0-9a77a66d2c1f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-44b8v" Mar 07 15:14:26 crc kubenswrapper[4943]: I0307 15:14:26.299151 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/24776482-420f-4a27-96d0-9a77a66d2c1f-swiftconf\") pod \"swift-ring-rebalance-debug-44b8v\" (UID: \"24776482-420f-4a27-96d0-9a77a66d2c1f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-44b8v" Mar 07 15:14:26 crc kubenswrapper[4943]: I0307 15:14:26.302469 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/24776482-420f-4a27-96d0-9a77a66d2c1f-dispersionconf\") pod \"swift-ring-rebalance-debug-44b8v\" (UID: \"24776482-420f-4a27-96d0-9a77a66d2c1f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-44b8v" Mar 07 15:14:26 crc kubenswrapper[4943]: I0307 15:14:26.318414 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2ddh\" (UniqueName: \"kubernetes.io/projected/24776482-420f-4a27-96d0-9a77a66d2c1f-kube-api-access-j2ddh\") pod \"swift-ring-rebalance-debug-44b8v\" (UID: \"24776482-420f-4a27-96d0-9a77a66d2c1f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-44b8v" Mar 07 15:14:26 crc kubenswrapper[4943]: I0307 15:14:26.363333 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-44b8v" Mar 07 15:14:26 crc kubenswrapper[4943]: I0307 15:14:26.767192 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f1cf6d1-5e2d-4fac-b87d-b783b1421161" path="/var/lib/kubelet/pods/6f1cf6d1-5e2d-4fac-b87d-b783b1421161/volumes" Mar 07 15:14:26 crc kubenswrapper[4943]: I0307 15:14:26.920488 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-44b8v"] Mar 07 15:14:26 crc kubenswrapper[4943]: W0307 15:14:26.928168 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod24776482_420f_4a27_96d0_9a77a66d2c1f.slice/crio-78b05ddfddafed6e28ddcae4c0d3a7b96fff426e4bf6f0104a635b35b75a2c0e WatchSource:0}: Error finding container 78b05ddfddafed6e28ddcae4c0d3a7b96fff426e4bf6f0104a635b35b75a2c0e: Status 404 returned error can't find the container with id 78b05ddfddafed6e28ddcae4c0d3a7b96fff426e4bf6f0104a635b35b75a2c0e Mar 07 15:14:27 crc kubenswrapper[4943]: I0307 15:14:27.437316 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-44b8v" event={"ID":"24776482-420f-4a27-96d0-9a77a66d2c1f","Type":"ContainerStarted","Data":"6b42dfcd362de484f87d1214e9c529ee4264aa022e5f08f253e6268435f5a2de"} Mar 07 15:14:27 crc kubenswrapper[4943]: I0307 15:14:27.437510 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-44b8v" event={"ID":"24776482-420f-4a27-96d0-9a77a66d2c1f","Type":"ContainerStarted","Data":"78b05ddfddafed6e28ddcae4c0d3a7b96fff426e4bf6f0104a635b35b75a2c0e"} Mar 07 15:14:27 crc kubenswrapper[4943]: I0307 15:14:27.477259 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-44b8v" podStartSLOduration=2.477202772 podStartE2EDuration="2.477202772s" podCreationTimestamp="2026-03-07 15:14:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:14:27.462766478 +0000 UTC m=+2109.414902996" watchObservedRunningTime="2026-03-07 15:14:27.477202772 +0000 UTC m=+2109.429339310" Mar 07 15:14:28 crc kubenswrapper[4943]: I0307 15:14:28.448305 4943 generic.go:334] "Generic (PLEG): container finished" podID="24776482-420f-4a27-96d0-9a77a66d2c1f" containerID="6b42dfcd362de484f87d1214e9c529ee4264aa022e5f08f253e6268435f5a2de" exitCode=0 Mar 07 15:14:28 crc kubenswrapper[4943]: I0307 15:14:28.448349 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-44b8v" event={"ID":"24776482-420f-4a27-96d0-9a77a66d2c1f","Type":"ContainerDied","Data":"6b42dfcd362de484f87d1214e9c529ee4264aa022e5f08f253e6268435f5a2de"} Mar 07 15:14:29 crc kubenswrapper[4943]: I0307 15:14:29.761640 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-44b8v" Mar 07 15:14:29 crc kubenswrapper[4943]: I0307 15:14:29.801162 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-44b8v"] Mar 07 15:14:29 crc kubenswrapper[4943]: I0307 15:14:29.808433 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-44b8v"] Mar 07 15:14:29 crc kubenswrapper[4943]: I0307 15:14:29.882092 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/24776482-420f-4a27-96d0-9a77a66d2c1f-dispersionconf\") pod \"24776482-420f-4a27-96d0-9a77a66d2c1f\" (UID: \"24776482-420f-4a27-96d0-9a77a66d2c1f\") " Mar 07 15:14:29 crc kubenswrapper[4943]: I0307 15:14:29.882125 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/24776482-420f-4a27-96d0-9a77a66d2c1f-swiftconf\") pod \"24776482-420f-4a27-96d0-9a77a66d2c1f\" (UID: \"24776482-420f-4a27-96d0-9a77a66d2c1f\") " Mar 07 15:14:29 crc kubenswrapper[4943]: I0307 15:14:29.882155 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/24776482-420f-4a27-96d0-9a77a66d2c1f-scripts\") pod \"24776482-420f-4a27-96d0-9a77a66d2c1f\" (UID: \"24776482-420f-4a27-96d0-9a77a66d2c1f\") " Mar 07 15:14:29 crc kubenswrapper[4943]: I0307 15:14:29.882189 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j2ddh\" (UniqueName: \"kubernetes.io/projected/24776482-420f-4a27-96d0-9a77a66d2c1f-kube-api-access-j2ddh\") pod \"24776482-420f-4a27-96d0-9a77a66d2c1f\" (UID: \"24776482-420f-4a27-96d0-9a77a66d2c1f\") " Mar 07 15:14:29 crc kubenswrapper[4943]: I0307 15:14:29.882240 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/24776482-420f-4a27-96d0-9a77a66d2c1f-etc-swift\") pod \"24776482-420f-4a27-96d0-9a77a66d2c1f\" (UID: \"24776482-420f-4a27-96d0-9a77a66d2c1f\") " Mar 07 15:14:29 crc kubenswrapper[4943]: I0307 15:14:29.882314 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/24776482-420f-4a27-96d0-9a77a66d2c1f-ring-data-devices\") pod \"24776482-420f-4a27-96d0-9a77a66d2c1f\" (UID: \"24776482-420f-4a27-96d0-9a77a66d2c1f\") " Mar 07 15:14:29 crc kubenswrapper[4943]: I0307 15:14:29.883743 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/24776482-420f-4a27-96d0-9a77a66d2c1f-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "24776482-420f-4a27-96d0-9a77a66d2c1f" (UID: "24776482-420f-4a27-96d0-9a77a66d2c1f"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:14:29 crc kubenswrapper[4943]: I0307 15:14:29.884089 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/24776482-420f-4a27-96d0-9a77a66d2c1f-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "24776482-420f-4a27-96d0-9a77a66d2c1f" (UID: "24776482-420f-4a27-96d0-9a77a66d2c1f"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:14:29 crc kubenswrapper[4943]: I0307 15:14:29.895198 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24776482-420f-4a27-96d0-9a77a66d2c1f-kube-api-access-j2ddh" (OuterVolumeSpecName: "kube-api-access-j2ddh") pod "24776482-420f-4a27-96d0-9a77a66d2c1f" (UID: "24776482-420f-4a27-96d0-9a77a66d2c1f"). InnerVolumeSpecName "kube-api-access-j2ddh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:14:29 crc kubenswrapper[4943]: I0307 15:14:29.904959 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/24776482-420f-4a27-96d0-9a77a66d2c1f-scripts" (OuterVolumeSpecName: "scripts") pod "24776482-420f-4a27-96d0-9a77a66d2c1f" (UID: "24776482-420f-4a27-96d0-9a77a66d2c1f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:14:29 crc kubenswrapper[4943]: I0307 15:14:29.908302 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24776482-420f-4a27-96d0-9a77a66d2c1f-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "24776482-420f-4a27-96d0-9a77a66d2c1f" (UID: "24776482-420f-4a27-96d0-9a77a66d2c1f"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:14:29 crc kubenswrapper[4943]: I0307 15:14:29.910251 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24776482-420f-4a27-96d0-9a77a66d2c1f-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "24776482-420f-4a27-96d0-9a77a66d2c1f" (UID: "24776482-420f-4a27-96d0-9a77a66d2c1f"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:14:29 crc kubenswrapper[4943]: I0307 15:14:29.983753 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/24776482-420f-4a27-96d0-9a77a66d2c1f-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:29 crc kubenswrapper[4943]: I0307 15:14:29.983797 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/24776482-420f-4a27-96d0-9a77a66d2c1f-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:29 crc kubenswrapper[4943]: I0307 15:14:29.983811 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/24776482-420f-4a27-96d0-9a77a66d2c1f-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:29 crc kubenswrapper[4943]: I0307 15:14:29.983822 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/24776482-420f-4a27-96d0-9a77a66d2c1f-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:29 crc kubenswrapper[4943]: I0307 15:14:29.983836 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j2ddh\" (UniqueName: \"kubernetes.io/projected/24776482-420f-4a27-96d0-9a77a66d2c1f-kube-api-access-j2ddh\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:29 crc kubenswrapper[4943]: I0307 15:14:29.983848 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/24776482-420f-4a27-96d0-9a77a66d2c1f-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:30 crc kubenswrapper[4943]: I0307 15:14:30.469909 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="78b05ddfddafed6e28ddcae4c0d3a7b96fff426e4bf6f0104a635b35b75a2c0e" Mar 07 15:14:30 crc kubenswrapper[4943]: I0307 15:14:30.470007 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-44b8v" Mar 07 15:14:30 crc kubenswrapper[4943]: I0307 15:14:30.770573 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24776482-420f-4a27-96d0-9a77a66d2c1f" path="/var/lib/kubelet/pods/24776482-420f-4a27-96d0-9a77a66d2c1f/volumes" Mar 07 15:14:30 crc kubenswrapper[4943]: I0307 15:14:30.938311 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-xsgmp"] Mar 07 15:14:30 crc kubenswrapper[4943]: E0307 15:14:30.938752 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24776482-420f-4a27-96d0-9a77a66d2c1f" containerName="swift-ring-rebalance" Mar 07 15:14:30 crc kubenswrapper[4943]: I0307 15:14:30.938776 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="24776482-420f-4a27-96d0-9a77a66d2c1f" containerName="swift-ring-rebalance" Mar 07 15:14:30 crc kubenswrapper[4943]: I0307 15:14:30.939196 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="24776482-420f-4a27-96d0-9a77a66d2c1f" containerName="swift-ring-rebalance" Mar 07 15:14:30 crc kubenswrapper[4943]: I0307 15:14:30.940065 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-xsgmp" Mar 07 15:14:30 crc kubenswrapper[4943]: I0307 15:14:30.945993 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-xsgmp"] Mar 07 15:14:30 crc kubenswrapper[4943]: I0307 15:14:30.946755 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:14:30 crc kubenswrapper[4943]: I0307 15:14:30.946756 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:14:31 crc kubenswrapper[4943]: I0307 15:14:31.098023 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b133b8cb-ab81-4a40-83ec-d75ec9b46d7e-etc-swift\") pod \"swift-ring-rebalance-debug-xsgmp\" (UID: \"b133b8cb-ab81-4a40-83ec-d75ec9b46d7e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xsgmp" Mar 07 15:14:31 crc kubenswrapper[4943]: I0307 15:14:31.098199 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b133b8cb-ab81-4a40-83ec-d75ec9b46d7e-scripts\") pod \"swift-ring-rebalance-debug-xsgmp\" (UID: \"b133b8cb-ab81-4a40-83ec-d75ec9b46d7e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xsgmp" Mar 07 15:14:31 crc kubenswrapper[4943]: I0307 15:14:31.098313 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44dsv\" (UniqueName: \"kubernetes.io/projected/b133b8cb-ab81-4a40-83ec-d75ec9b46d7e-kube-api-access-44dsv\") pod \"swift-ring-rebalance-debug-xsgmp\" (UID: \"b133b8cb-ab81-4a40-83ec-d75ec9b46d7e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xsgmp" Mar 07 15:14:31 crc kubenswrapper[4943]: I0307 15:14:31.098409 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b133b8cb-ab81-4a40-83ec-d75ec9b46d7e-swiftconf\") pod \"swift-ring-rebalance-debug-xsgmp\" (UID: \"b133b8cb-ab81-4a40-83ec-d75ec9b46d7e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xsgmp" Mar 07 15:14:31 crc kubenswrapper[4943]: I0307 15:14:31.098564 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b133b8cb-ab81-4a40-83ec-d75ec9b46d7e-dispersionconf\") pod \"swift-ring-rebalance-debug-xsgmp\" (UID: \"b133b8cb-ab81-4a40-83ec-d75ec9b46d7e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xsgmp" Mar 07 15:14:31 crc kubenswrapper[4943]: I0307 15:14:31.098742 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b133b8cb-ab81-4a40-83ec-d75ec9b46d7e-ring-data-devices\") pod \"swift-ring-rebalance-debug-xsgmp\" (UID: \"b133b8cb-ab81-4a40-83ec-d75ec9b46d7e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xsgmp" Mar 07 15:14:31 crc kubenswrapper[4943]: I0307 15:14:31.199734 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b133b8cb-ab81-4a40-83ec-d75ec9b46d7e-swiftconf\") pod \"swift-ring-rebalance-debug-xsgmp\" (UID: \"b133b8cb-ab81-4a40-83ec-d75ec9b46d7e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xsgmp" Mar 07 15:14:31 crc kubenswrapper[4943]: I0307 15:14:31.199791 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b133b8cb-ab81-4a40-83ec-d75ec9b46d7e-dispersionconf\") pod \"swift-ring-rebalance-debug-xsgmp\" (UID: \"b133b8cb-ab81-4a40-83ec-d75ec9b46d7e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xsgmp" Mar 07 15:14:31 crc kubenswrapper[4943]: I0307 15:14:31.199849 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b133b8cb-ab81-4a40-83ec-d75ec9b46d7e-ring-data-devices\") pod \"swift-ring-rebalance-debug-xsgmp\" (UID: \"b133b8cb-ab81-4a40-83ec-d75ec9b46d7e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xsgmp" Mar 07 15:14:31 crc kubenswrapper[4943]: I0307 15:14:31.199908 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b133b8cb-ab81-4a40-83ec-d75ec9b46d7e-etc-swift\") pod \"swift-ring-rebalance-debug-xsgmp\" (UID: \"b133b8cb-ab81-4a40-83ec-d75ec9b46d7e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xsgmp" Mar 07 15:14:31 crc kubenswrapper[4943]: I0307 15:14:31.199958 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b133b8cb-ab81-4a40-83ec-d75ec9b46d7e-scripts\") pod \"swift-ring-rebalance-debug-xsgmp\" (UID: \"b133b8cb-ab81-4a40-83ec-d75ec9b46d7e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xsgmp" Mar 07 15:14:31 crc kubenswrapper[4943]: I0307 15:14:31.200031 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44dsv\" (UniqueName: \"kubernetes.io/projected/b133b8cb-ab81-4a40-83ec-d75ec9b46d7e-kube-api-access-44dsv\") pod \"swift-ring-rebalance-debug-xsgmp\" (UID: \"b133b8cb-ab81-4a40-83ec-d75ec9b46d7e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xsgmp" Mar 07 15:14:31 crc kubenswrapper[4943]: I0307 15:14:31.200617 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b133b8cb-ab81-4a40-83ec-d75ec9b46d7e-etc-swift\") pod \"swift-ring-rebalance-debug-xsgmp\" (UID: \"b133b8cb-ab81-4a40-83ec-d75ec9b46d7e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xsgmp" Mar 07 15:14:31 crc kubenswrapper[4943]: I0307 15:14:31.200956 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b133b8cb-ab81-4a40-83ec-d75ec9b46d7e-scripts\") pod \"swift-ring-rebalance-debug-xsgmp\" (UID: \"b133b8cb-ab81-4a40-83ec-d75ec9b46d7e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xsgmp" Mar 07 15:14:31 crc kubenswrapper[4943]: I0307 15:14:31.201411 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b133b8cb-ab81-4a40-83ec-d75ec9b46d7e-ring-data-devices\") pod \"swift-ring-rebalance-debug-xsgmp\" (UID: \"b133b8cb-ab81-4a40-83ec-d75ec9b46d7e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xsgmp" Mar 07 15:14:31 crc kubenswrapper[4943]: I0307 15:14:31.204567 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b133b8cb-ab81-4a40-83ec-d75ec9b46d7e-swiftconf\") pod \"swift-ring-rebalance-debug-xsgmp\" (UID: \"b133b8cb-ab81-4a40-83ec-d75ec9b46d7e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xsgmp" Mar 07 15:14:31 crc kubenswrapper[4943]: I0307 15:14:31.205291 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b133b8cb-ab81-4a40-83ec-d75ec9b46d7e-dispersionconf\") pod \"swift-ring-rebalance-debug-xsgmp\" (UID: \"b133b8cb-ab81-4a40-83ec-d75ec9b46d7e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xsgmp" Mar 07 15:14:31 crc kubenswrapper[4943]: I0307 15:14:31.232118 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44dsv\" (UniqueName: \"kubernetes.io/projected/b133b8cb-ab81-4a40-83ec-d75ec9b46d7e-kube-api-access-44dsv\") pod \"swift-ring-rebalance-debug-xsgmp\" (UID: \"b133b8cb-ab81-4a40-83ec-d75ec9b46d7e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xsgmp" Mar 07 15:14:31 crc kubenswrapper[4943]: I0307 15:14:31.272656 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-xsgmp" Mar 07 15:14:31 crc kubenswrapper[4943]: I0307 15:14:31.518104 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-xsgmp"] Mar 07 15:14:31 crc kubenswrapper[4943]: W0307 15:14:31.526002 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb133b8cb_ab81_4a40_83ec_d75ec9b46d7e.slice/crio-edb9737075db3440631464b5b7ebe53b3865a9a1310b43a53023c854a81642a9 WatchSource:0}: Error finding container edb9737075db3440631464b5b7ebe53b3865a9a1310b43a53023c854a81642a9: Status 404 returned error can't find the container with id edb9737075db3440631464b5b7ebe53b3865a9a1310b43a53023c854a81642a9 Mar 07 15:14:32 crc kubenswrapper[4943]: I0307 15:14:32.493132 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-xsgmp" event={"ID":"b133b8cb-ab81-4a40-83ec-d75ec9b46d7e","Type":"ContainerStarted","Data":"f4752e484e0ccbe4760ca08e61be7c10dc834d665b03dd98b780ea21f8da2243"} Mar 07 15:14:32 crc kubenswrapper[4943]: I0307 15:14:32.493442 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-xsgmp" event={"ID":"b133b8cb-ab81-4a40-83ec-d75ec9b46d7e","Type":"ContainerStarted","Data":"edb9737075db3440631464b5b7ebe53b3865a9a1310b43a53023c854a81642a9"} Mar 07 15:14:32 crc kubenswrapper[4943]: I0307 15:14:32.511966 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-xsgmp" podStartSLOduration=2.511950851 podStartE2EDuration="2.511950851s" podCreationTimestamp="2026-03-07 15:14:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:14:32.508430335 +0000 UTC m=+2114.460566833" watchObservedRunningTime="2026-03-07 15:14:32.511950851 +0000 UTC m=+2114.464087349" Mar 07 15:14:33 crc kubenswrapper[4943]: I0307 15:14:33.504177 4943 generic.go:334] "Generic (PLEG): container finished" podID="b133b8cb-ab81-4a40-83ec-d75ec9b46d7e" containerID="f4752e484e0ccbe4760ca08e61be7c10dc834d665b03dd98b780ea21f8da2243" exitCode=0 Mar 07 15:14:33 crc kubenswrapper[4943]: I0307 15:14:33.504227 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-xsgmp" event={"ID":"b133b8cb-ab81-4a40-83ec-d75ec9b46d7e","Type":"ContainerDied","Data":"f4752e484e0ccbe4760ca08e61be7c10dc834d665b03dd98b780ea21f8da2243"} Mar 07 15:14:34 crc kubenswrapper[4943]: I0307 15:14:34.892417 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-xsgmp" Mar 07 15:14:34 crc kubenswrapper[4943]: I0307 15:14:34.942050 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-xsgmp"] Mar 07 15:14:34 crc kubenswrapper[4943]: I0307 15:14:34.949831 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-xsgmp"] Mar 07 15:14:35 crc kubenswrapper[4943]: I0307 15:14:35.058667 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b133b8cb-ab81-4a40-83ec-d75ec9b46d7e-dispersionconf\") pod \"b133b8cb-ab81-4a40-83ec-d75ec9b46d7e\" (UID: \"b133b8cb-ab81-4a40-83ec-d75ec9b46d7e\") " Mar 07 15:14:35 crc kubenswrapper[4943]: I0307 15:14:35.058732 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b133b8cb-ab81-4a40-83ec-d75ec9b46d7e-ring-data-devices\") pod \"b133b8cb-ab81-4a40-83ec-d75ec9b46d7e\" (UID: \"b133b8cb-ab81-4a40-83ec-d75ec9b46d7e\") " Mar 07 15:14:35 crc kubenswrapper[4943]: I0307 15:14:35.058786 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b133b8cb-ab81-4a40-83ec-d75ec9b46d7e-scripts\") pod \"b133b8cb-ab81-4a40-83ec-d75ec9b46d7e\" (UID: \"b133b8cb-ab81-4a40-83ec-d75ec9b46d7e\") " Mar 07 15:14:35 crc kubenswrapper[4943]: I0307 15:14:35.058823 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-44dsv\" (UniqueName: \"kubernetes.io/projected/b133b8cb-ab81-4a40-83ec-d75ec9b46d7e-kube-api-access-44dsv\") pod \"b133b8cb-ab81-4a40-83ec-d75ec9b46d7e\" (UID: \"b133b8cb-ab81-4a40-83ec-d75ec9b46d7e\") " Mar 07 15:14:35 crc kubenswrapper[4943]: I0307 15:14:35.058868 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b133b8cb-ab81-4a40-83ec-d75ec9b46d7e-etc-swift\") pod \"b133b8cb-ab81-4a40-83ec-d75ec9b46d7e\" (UID: \"b133b8cb-ab81-4a40-83ec-d75ec9b46d7e\") " Mar 07 15:14:35 crc kubenswrapper[4943]: I0307 15:14:35.058918 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b133b8cb-ab81-4a40-83ec-d75ec9b46d7e-swiftconf\") pod \"b133b8cb-ab81-4a40-83ec-d75ec9b46d7e\" (UID: \"b133b8cb-ab81-4a40-83ec-d75ec9b46d7e\") " Mar 07 15:14:35 crc kubenswrapper[4943]: I0307 15:14:35.064439 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b133b8cb-ab81-4a40-83ec-d75ec9b46d7e-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "b133b8cb-ab81-4a40-83ec-d75ec9b46d7e" (UID: "b133b8cb-ab81-4a40-83ec-d75ec9b46d7e"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:14:35 crc kubenswrapper[4943]: I0307 15:14:35.064570 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b133b8cb-ab81-4a40-83ec-d75ec9b46d7e-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "b133b8cb-ab81-4a40-83ec-d75ec9b46d7e" (UID: "b133b8cb-ab81-4a40-83ec-d75ec9b46d7e"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:14:35 crc kubenswrapper[4943]: I0307 15:14:35.066304 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b133b8cb-ab81-4a40-83ec-d75ec9b46d7e-kube-api-access-44dsv" (OuterVolumeSpecName: "kube-api-access-44dsv") pod "b133b8cb-ab81-4a40-83ec-d75ec9b46d7e" (UID: "b133b8cb-ab81-4a40-83ec-d75ec9b46d7e"). InnerVolumeSpecName "kube-api-access-44dsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:14:35 crc kubenswrapper[4943]: I0307 15:14:35.089200 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b133b8cb-ab81-4a40-83ec-d75ec9b46d7e-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "b133b8cb-ab81-4a40-83ec-d75ec9b46d7e" (UID: "b133b8cb-ab81-4a40-83ec-d75ec9b46d7e"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:14:35 crc kubenswrapper[4943]: I0307 15:14:35.089761 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b133b8cb-ab81-4a40-83ec-d75ec9b46d7e-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "b133b8cb-ab81-4a40-83ec-d75ec9b46d7e" (UID: "b133b8cb-ab81-4a40-83ec-d75ec9b46d7e"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:14:35 crc kubenswrapper[4943]: I0307 15:14:35.090281 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b133b8cb-ab81-4a40-83ec-d75ec9b46d7e-scripts" (OuterVolumeSpecName: "scripts") pod "b133b8cb-ab81-4a40-83ec-d75ec9b46d7e" (UID: "b133b8cb-ab81-4a40-83ec-d75ec9b46d7e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:14:35 crc kubenswrapper[4943]: I0307 15:14:35.160566 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b133b8cb-ab81-4a40-83ec-d75ec9b46d7e-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:35 crc kubenswrapper[4943]: I0307 15:14:35.160601 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b133b8cb-ab81-4a40-83ec-d75ec9b46d7e-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:35 crc kubenswrapper[4943]: I0307 15:14:35.160614 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b133b8cb-ab81-4a40-83ec-d75ec9b46d7e-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:35 crc kubenswrapper[4943]: I0307 15:14:35.160623 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b133b8cb-ab81-4a40-83ec-d75ec9b46d7e-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:35 crc kubenswrapper[4943]: I0307 15:14:35.160632 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-44dsv\" (UniqueName: \"kubernetes.io/projected/b133b8cb-ab81-4a40-83ec-d75ec9b46d7e-kube-api-access-44dsv\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:35 crc kubenswrapper[4943]: I0307 15:14:35.160640 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b133b8cb-ab81-4a40-83ec-d75ec9b46d7e-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:35 crc kubenswrapper[4943]: I0307 15:14:35.527046 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="edb9737075db3440631464b5b7ebe53b3865a9a1310b43a53023c854a81642a9" Mar 07 15:14:35 crc kubenswrapper[4943]: I0307 15:14:35.527134 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-xsgmp" Mar 07 15:14:36 crc kubenswrapper[4943]: I0307 15:14:36.168985 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-dm9wg"] Mar 07 15:14:36 crc kubenswrapper[4943]: E0307 15:14:36.169285 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b133b8cb-ab81-4a40-83ec-d75ec9b46d7e" containerName="swift-ring-rebalance" Mar 07 15:14:36 crc kubenswrapper[4943]: I0307 15:14:36.169298 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="b133b8cb-ab81-4a40-83ec-d75ec9b46d7e" containerName="swift-ring-rebalance" Mar 07 15:14:36 crc kubenswrapper[4943]: I0307 15:14:36.169420 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="b133b8cb-ab81-4a40-83ec-d75ec9b46d7e" containerName="swift-ring-rebalance" Mar 07 15:14:36 crc kubenswrapper[4943]: I0307 15:14:36.169886 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-dm9wg" Mar 07 15:14:36 crc kubenswrapper[4943]: I0307 15:14:36.175082 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:14:36 crc kubenswrapper[4943]: I0307 15:14:36.175169 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:14:36 crc kubenswrapper[4943]: I0307 15:14:36.195620 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-dm9wg"] Mar 07 15:14:36 crc kubenswrapper[4943]: I0307 15:14:36.279017 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ee702163-4146-4052-9e5c-e215e0d6420f-dispersionconf\") pod \"swift-ring-rebalance-debug-dm9wg\" (UID: \"ee702163-4146-4052-9e5c-e215e0d6420f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dm9wg" Mar 07 15:14:36 crc kubenswrapper[4943]: I0307 15:14:36.279074 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ee702163-4146-4052-9e5c-e215e0d6420f-etc-swift\") pod \"swift-ring-rebalance-debug-dm9wg\" (UID: \"ee702163-4146-4052-9e5c-e215e0d6420f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dm9wg" Mar 07 15:14:36 crc kubenswrapper[4943]: I0307 15:14:36.279105 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ee702163-4146-4052-9e5c-e215e0d6420f-scripts\") pod \"swift-ring-rebalance-debug-dm9wg\" (UID: \"ee702163-4146-4052-9e5c-e215e0d6420f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dm9wg" Mar 07 15:14:36 crc kubenswrapper[4943]: I0307 15:14:36.279155 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ee702163-4146-4052-9e5c-e215e0d6420f-swiftconf\") pod \"swift-ring-rebalance-debug-dm9wg\" (UID: \"ee702163-4146-4052-9e5c-e215e0d6420f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dm9wg" Mar 07 15:14:36 crc kubenswrapper[4943]: I0307 15:14:36.279175 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ee702163-4146-4052-9e5c-e215e0d6420f-ring-data-devices\") pod \"swift-ring-rebalance-debug-dm9wg\" (UID: \"ee702163-4146-4052-9e5c-e215e0d6420f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dm9wg" Mar 07 15:14:36 crc kubenswrapper[4943]: I0307 15:14:36.279195 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbv6w\" (UniqueName: \"kubernetes.io/projected/ee702163-4146-4052-9e5c-e215e0d6420f-kube-api-access-hbv6w\") pod \"swift-ring-rebalance-debug-dm9wg\" (UID: \"ee702163-4146-4052-9e5c-e215e0d6420f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dm9wg" Mar 07 15:14:36 crc kubenswrapper[4943]: I0307 15:14:36.381388 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ee702163-4146-4052-9e5c-e215e0d6420f-dispersionconf\") pod \"swift-ring-rebalance-debug-dm9wg\" (UID: \"ee702163-4146-4052-9e5c-e215e0d6420f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dm9wg" Mar 07 15:14:36 crc kubenswrapper[4943]: I0307 15:14:36.381656 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ee702163-4146-4052-9e5c-e215e0d6420f-etc-swift\") pod \"swift-ring-rebalance-debug-dm9wg\" (UID: \"ee702163-4146-4052-9e5c-e215e0d6420f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dm9wg" Mar 07 15:14:36 crc kubenswrapper[4943]: I0307 15:14:36.381686 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ee702163-4146-4052-9e5c-e215e0d6420f-scripts\") pod \"swift-ring-rebalance-debug-dm9wg\" (UID: \"ee702163-4146-4052-9e5c-e215e0d6420f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dm9wg" Mar 07 15:14:36 crc kubenswrapper[4943]: I0307 15:14:36.381727 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ee702163-4146-4052-9e5c-e215e0d6420f-swiftconf\") pod \"swift-ring-rebalance-debug-dm9wg\" (UID: \"ee702163-4146-4052-9e5c-e215e0d6420f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dm9wg" Mar 07 15:14:36 crc kubenswrapper[4943]: I0307 15:14:36.381743 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ee702163-4146-4052-9e5c-e215e0d6420f-ring-data-devices\") pod \"swift-ring-rebalance-debug-dm9wg\" (UID: \"ee702163-4146-4052-9e5c-e215e0d6420f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dm9wg" Mar 07 15:14:36 crc kubenswrapper[4943]: I0307 15:14:36.381761 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbv6w\" (UniqueName: \"kubernetes.io/projected/ee702163-4146-4052-9e5c-e215e0d6420f-kube-api-access-hbv6w\") pod \"swift-ring-rebalance-debug-dm9wg\" (UID: \"ee702163-4146-4052-9e5c-e215e0d6420f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dm9wg" Mar 07 15:14:36 crc kubenswrapper[4943]: I0307 15:14:36.382315 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ee702163-4146-4052-9e5c-e215e0d6420f-etc-swift\") pod \"swift-ring-rebalance-debug-dm9wg\" (UID: \"ee702163-4146-4052-9e5c-e215e0d6420f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dm9wg" Mar 07 15:14:36 crc kubenswrapper[4943]: I0307 15:14:36.382871 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ee702163-4146-4052-9e5c-e215e0d6420f-scripts\") pod \"swift-ring-rebalance-debug-dm9wg\" (UID: \"ee702163-4146-4052-9e5c-e215e0d6420f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dm9wg" Mar 07 15:14:36 crc kubenswrapper[4943]: I0307 15:14:36.383011 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ee702163-4146-4052-9e5c-e215e0d6420f-ring-data-devices\") pod \"swift-ring-rebalance-debug-dm9wg\" (UID: \"ee702163-4146-4052-9e5c-e215e0d6420f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dm9wg" Mar 07 15:14:36 crc kubenswrapper[4943]: I0307 15:14:36.386405 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ee702163-4146-4052-9e5c-e215e0d6420f-swiftconf\") pod \"swift-ring-rebalance-debug-dm9wg\" (UID: \"ee702163-4146-4052-9e5c-e215e0d6420f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dm9wg" Mar 07 15:14:36 crc kubenswrapper[4943]: I0307 15:14:36.388207 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ee702163-4146-4052-9e5c-e215e0d6420f-dispersionconf\") pod \"swift-ring-rebalance-debug-dm9wg\" (UID: \"ee702163-4146-4052-9e5c-e215e0d6420f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dm9wg" Mar 07 15:14:36 crc kubenswrapper[4943]: I0307 15:14:36.404198 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbv6w\" (UniqueName: \"kubernetes.io/projected/ee702163-4146-4052-9e5c-e215e0d6420f-kube-api-access-hbv6w\") pod \"swift-ring-rebalance-debug-dm9wg\" (UID: \"ee702163-4146-4052-9e5c-e215e0d6420f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dm9wg" Mar 07 15:14:36 crc kubenswrapper[4943]: I0307 15:14:36.486105 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-dm9wg" Mar 07 15:14:36 crc kubenswrapper[4943]: I0307 15:14:36.762725 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b133b8cb-ab81-4a40-83ec-d75ec9b46d7e" path="/var/lib/kubelet/pods/b133b8cb-ab81-4a40-83ec-d75ec9b46d7e/volumes" Mar 07 15:14:36 crc kubenswrapper[4943]: I0307 15:14:36.997724 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-dm9wg"] Mar 07 15:14:37 crc kubenswrapper[4943]: I0307 15:14:37.552470 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-dm9wg" event={"ID":"ee702163-4146-4052-9e5c-e215e0d6420f","Type":"ContainerStarted","Data":"79267f52fc465e59ae9257b9280f1da6610e63b4134f932f0f536fa3f529ec0d"} Mar 07 15:14:37 crc kubenswrapper[4943]: I0307 15:14:37.552902 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-dm9wg" event={"ID":"ee702163-4146-4052-9e5c-e215e0d6420f","Type":"ContainerStarted","Data":"33bd0b8504d88c621eb267e86df991d073f6268a25dc7bc68cabded99c4af749"} Mar 07 15:14:37 crc kubenswrapper[4943]: I0307 15:14:37.577859 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-dm9wg" podStartSLOduration=1.577840143 podStartE2EDuration="1.577840143s" podCreationTimestamp="2026-03-07 15:14:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:14:37.571920998 +0000 UTC m=+2119.524057496" watchObservedRunningTime="2026-03-07 15:14:37.577840143 +0000 UTC m=+2119.529976641" Mar 07 15:14:38 crc kubenswrapper[4943]: I0307 15:14:38.566583 4943 generic.go:334] "Generic (PLEG): container finished" podID="ee702163-4146-4052-9e5c-e215e0d6420f" containerID="79267f52fc465e59ae9257b9280f1da6610e63b4134f932f0f536fa3f529ec0d" exitCode=0 Mar 07 15:14:38 crc kubenswrapper[4943]: I0307 15:14:38.566627 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-dm9wg" event={"ID":"ee702163-4146-4052-9e5c-e215e0d6420f","Type":"ContainerDied","Data":"79267f52fc465e59ae9257b9280f1da6610e63b4134f932f0f536fa3f529ec0d"} Mar 07 15:14:39 crc kubenswrapper[4943]: I0307 15:14:39.953490 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-dm9wg" Mar 07 15:14:39 crc kubenswrapper[4943]: I0307 15:14:39.999881 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-dm9wg"] Mar 07 15:14:40 crc kubenswrapper[4943]: I0307 15:14:40.007392 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-dm9wg"] Mar 07 15:14:40 crc kubenswrapper[4943]: I0307 15:14:40.048847 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ee702163-4146-4052-9e5c-e215e0d6420f-etc-swift\") pod \"ee702163-4146-4052-9e5c-e215e0d6420f\" (UID: \"ee702163-4146-4052-9e5c-e215e0d6420f\") " Mar 07 15:14:40 crc kubenswrapper[4943]: I0307 15:14:40.048975 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ee702163-4146-4052-9e5c-e215e0d6420f-dispersionconf\") pod \"ee702163-4146-4052-9e5c-e215e0d6420f\" (UID: \"ee702163-4146-4052-9e5c-e215e0d6420f\") " Mar 07 15:14:40 crc kubenswrapper[4943]: I0307 15:14:40.049025 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ee702163-4146-4052-9e5c-e215e0d6420f-scripts\") pod \"ee702163-4146-4052-9e5c-e215e0d6420f\" (UID: \"ee702163-4146-4052-9e5c-e215e0d6420f\") " Mar 07 15:14:40 crc kubenswrapper[4943]: I0307 15:14:40.049052 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ee702163-4146-4052-9e5c-e215e0d6420f-swiftconf\") pod \"ee702163-4146-4052-9e5c-e215e0d6420f\" (UID: \"ee702163-4146-4052-9e5c-e215e0d6420f\") " Mar 07 15:14:40 crc kubenswrapper[4943]: I0307 15:14:40.049075 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ee702163-4146-4052-9e5c-e215e0d6420f-ring-data-devices\") pod \"ee702163-4146-4052-9e5c-e215e0d6420f\" (UID: \"ee702163-4146-4052-9e5c-e215e0d6420f\") " Mar 07 15:14:40 crc kubenswrapper[4943]: I0307 15:14:40.049098 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hbv6w\" (UniqueName: \"kubernetes.io/projected/ee702163-4146-4052-9e5c-e215e0d6420f-kube-api-access-hbv6w\") pod \"ee702163-4146-4052-9e5c-e215e0d6420f\" (UID: \"ee702163-4146-4052-9e5c-e215e0d6420f\") " Mar 07 15:14:40 crc kubenswrapper[4943]: I0307 15:14:40.049593 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee702163-4146-4052-9e5c-e215e0d6420f-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "ee702163-4146-4052-9e5c-e215e0d6420f" (UID: "ee702163-4146-4052-9e5c-e215e0d6420f"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:14:40 crc kubenswrapper[4943]: I0307 15:14:40.049759 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee702163-4146-4052-9e5c-e215e0d6420f-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "ee702163-4146-4052-9e5c-e215e0d6420f" (UID: "ee702163-4146-4052-9e5c-e215e0d6420f"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:14:40 crc kubenswrapper[4943]: I0307 15:14:40.050089 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ee702163-4146-4052-9e5c-e215e0d6420f-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:40 crc kubenswrapper[4943]: I0307 15:14:40.050109 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ee702163-4146-4052-9e5c-e215e0d6420f-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:40 crc kubenswrapper[4943]: I0307 15:14:40.056521 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee702163-4146-4052-9e5c-e215e0d6420f-kube-api-access-hbv6w" (OuterVolumeSpecName: "kube-api-access-hbv6w") pod "ee702163-4146-4052-9e5c-e215e0d6420f" (UID: "ee702163-4146-4052-9e5c-e215e0d6420f"). InnerVolumeSpecName "kube-api-access-hbv6w". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:14:40 crc kubenswrapper[4943]: I0307 15:14:40.069191 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee702163-4146-4052-9e5c-e215e0d6420f-scripts" (OuterVolumeSpecName: "scripts") pod "ee702163-4146-4052-9e5c-e215e0d6420f" (UID: "ee702163-4146-4052-9e5c-e215e0d6420f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:14:40 crc kubenswrapper[4943]: I0307 15:14:40.069370 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee702163-4146-4052-9e5c-e215e0d6420f-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "ee702163-4146-4052-9e5c-e215e0d6420f" (UID: "ee702163-4146-4052-9e5c-e215e0d6420f"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:14:40 crc kubenswrapper[4943]: I0307 15:14:40.091300 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee702163-4146-4052-9e5c-e215e0d6420f-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "ee702163-4146-4052-9e5c-e215e0d6420f" (UID: "ee702163-4146-4052-9e5c-e215e0d6420f"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:14:40 crc kubenswrapper[4943]: I0307 15:14:40.151518 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ee702163-4146-4052-9e5c-e215e0d6420f-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:40 crc kubenswrapper[4943]: I0307 15:14:40.151556 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ee702163-4146-4052-9e5c-e215e0d6420f-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:40 crc kubenswrapper[4943]: I0307 15:14:40.151565 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ee702163-4146-4052-9e5c-e215e0d6420f-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:40 crc kubenswrapper[4943]: I0307 15:14:40.151574 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hbv6w\" (UniqueName: \"kubernetes.io/projected/ee702163-4146-4052-9e5c-e215e0d6420f-kube-api-access-hbv6w\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:40 crc kubenswrapper[4943]: I0307 15:14:40.590759 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="33bd0b8504d88c621eb267e86df991d073f6268a25dc7bc68cabded99c4af749" Mar 07 15:14:40 crc kubenswrapper[4943]: I0307 15:14:40.590850 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-dm9wg" Mar 07 15:14:40 crc kubenswrapper[4943]: I0307 15:14:40.771338 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee702163-4146-4052-9e5c-e215e0d6420f" path="/var/lib/kubelet/pods/ee702163-4146-4052-9e5c-e215e0d6420f/volumes" Mar 07 15:14:41 crc kubenswrapper[4943]: I0307 15:14:41.130970 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-hsrq7"] Mar 07 15:14:41 crc kubenswrapper[4943]: E0307 15:14:41.131395 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee702163-4146-4052-9e5c-e215e0d6420f" containerName="swift-ring-rebalance" Mar 07 15:14:41 crc kubenswrapper[4943]: I0307 15:14:41.131418 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee702163-4146-4052-9e5c-e215e0d6420f" containerName="swift-ring-rebalance" Mar 07 15:14:41 crc kubenswrapper[4943]: I0307 15:14:41.131672 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee702163-4146-4052-9e5c-e215e0d6420f" containerName="swift-ring-rebalance" Mar 07 15:14:41 crc kubenswrapper[4943]: I0307 15:14:41.132404 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-hsrq7" Mar 07 15:14:41 crc kubenswrapper[4943]: I0307 15:14:41.135590 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:14:41 crc kubenswrapper[4943]: I0307 15:14:41.135664 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:14:41 crc kubenswrapper[4943]: I0307 15:14:41.141668 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-hsrq7"] Mar 07 15:14:41 crc kubenswrapper[4943]: I0307 15:14:41.267369 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/71db4d2a-006a-478a-8445-3ace5f5082ab-scripts\") pod \"swift-ring-rebalance-debug-hsrq7\" (UID: \"71db4d2a-006a-478a-8445-3ace5f5082ab\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hsrq7" Mar 07 15:14:41 crc kubenswrapper[4943]: I0307 15:14:41.267439 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/71db4d2a-006a-478a-8445-3ace5f5082ab-swiftconf\") pod \"swift-ring-rebalance-debug-hsrq7\" (UID: \"71db4d2a-006a-478a-8445-3ace5f5082ab\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hsrq7" Mar 07 15:14:41 crc kubenswrapper[4943]: I0307 15:14:41.267561 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/71db4d2a-006a-478a-8445-3ace5f5082ab-dispersionconf\") pod \"swift-ring-rebalance-debug-hsrq7\" (UID: \"71db4d2a-006a-478a-8445-3ace5f5082ab\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hsrq7" Mar 07 15:14:41 crc kubenswrapper[4943]: I0307 15:14:41.267676 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/71db4d2a-006a-478a-8445-3ace5f5082ab-etc-swift\") pod \"swift-ring-rebalance-debug-hsrq7\" (UID: \"71db4d2a-006a-478a-8445-3ace5f5082ab\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hsrq7" Mar 07 15:14:41 crc kubenswrapper[4943]: I0307 15:14:41.267750 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xm9n8\" (UniqueName: \"kubernetes.io/projected/71db4d2a-006a-478a-8445-3ace5f5082ab-kube-api-access-xm9n8\") pod \"swift-ring-rebalance-debug-hsrq7\" (UID: \"71db4d2a-006a-478a-8445-3ace5f5082ab\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hsrq7" Mar 07 15:14:41 crc kubenswrapper[4943]: I0307 15:14:41.267855 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/71db4d2a-006a-478a-8445-3ace5f5082ab-ring-data-devices\") pod \"swift-ring-rebalance-debug-hsrq7\" (UID: \"71db4d2a-006a-478a-8445-3ace5f5082ab\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hsrq7" Mar 07 15:14:41 crc kubenswrapper[4943]: I0307 15:14:41.369973 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/71db4d2a-006a-478a-8445-3ace5f5082ab-etc-swift\") pod \"swift-ring-rebalance-debug-hsrq7\" (UID: \"71db4d2a-006a-478a-8445-3ace5f5082ab\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hsrq7" Mar 07 15:14:41 crc kubenswrapper[4943]: I0307 15:14:41.370113 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xm9n8\" (UniqueName: \"kubernetes.io/projected/71db4d2a-006a-478a-8445-3ace5f5082ab-kube-api-access-xm9n8\") pod \"swift-ring-rebalance-debug-hsrq7\" (UID: \"71db4d2a-006a-478a-8445-3ace5f5082ab\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hsrq7" Mar 07 15:14:41 crc kubenswrapper[4943]: I0307 15:14:41.370214 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/71db4d2a-006a-478a-8445-3ace5f5082ab-ring-data-devices\") pod \"swift-ring-rebalance-debug-hsrq7\" (UID: \"71db4d2a-006a-478a-8445-3ace5f5082ab\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hsrq7" Mar 07 15:14:41 crc kubenswrapper[4943]: I0307 15:14:41.370377 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/71db4d2a-006a-478a-8445-3ace5f5082ab-scripts\") pod \"swift-ring-rebalance-debug-hsrq7\" (UID: \"71db4d2a-006a-478a-8445-3ace5f5082ab\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hsrq7" Mar 07 15:14:41 crc kubenswrapper[4943]: I0307 15:14:41.371254 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/71db4d2a-006a-478a-8445-3ace5f5082ab-etc-swift\") pod \"swift-ring-rebalance-debug-hsrq7\" (UID: \"71db4d2a-006a-478a-8445-3ace5f5082ab\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hsrq7" Mar 07 15:14:41 crc kubenswrapper[4943]: I0307 15:14:41.371473 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/71db4d2a-006a-478a-8445-3ace5f5082ab-ring-data-devices\") pod \"swift-ring-rebalance-debug-hsrq7\" (UID: \"71db4d2a-006a-478a-8445-3ace5f5082ab\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hsrq7" Mar 07 15:14:41 crc kubenswrapper[4943]: I0307 15:14:41.371563 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/71db4d2a-006a-478a-8445-3ace5f5082ab-scripts\") pod \"swift-ring-rebalance-debug-hsrq7\" (UID: \"71db4d2a-006a-478a-8445-3ace5f5082ab\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hsrq7" Mar 07 15:14:41 crc kubenswrapper[4943]: I0307 15:14:41.371820 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/71db4d2a-006a-478a-8445-3ace5f5082ab-swiftconf\") pod \"swift-ring-rebalance-debug-hsrq7\" (UID: \"71db4d2a-006a-478a-8445-3ace5f5082ab\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hsrq7" Mar 07 15:14:41 crc kubenswrapper[4943]: I0307 15:14:41.373113 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/71db4d2a-006a-478a-8445-3ace5f5082ab-dispersionconf\") pod \"swift-ring-rebalance-debug-hsrq7\" (UID: \"71db4d2a-006a-478a-8445-3ace5f5082ab\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hsrq7" Mar 07 15:14:41 crc kubenswrapper[4943]: I0307 15:14:41.388900 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/71db4d2a-006a-478a-8445-3ace5f5082ab-dispersionconf\") pod \"swift-ring-rebalance-debug-hsrq7\" (UID: \"71db4d2a-006a-478a-8445-3ace5f5082ab\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hsrq7" Mar 07 15:14:41 crc kubenswrapper[4943]: I0307 15:14:41.391734 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/71db4d2a-006a-478a-8445-3ace5f5082ab-swiftconf\") pod \"swift-ring-rebalance-debug-hsrq7\" (UID: \"71db4d2a-006a-478a-8445-3ace5f5082ab\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hsrq7" Mar 07 15:14:41 crc kubenswrapper[4943]: I0307 15:14:41.395984 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xm9n8\" (UniqueName: \"kubernetes.io/projected/71db4d2a-006a-478a-8445-3ace5f5082ab-kube-api-access-xm9n8\") pod \"swift-ring-rebalance-debug-hsrq7\" (UID: \"71db4d2a-006a-478a-8445-3ace5f5082ab\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hsrq7" Mar 07 15:14:41 crc kubenswrapper[4943]: I0307 15:14:41.490096 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-hsrq7" Mar 07 15:14:42 crc kubenswrapper[4943]: I0307 15:14:42.009102 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-hsrq7"] Mar 07 15:14:42 crc kubenswrapper[4943]: W0307 15:14:42.015428 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod71db4d2a_006a_478a_8445_3ace5f5082ab.slice/crio-b64e2554711e4c7f079a3b68e9a0328040e99bfb1d2d1d1a5891485e772eeb42 WatchSource:0}: Error finding container b64e2554711e4c7f079a3b68e9a0328040e99bfb1d2d1d1a5891485e772eeb42: Status 404 returned error can't find the container with id b64e2554711e4c7f079a3b68e9a0328040e99bfb1d2d1d1a5891485e772eeb42 Mar 07 15:14:42 crc kubenswrapper[4943]: I0307 15:14:42.620489 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-hsrq7" event={"ID":"71db4d2a-006a-478a-8445-3ace5f5082ab","Type":"ContainerStarted","Data":"c4c7288d2582ac16fdf710eca40a5eccdaa8f97875f3ce161802f9ef99ca3ed9"} Mar 07 15:14:42 crc kubenswrapper[4943]: I0307 15:14:42.620873 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-hsrq7" event={"ID":"71db4d2a-006a-478a-8445-3ace5f5082ab","Type":"ContainerStarted","Data":"b64e2554711e4c7f079a3b68e9a0328040e99bfb1d2d1d1a5891485e772eeb42"} Mar 07 15:14:42 crc kubenswrapper[4943]: I0307 15:14:42.650873 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-hsrq7" podStartSLOduration=1.650857071 podStartE2EDuration="1.650857071s" podCreationTimestamp="2026-03-07 15:14:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:14:42.643998903 +0000 UTC m=+2124.596135441" watchObservedRunningTime="2026-03-07 15:14:42.650857071 +0000 UTC m=+2124.602993579" Mar 07 15:14:43 crc kubenswrapper[4943]: I0307 15:14:43.632231 4943 generic.go:334] "Generic (PLEG): container finished" podID="71db4d2a-006a-478a-8445-3ace5f5082ab" containerID="c4c7288d2582ac16fdf710eca40a5eccdaa8f97875f3ce161802f9ef99ca3ed9" exitCode=0 Mar 07 15:14:43 crc kubenswrapper[4943]: I0307 15:14:43.632636 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-hsrq7" event={"ID":"71db4d2a-006a-478a-8445-3ace5f5082ab","Type":"ContainerDied","Data":"c4c7288d2582ac16fdf710eca40a5eccdaa8f97875f3ce161802f9ef99ca3ed9"} Mar 07 15:14:44 crc kubenswrapper[4943]: I0307 15:14:44.981651 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-hsrq7" Mar 07 15:14:45 crc kubenswrapper[4943]: I0307 15:14:45.023167 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-hsrq7"] Mar 07 15:14:45 crc kubenswrapper[4943]: I0307 15:14:45.033053 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-hsrq7"] Mar 07 15:14:45 crc kubenswrapper[4943]: I0307 15:14:45.134773 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/71db4d2a-006a-478a-8445-3ace5f5082ab-dispersionconf\") pod \"71db4d2a-006a-478a-8445-3ace5f5082ab\" (UID: \"71db4d2a-006a-478a-8445-3ace5f5082ab\") " Mar 07 15:14:45 crc kubenswrapper[4943]: I0307 15:14:45.134859 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/71db4d2a-006a-478a-8445-3ace5f5082ab-ring-data-devices\") pod \"71db4d2a-006a-478a-8445-3ace5f5082ab\" (UID: \"71db4d2a-006a-478a-8445-3ace5f5082ab\") " Mar 07 15:14:45 crc kubenswrapper[4943]: I0307 15:14:45.134908 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/71db4d2a-006a-478a-8445-3ace5f5082ab-etc-swift\") pod \"71db4d2a-006a-478a-8445-3ace5f5082ab\" (UID: \"71db4d2a-006a-478a-8445-3ace5f5082ab\") " Mar 07 15:14:45 crc kubenswrapper[4943]: I0307 15:14:45.134968 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/71db4d2a-006a-478a-8445-3ace5f5082ab-scripts\") pod \"71db4d2a-006a-478a-8445-3ace5f5082ab\" (UID: \"71db4d2a-006a-478a-8445-3ace5f5082ab\") " Mar 07 15:14:45 crc kubenswrapper[4943]: I0307 15:14:45.135077 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xm9n8\" (UniqueName: \"kubernetes.io/projected/71db4d2a-006a-478a-8445-3ace5f5082ab-kube-api-access-xm9n8\") pod \"71db4d2a-006a-478a-8445-3ace5f5082ab\" (UID: \"71db4d2a-006a-478a-8445-3ace5f5082ab\") " Mar 07 15:14:45 crc kubenswrapper[4943]: I0307 15:14:45.135174 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/71db4d2a-006a-478a-8445-3ace5f5082ab-swiftconf\") pod \"71db4d2a-006a-478a-8445-3ace5f5082ab\" (UID: \"71db4d2a-006a-478a-8445-3ace5f5082ab\") " Mar 07 15:14:45 crc kubenswrapper[4943]: I0307 15:14:45.135865 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71db4d2a-006a-478a-8445-3ace5f5082ab-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "71db4d2a-006a-478a-8445-3ace5f5082ab" (UID: "71db4d2a-006a-478a-8445-3ace5f5082ab"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:14:45 crc kubenswrapper[4943]: I0307 15:14:45.136469 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/71db4d2a-006a-478a-8445-3ace5f5082ab-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:45 crc kubenswrapper[4943]: I0307 15:14:45.136858 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71db4d2a-006a-478a-8445-3ace5f5082ab-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "71db4d2a-006a-478a-8445-3ace5f5082ab" (UID: "71db4d2a-006a-478a-8445-3ace5f5082ab"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:14:45 crc kubenswrapper[4943]: I0307 15:14:45.141418 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71db4d2a-006a-478a-8445-3ace5f5082ab-kube-api-access-xm9n8" (OuterVolumeSpecName: "kube-api-access-xm9n8") pod "71db4d2a-006a-478a-8445-3ace5f5082ab" (UID: "71db4d2a-006a-478a-8445-3ace5f5082ab"). InnerVolumeSpecName "kube-api-access-xm9n8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:14:45 crc kubenswrapper[4943]: I0307 15:14:45.159427 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71db4d2a-006a-478a-8445-3ace5f5082ab-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "71db4d2a-006a-478a-8445-3ace5f5082ab" (UID: "71db4d2a-006a-478a-8445-3ace5f5082ab"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:14:45 crc kubenswrapper[4943]: I0307 15:14:45.171233 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71db4d2a-006a-478a-8445-3ace5f5082ab-scripts" (OuterVolumeSpecName: "scripts") pod "71db4d2a-006a-478a-8445-3ace5f5082ab" (UID: "71db4d2a-006a-478a-8445-3ace5f5082ab"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:14:45 crc kubenswrapper[4943]: I0307 15:14:45.175029 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71db4d2a-006a-478a-8445-3ace5f5082ab-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "71db4d2a-006a-478a-8445-3ace5f5082ab" (UID: "71db4d2a-006a-478a-8445-3ace5f5082ab"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:14:45 crc kubenswrapper[4943]: I0307 15:14:45.238348 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/71db4d2a-006a-478a-8445-3ace5f5082ab-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:45 crc kubenswrapper[4943]: I0307 15:14:45.238380 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/71db4d2a-006a-478a-8445-3ace5f5082ab-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:45 crc kubenswrapper[4943]: I0307 15:14:45.238390 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xm9n8\" (UniqueName: \"kubernetes.io/projected/71db4d2a-006a-478a-8445-3ace5f5082ab-kube-api-access-xm9n8\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:45 crc kubenswrapper[4943]: I0307 15:14:45.238404 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/71db4d2a-006a-478a-8445-3ace5f5082ab-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:45 crc kubenswrapper[4943]: I0307 15:14:45.238412 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/71db4d2a-006a-478a-8445-3ace5f5082ab-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:45 crc kubenswrapper[4943]: I0307 15:14:45.658432 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b64e2554711e4c7f079a3b68e9a0328040e99bfb1d2d1d1a5891485e772eeb42" Mar 07 15:14:45 crc kubenswrapper[4943]: I0307 15:14:45.658574 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-hsrq7" Mar 07 15:14:46 crc kubenswrapper[4943]: I0307 15:14:46.197866 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-gq462"] Mar 07 15:14:46 crc kubenswrapper[4943]: E0307 15:14:46.198239 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71db4d2a-006a-478a-8445-3ace5f5082ab" containerName="swift-ring-rebalance" Mar 07 15:14:46 crc kubenswrapper[4943]: I0307 15:14:46.198254 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="71db4d2a-006a-478a-8445-3ace5f5082ab" containerName="swift-ring-rebalance" Mar 07 15:14:46 crc kubenswrapper[4943]: I0307 15:14:46.198443 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="71db4d2a-006a-478a-8445-3ace5f5082ab" containerName="swift-ring-rebalance" Mar 07 15:14:46 crc kubenswrapper[4943]: I0307 15:14:46.199000 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-gq462" Mar 07 15:14:46 crc kubenswrapper[4943]: I0307 15:14:46.203646 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:14:46 crc kubenswrapper[4943]: I0307 15:14:46.205396 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:14:46 crc kubenswrapper[4943]: I0307 15:14:46.218398 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-gq462"] Mar 07 15:14:46 crc kubenswrapper[4943]: I0307 15:14:46.358917 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bc01b720-2ff8-4c06-8b02-ee74e1b8f439-swiftconf\") pod \"swift-ring-rebalance-debug-gq462\" (UID: \"bc01b720-2ff8-4c06-8b02-ee74e1b8f439\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gq462" Mar 07 15:14:46 crc kubenswrapper[4943]: I0307 15:14:46.358992 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcxgf\" (UniqueName: \"kubernetes.io/projected/bc01b720-2ff8-4c06-8b02-ee74e1b8f439-kube-api-access-wcxgf\") pod \"swift-ring-rebalance-debug-gq462\" (UID: \"bc01b720-2ff8-4c06-8b02-ee74e1b8f439\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gq462" Mar 07 15:14:46 crc kubenswrapper[4943]: I0307 15:14:46.359023 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bc01b720-2ff8-4c06-8b02-ee74e1b8f439-dispersionconf\") pod \"swift-ring-rebalance-debug-gq462\" (UID: \"bc01b720-2ff8-4c06-8b02-ee74e1b8f439\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gq462" Mar 07 15:14:46 crc kubenswrapper[4943]: I0307 15:14:46.359357 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bc01b720-2ff8-4c06-8b02-ee74e1b8f439-etc-swift\") pod \"swift-ring-rebalance-debug-gq462\" (UID: \"bc01b720-2ff8-4c06-8b02-ee74e1b8f439\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gq462" Mar 07 15:14:46 crc kubenswrapper[4943]: I0307 15:14:46.359418 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bc01b720-2ff8-4c06-8b02-ee74e1b8f439-scripts\") pod \"swift-ring-rebalance-debug-gq462\" (UID: \"bc01b720-2ff8-4c06-8b02-ee74e1b8f439\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gq462" Mar 07 15:14:46 crc kubenswrapper[4943]: I0307 15:14:46.359456 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bc01b720-2ff8-4c06-8b02-ee74e1b8f439-ring-data-devices\") pod \"swift-ring-rebalance-debug-gq462\" (UID: \"bc01b720-2ff8-4c06-8b02-ee74e1b8f439\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gq462" Mar 07 15:14:46 crc kubenswrapper[4943]: I0307 15:14:46.461278 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bc01b720-2ff8-4c06-8b02-ee74e1b8f439-swiftconf\") pod \"swift-ring-rebalance-debug-gq462\" (UID: \"bc01b720-2ff8-4c06-8b02-ee74e1b8f439\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gq462" Mar 07 15:14:46 crc kubenswrapper[4943]: I0307 15:14:46.461375 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcxgf\" (UniqueName: \"kubernetes.io/projected/bc01b720-2ff8-4c06-8b02-ee74e1b8f439-kube-api-access-wcxgf\") pod \"swift-ring-rebalance-debug-gq462\" (UID: \"bc01b720-2ff8-4c06-8b02-ee74e1b8f439\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gq462" Mar 07 15:14:46 crc kubenswrapper[4943]: I0307 15:14:46.461420 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bc01b720-2ff8-4c06-8b02-ee74e1b8f439-dispersionconf\") pod \"swift-ring-rebalance-debug-gq462\" (UID: \"bc01b720-2ff8-4c06-8b02-ee74e1b8f439\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gq462" Mar 07 15:14:46 crc kubenswrapper[4943]: I0307 15:14:46.461517 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bc01b720-2ff8-4c06-8b02-ee74e1b8f439-etc-swift\") pod \"swift-ring-rebalance-debug-gq462\" (UID: \"bc01b720-2ff8-4c06-8b02-ee74e1b8f439\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gq462" Mar 07 15:14:46 crc kubenswrapper[4943]: I0307 15:14:46.461550 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bc01b720-2ff8-4c06-8b02-ee74e1b8f439-scripts\") pod \"swift-ring-rebalance-debug-gq462\" (UID: \"bc01b720-2ff8-4c06-8b02-ee74e1b8f439\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gq462" Mar 07 15:14:46 crc kubenswrapper[4943]: I0307 15:14:46.461586 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bc01b720-2ff8-4c06-8b02-ee74e1b8f439-ring-data-devices\") pod \"swift-ring-rebalance-debug-gq462\" (UID: \"bc01b720-2ff8-4c06-8b02-ee74e1b8f439\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gq462" Mar 07 15:14:46 crc kubenswrapper[4943]: I0307 15:14:46.462359 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bc01b720-2ff8-4c06-8b02-ee74e1b8f439-etc-swift\") pod \"swift-ring-rebalance-debug-gq462\" (UID: \"bc01b720-2ff8-4c06-8b02-ee74e1b8f439\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gq462" Mar 07 15:14:46 crc kubenswrapper[4943]: I0307 15:14:46.463115 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bc01b720-2ff8-4c06-8b02-ee74e1b8f439-scripts\") pod \"swift-ring-rebalance-debug-gq462\" (UID: \"bc01b720-2ff8-4c06-8b02-ee74e1b8f439\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gq462" Mar 07 15:14:46 crc kubenswrapper[4943]: I0307 15:14:46.463194 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bc01b720-2ff8-4c06-8b02-ee74e1b8f439-ring-data-devices\") pod \"swift-ring-rebalance-debug-gq462\" (UID: \"bc01b720-2ff8-4c06-8b02-ee74e1b8f439\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gq462" Mar 07 15:14:46 crc kubenswrapper[4943]: I0307 15:14:46.469003 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bc01b720-2ff8-4c06-8b02-ee74e1b8f439-swiftconf\") pod \"swift-ring-rebalance-debug-gq462\" (UID: \"bc01b720-2ff8-4c06-8b02-ee74e1b8f439\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gq462" Mar 07 15:14:46 crc kubenswrapper[4943]: I0307 15:14:46.482972 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bc01b720-2ff8-4c06-8b02-ee74e1b8f439-dispersionconf\") pod \"swift-ring-rebalance-debug-gq462\" (UID: \"bc01b720-2ff8-4c06-8b02-ee74e1b8f439\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gq462" Mar 07 15:14:46 crc kubenswrapper[4943]: I0307 15:14:46.483124 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcxgf\" (UniqueName: \"kubernetes.io/projected/bc01b720-2ff8-4c06-8b02-ee74e1b8f439-kube-api-access-wcxgf\") pod \"swift-ring-rebalance-debug-gq462\" (UID: \"bc01b720-2ff8-4c06-8b02-ee74e1b8f439\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gq462" Mar 07 15:14:46 crc kubenswrapper[4943]: I0307 15:14:46.518081 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-gq462" Mar 07 15:14:46 crc kubenswrapper[4943]: I0307 15:14:46.770893 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71db4d2a-006a-478a-8445-3ace5f5082ab" path="/var/lib/kubelet/pods/71db4d2a-006a-478a-8445-3ace5f5082ab/volumes" Mar 07 15:14:46 crc kubenswrapper[4943]: I0307 15:14:46.773698 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-gq462"] Mar 07 15:14:46 crc kubenswrapper[4943]: W0307 15:14:46.780186 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbc01b720_2ff8_4c06_8b02_ee74e1b8f439.slice/crio-fdd9014553fccf28f273635f8a4f8ac56794355f3d4d461f93032409332cbd92 WatchSource:0}: Error finding container fdd9014553fccf28f273635f8a4f8ac56794355f3d4d461f93032409332cbd92: Status 404 returned error can't find the container with id fdd9014553fccf28f273635f8a4f8ac56794355f3d4d461f93032409332cbd92 Mar 07 15:14:47 crc kubenswrapper[4943]: I0307 15:14:47.699360 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-gq462" event={"ID":"bc01b720-2ff8-4c06-8b02-ee74e1b8f439","Type":"ContainerStarted","Data":"75b39e0895817d203e2ba4b4ad29a861fd66d7b554f98549272c08813e57cc6f"} Mar 07 15:14:47 crc kubenswrapper[4943]: I0307 15:14:47.699988 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-gq462" event={"ID":"bc01b720-2ff8-4c06-8b02-ee74e1b8f439","Type":"ContainerStarted","Data":"fdd9014553fccf28f273635f8a4f8ac56794355f3d4d461f93032409332cbd92"} Mar 07 15:14:48 crc kubenswrapper[4943]: I0307 15:14:48.712501 4943 generic.go:334] "Generic (PLEG): container finished" podID="bc01b720-2ff8-4c06-8b02-ee74e1b8f439" containerID="75b39e0895817d203e2ba4b4ad29a861fd66d7b554f98549272c08813e57cc6f" exitCode=0 Mar 07 15:14:48 crc kubenswrapper[4943]: I0307 15:14:48.712550 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-gq462" event={"ID":"bc01b720-2ff8-4c06-8b02-ee74e1b8f439","Type":"ContainerDied","Data":"75b39e0895817d203e2ba4b4ad29a861fd66d7b554f98549272c08813e57cc6f"} Mar 07 15:14:50 crc kubenswrapper[4943]: I0307 15:14:50.089717 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-gq462" Mar 07 15:14:50 crc kubenswrapper[4943]: I0307 15:14:50.138431 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-gq462"] Mar 07 15:14:50 crc kubenswrapper[4943]: I0307 15:14:50.144966 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-gq462"] Mar 07 15:14:50 crc kubenswrapper[4943]: I0307 15:14:50.234409 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wcxgf\" (UniqueName: \"kubernetes.io/projected/bc01b720-2ff8-4c06-8b02-ee74e1b8f439-kube-api-access-wcxgf\") pod \"bc01b720-2ff8-4c06-8b02-ee74e1b8f439\" (UID: \"bc01b720-2ff8-4c06-8b02-ee74e1b8f439\") " Mar 07 15:14:50 crc kubenswrapper[4943]: I0307 15:14:50.234451 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bc01b720-2ff8-4c06-8b02-ee74e1b8f439-swiftconf\") pod \"bc01b720-2ff8-4c06-8b02-ee74e1b8f439\" (UID: \"bc01b720-2ff8-4c06-8b02-ee74e1b8f439\") " Mar 07 15:14:50 crc kubenswrapper[4943]: I0307 15:14:50.234500 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bc01b720-2ff8-4c06-8b02-ee74e1b8f439-ring-data-devices\") pod \"bc01b720-2ff8-4c06-8b02-ee74e1b8f439\" (UID: \"bc01b720-2ff8-4c06-8b02-ee74e1b8f439\") " Mar 07 15:14:50 crc kubenswrapper[4943]: I0307 15:14:50.234567 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bc01b720-2ff8-4c06-8b02-ee74e1b8f439-scripts\") pod \"bc01b720-2ff8-4c06-8b02-ee74e1b8f439\" (UID: \"bc01b720-2ff8-4c06-8b02-ee74e1b8f439\") " Mar 07 15:14:50 crc kubenswrapper[4943]: I0307 15:14:50.234587 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bc01b720-2ff8-4c06-8b02-ee74e1b8f439-etc-swift\") pod \"bc01b720-2ff8-4c06-8b02-ee74e1b8f439\" (UID: \"bc01b720-2ff8-4c06-8b02-ee74e1b8f439\") " Mar 07 15:14:50 crc kubenswrapper[4943]: I0307 15:14:50.234667 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bc01b720-2ff8-4c06-8b02-ee74e1b8f439-dispersionconf\") pod \"bc01b720-2ff8-4c06-8b02-ee74e1b8f439\" (UID: \"bc01b720-2ff8-4c06-8b02-ee74e1b8f439\") " Mar 07 15:14:50 crc kubenswrapper[4943]: I0307 15:14:50.235353 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc01b720-2ff8-4c06-8b02-ee74e1b8f439-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "bc01b720-2ff8-4c06-8b02-ee74e1b8f439" (UID: "bc01b720-2ff8-4c06-8b02-ee74e1b8f439"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:14:50 crc kubenswrapper[4943]: I0307 15:14:50.235609 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc01b720-2ff8-4c06-8b02-ee74e1b8f439-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "bc01b720-2ff8-4c06-8b02-ee74e1b8f439" (UID: "bc01b720-2ff8-4c06-8b02-ee74e1b8f439"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:14:50 crc kubenswrapper[4943]: I0307 15:14:50.240649 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc01b720-2ff8-4c06-8b02-ee74e1b8f439-kube-api-access-wcxgf" (OuterVolumeSpecName: "kube-api-access-wcxgf") pod "bc01b720-2ff8-4c06-8b02-ee74e1b8f439" (UID: "bc01b720-2ff8-4c06-8b02-ee74e1b8f439"). InnerVolumeSpecName "kube-api-access-wcxgf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:14:50 crc kubenswrapper[4943]: I0307 15:14:50.255496 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc01b720-2ff8-4c06-8b02-ee74e1b8f439-scripts" (OuterVolumeSpecName: "scripts") pod "bc01b720-2ff8-4c06-8b02-ee74e1b8f439" (UID: "bc01b720-2ff8-4c06-8b02-ee74e1b8f439"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:14:50 crc kubenswrapper[4943]: I0307 15:14:50.259109 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc01b720-2ff8-4c06-8b02-ee74e1b8f439-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "bc01b720-2ff8-4c06-8b02-ee74e1b8f439" (UID: "bc01b720-2ff8-4c06-8b02-ee74e1b8f439"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:14:50 crc kubenswrapper[4943]: I0307 15:14:50.259329 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc01b720-2ff8-4c06-8b02-ee74e1b8f439-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "bc01b720-2ff8-4c06-8b02-ee74e1b8f439" (UID: "bc01b720-2ff8-4c06-8b02-ee74e1b8f439"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:14:50 crc kubenswrapper[4943]: I0307 15:14:50.336623 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bc01b720-2ff8-4c06-8b02-ee74e1b8f439-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:50 crc kubenswrapper[4943]: I0307 15:14:50.336648 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wcxgf\" (UniqueName: \"kubernetes.io/projected/bc01b720-2ff8-4c06-8b02-ee74e1b8f439-kube-api-access-wcxgf\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:50 crc kubenswrapper[4943]: I0307 15:14:50.336660 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bc01b720-2ff8-4c06-8b02-ee74e1b8f439-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:50 crc kubenswrapper[4943]: I0307 15:14:50.336669 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bc01b720-2ff8-4c06-8b02-ee74e1b8f439-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:50 crc kubenswrapper[4943]: I0307 15:14:50.336679 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bc01b720-2ff8-4c06-8b02-ee74e1b8f439-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:50 crc kubenswrapper[4943]: I0307 15:14:50.336687 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bc01b720-2ff8-4c06-8b02-ee74e1b8f439-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:50 crc kubenswrapper[4943]: I0307 15:14:50.733098 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fdd9014553fccf28f273635f8a4f8ac56794355f3d4d461f93032409332cbd92" Mar 07 15:14:50 crc kubenswrapper[4943]: I0307 15:14:50.733158 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-gq462" Mar 07 15:14:50 crc kubenswrapper[4943]: I0307 15:14:50.769143 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc01b720-2ff8-4c06-8b02-ee74e1b8f439" path="/var/lib/kubelet/pods/bc01b720-2ff8-4c06-8b02-ee74e1b8f439/volumes" Mar 07 15:14:51 crc kubenswrapper[4943]: I0307 15:14:51.285568 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-z9dr4"] Mar 07 15:14:51 crc kubenswrapper[4943]: E0307 15:14:51.286332 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc01b720-2ff8-4c06-8b02-ee74e1b8f439" containerName="swift-ring-rebalance" Mar 07 15:14:51 crc kubenswrapper[4943]: I0307 15:14:51.286354 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc01b720-2ff8-4c06-8b02-ee74e1b8f439" containerName="swift-ring-rebalance" Mar 07 15:14:51 crc kubenswrapper[4943]: I0307 15:14:51.286568 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc01b720-2ff8-4c06-8b02-ee74e1b8f439" containerName="swift-ring-rebalance" Mar 07 15:14:51 crc kubenswrapper[4943]: I0307 15:14:51.287253 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-z9dr4" Mar 07 15:14:51 crc kubenswrapper[4943]: I0307 15:14:51.289516 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:14:51 crc kubenswrapper[4943]: I0307 15:14:51.290547 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:14:51 crc kubenswrapper[4943]: I0307 15:14:51.296178 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-z9dr4"] Mar 07 15:14:51 crc kubenswrapper[4943]: I0307 15:14:51.352013 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hnzh\" (UniqueName: \"kubernetes.io/projected/3dd1b036-d74a-4f01-82c4-748c0a37fad4-kube-api-access-4hnzh\") pod \"swift-ring-rebalance-debug-z9dr4\" (UID: \"3dd1b036-d74a-4f01-82c4-748c0a37fad4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-z9dr4" Mar 07 15:14:51 crc kubenswrapper[4943]: I0307 15:14:51.352077 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3dd1b036-d74a-4f01-82c4-748c0a37fad4-scripts\") pod \"swift-ring-rebalance-debug-z9dr4\" (UID: \"3dd1b036-d74a-4f01-82c4-748c0a37fad4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-z9dr4" Mar 07 15:14:51 crc kubenswrapper[4943]: I0307 15:14:51.352108 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3dd1b036-d74a-4f01-82c4-748c0a37fad4-etc-swift\") pod \"swift-ring-rebalance-debug-z9dr4\" (UID: \"3dd1b036-d74a-4f01-82c4-748c0a37fad4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-z9dr4" Mar 07 15:14:51 crc kubenswrapper[4943]: I0307 15:14:51.352130 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3dd1b036-d74a-4f01-82c4-748c0a37fad4-ring-data-devices\") pod \"swift-ring-rebalance-debug-z9dr4\" (UID: \"3dd1b036-d74a-4f01-82c4-748c0a37fad4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-z9dr4" Mar 07 15:14:51 crc kubenswrapper[4943]: I0307 15:14:51.352452 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3dd1b036-d74a-4f01-82c4-748c0a37fad4-dispersionconf\") pod \"swift-ring-rebalance-debug-z9dr4\" (UID: \"3dd1b036-d74a-4f01-82c4-748c0a37fad4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-z9dr4" Mar 07 15:14:51 crc kubenswrapper[4943]: I0307 15:14:51.352528 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3dd1b036-d74a-4f01-82c4-748c0a37fad4-swiftconf\") pod \"swift-ring-rebalance-debug-z9dr4\" (UID: \"3dd1b036-d74a-4f01-82c4-748c0a37fad4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-z9dr4" Mar 07 15:14:51 crc kubenswrapper[4943]: I0307 15:14:51.453827 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3dd1b036-d74a-4f01-82c4-748c0a37fad4-etc-swift\") pod \"swift-ring-rebalance-debug-z9dr4\" (UID: \"3dd1b036-d74a-4f01-82c4-748c0a37fad4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-z9dr4" Mar 07 15:14:51 crc kubenswrapper[4943]: I0307 15:14:51.453877 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3dd1b036-d74a-4f01-82c4-748c0a37fad4-ring-data-devices\") pod \"swift-ring-rebalance-debug-z9dr4\" (UID: \"3dd1b036-d74a-4f01-82c4-748c0a37fad4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-z9dr4" Mar 07 15:14:51 crc kubenswrapper[4943]: I0307 15:14:51.453992 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3dd1b036-d74a-4f01-82c4-748c0a37fad4-dispersionconf\") pod \"swift-ring-rebalance-debug-z9dr4\" (UID: \"3dd1b036-d74a-4f01-82c4-748c0a37fad4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-z9dr4" Mar 07 15:14:51 crc kubenswrapper[4943]: I0307 15:14:51.454019 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3dd1b036-d74a-4f01-82c4-748c0a37fad4-swiftconf\") pod \"swift-ring-rebalance-debug-z9dr4\" (UID: \"3dd1b036-d74a-4f01-82c4-748c0a37fad4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-z9dr4" Mar 07 15:14:51 crc kubenswrapper[4943]: I0307 15:14:51.454074 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hnzh\" (UniqueName: \"kubernetes.io/projected/3dd1b036-d74a-4f01-82c4-748c0a37fad4-kube-api-access-4hnzh\") pod \"swift-ring-rebalance-debug-z9dr4\" (UID: \"3dd1b036-d74a-4f01-82c4-748c0a37fad4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-z9dr4" Mar 07 15:14:51 crc kubenswrapper[4943]: I0307 15:14:51.454102 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3dd1b036-d74a-4f01-82c4-748c0a37fad4-scripts\") pod \"swift-ring-rebalance-debug-z9dr4\" (UID: \"3dd1b036-d74a-4f01-82c4-748c0a37fad4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-z9dr4" Mar 07 15:14:51 crc kubenswrapper[4943]: I0307 15:14:51.455093 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3dd1b036-d74a-4f01-82c4-748c0a37fad4-etc-swift\") pod \"swift-ring-rebalance-debug-z9dr4\" (UID: \"3dd1b036-d74a-4f01-82c4-748c0a37fad4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-z9dr4" Mar 07 15:14:51 crc kubenswrapper[4943]: I0307 15:14:51.455352 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3dd1b036-d74a-4f01-82c4-748c0a37fad4-ring-data-devices\") pod \"swift-ring-rebalance-debug-z9dr4\" (UID: \"3dd1b036-d74a-4f01-82c4-748c0a37fad4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-z9dr4" Mar 07 15:14:51 crc kubenswrapper[4943]: I0307 15:14:51.455517 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3dd1b036-d74a-4f01-82c4-748c0a37fad4-scripts\") pod \"swift-ring-rebalance-debug-z9dr4\" (UID: \"3dd1b036-d74a-4f01-82c4-748c0a37fad4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-z9dr4" Mar 07 15:14:51 crc kubenswrapper[4943]: I0307 15:14:51.458507 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3dd1b036-d74a-4f01-82c4-748c0a37fad4-swiftconf\") pod \"swift-ring-rebalance-debug-z9dr4\" (UID: \"3dd1b036-d74a-4f01-82c4-748c0a37fad4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-z9dr4" Mar 07 15:14:51 crc kubenswrapper[4943]: I0307 15:14:51.459643 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3dd1b036-d74a-4f01-82c4-748c0a37fad4-dispersionconf\") pod \"swift-ring-rebalance-debug-z9dr4\" (UID: \"3dd1b036-d74a-4f01-82c4-748c0a37fad4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-z9dr4" Mar 07 15:14:51 crc kubenswrapper[4943]: I0307 15:14:51.484615 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hnzh\" (UniqueName: \"kubernetes.io/projected/3dd1b036-d74a-4f01-82c4-748c0a37fad4-kube-api-access-4hnzh\") pod \"swift-ring-rebalance-debug-z9dr4\" (UID: \"3dd1b036-d74a-4f01-82c4-748c0a37fad4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-z9dr4" Mar 07 15:14:51 crc kubenswrapper[4943]: I0307 15:14:51.617277 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-z9dr4" Mar 07 15:14:51 crc kubenswrapper[4943]: I0307 15:14:51.895150 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-z9dr4"] Mar 07 15:14:52 crc kubenswrapper[4943]: I0307 15:14:52.776724 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-z9dr4" event={"ID":"3dd1b036-d74a-4f01-82c4-748c0a37fad4","Type":"ContainerStarted","Data":"e5bc4c5fb3c660735c7885799282c09732ec1f0a5848df8f03dbe19751851494"} Mar 07 15:14:52 crc kubenswrapper[4943]: I0307 15:14:52.778263 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-z9dr4" event={"ID":"3dd1b036-d74a-4f01-82c4-748c0a37fad4","Type":"ContainerStarted","Data":"3149ca8e1b717e5203e0808961e0817ff4deb608154f72fc83be0c0e53617475"} Mar 07 15:14:52 crc kubenswrapper[4943]: I0307 15:14:52.814835 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-z9dr4" podStartSLOduration=1.8148137370000001 podStartE2EDuration="1.814813737s" podCreationTimestamp="2026-03-07 15:14:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:14:52.805380585 +0000 UTC m=+2134.757517173" watchObservedRunningTime="2026-03-07 15:14:52.814813737 +0000 UTC m=+2134.766950245" Mar 07 15:14:53 crc kubenswrapper[4943]: I0307 15:14:53.801038 4943 generic.go:334] "Generic (PLEG): container finished" podID="3dd1b036-d74a-4f01-82c4-748c0a37fad4" containerID="e5bc4c5fb3c660735c7885799282c09732ec1f0a5848df8f03dbe19751851494" exitCode=0 Mar 07 15:14:53 crc kubenswrapper[4943]: I0307 15:14:53.801104 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-z9dr4" event={"ID":"3dd1b036-d74a-4f01-82c4-748c0a37fad4","Type":"ContainerDied","Data":"e5bc4c5fb3c660735c7885799282c09732ec1f0a5848df8f03dbe19751851494"} Mar 07 15:14:55 crc kubenswrapper[4943]: I0307 15:14:55.224288 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-z9dr4" Mar 07 15:14:55 crc kubenswrapper[4943]: I0307 15:14:55.250809 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-z9dr4"] Mar 07 15:14:55 crc kubenswrapper[4943]: I0307 15:14:55.255134 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-z9dr4"] Mar 07 15:14:55 crc kubenswrapper[4943]: I0307 15:14:55.315202 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hnzh\" (UniqueName: \"kubernetes.io/projected/3dd1b036-d74a-4f01-82c4-748c0a37fad4-kube-api-access-4hnzh\") pod \"3dd1b036-d74a-4f01-82c4-748c0a37fad4\" (UID: \"3dd1b036-d74a-4f01-82c4-748c0a37fad4\") " Mar 07 15:14:55 crc kubenswrapper[4943]: I0307 15:14:55.315319 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3dd1b036-d74a-4f01-82c4-748c0a37fad4-ring-data-devices\") pod \"3dd1b036-d74a-4f01-82c4-748c0a37fad4\" (UID: \"3dd1b036-d74a-4f01-82c4-748c0a37fad4\") " Mar 07 15:14:55 crc kubenswrapper[4943]: I0307 15:14:55.315357 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3dd1b036-d74a-4f01-82c4-748c0a37fad4-swiftconf\") pod \"3dd1b036-d74a-4f01-82c4-748c0a37fad4\" (UID: \"3dd1b036-d74a-4f01-82c4-748c0a37fad4\") " Mar 07 15:14:55 crc kubenswrapper[4943]: I0307 15:14:55.315388 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3dd1b036-d74a-4f01-82c4-748c0a37fad4-dispersionconf\") pod \"3dd1b036-d74a-4f01-82c4-748c0a37fad4\" (UID: \"3dd1b036-d74a-4f01-82c4-748c0a37fad4\") " Mar 07 15:14:55 crc kubenswrapper[4943]: I0307 15:14:55.315423 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3dd1b036-d74a-4f01-82c4-748c0a37fad4-etc-swift\") pod \"3dd1b036-d74a-4f01-82c4-748c0a37fad4\" (UID: \"3dd1b036-d74a-4f01-82c4-748c0a37fad4\") " Mar 07 15:14:55 crc kubenswrapper[4943]: I0307 15:14:55.315504 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3dd1b036-d74a-4f01-82c4-748c0a37fad4-scripts\") pod \"3dd1b036-d74a-4f01-82c4-748c0a37fad4\" (UID: \"3dd1b036-d74a-4f01-82c4-748c0a37fad4\") " Mar 07 15:14:55 crc kubenswrapper[4943]: I0307 15:14:55.317013 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3dd1b036-d74a-4f01-82c4-748c0a37fad4-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "3dd1b036-d74a-4f01-82c4-748c0a37fad4" (UID: "3dd1b036-d74a-4f01-82c4-748c0a37fad4"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:14:55 crc kubenswrapper[4943]: I0307 15:14:55.317137 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3dd1b036-d74a-4f01-82c4-748c0a37fad4-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "3dd1b036-d74a-4f01-82c4-748c0a37fad4" (UID: "3dd1b036-d74a-4f01-82c4-748c0a37fad4"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:14:55 crc kubenswrapper[4943]: I0307 15:14:55.320477 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3dd1b036-d74a-4f01-82c4-748c0a37fad4-kube-api-access-4hnzh" (OuterVolumeSpecName: "kube-api-access-4hnzh") pod "3dd1b036-d74a-4f01-82c4-748c0a37fad4" (UID: "3dd1b036-d74a-4f01-82c4-748c0a37fad4"). InnerVolumeSpecName "kube-api-access-4hnzh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:14:55 crc kubenswrapper[4943]: I0307 15:14:55.337181 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3dd1b036-d74a-4f01-82c4-748c0a37fad4-scripts" (OuterVolumeSpecName: "scripts") pod "3dd1b036-d74a-4f01-82c4-748c0a37fad4" (UID: "3dd1b036-d74a-4f01-82c4-748c0a37fad4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:14:55 crc kubenswrapper[4943]: I0307 15:14:55.337422 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dd1b036-d74a-4f01-82c4-748c0a37fad4-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "3dd1b036-d74a-4f01-82c4-748c0a37fad4" (UID: "3dd1b036-d74a-4f01-82c4-748c0a37fad4"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:14:55 crc kubenswrapper[4943]: I0307 15:14:55.350194 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dd1b036-d74a-4f01-82c4-748c0a37fad4-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "3dd1b036-d74a-4f01-82c4-748c0a37fad4" (UID: "3dd1b036-d74a-4f01-82c4-748c0a37fad4"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:14:55 crc kubenswrapper[4943]: I0307 15:14:55.417481 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hnzh\" (UniqueName: \"kubernetes.io/projected/3dd1b036-d74a-4f01-82c4-748c0a37fad4-kube-api-access-4hnzh\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:55 crc kubenswrapper[4943]: I0307 15:14:55.417673 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3dd1b036-d74a-4f01-82c4-748c0a37fad4-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:55 crc kubenswrapper[4943]: I0307 15:14:55.417754 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3dd1b036-d74a-4f01-82c4-748c0a37fad4-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:55 crc kubenswrapper[4943]: I0307 15:14:55.417832 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3dd1b036-d74a-4f01-82c4-748c0a37fad4-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:55 crc kubenswrapper[4943]: I0307 15:14:55.417914 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3dd1b036-d74a-4f01-82c4-748c0a37fad4-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:55 crc kubenswrapper[4943]: I0307 15:14:55.418036 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3dd1b036-d74a-4f01-82c4-748c0a37fad4-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:14:55 crc kubenswrapper[4943]: I0307 15:14:55.822196 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3149ca8e1b717e5203e0808961e0817ff4deb608154f72fc83be0c0e53617475" Mar 07 15:14:55 crc kubenswrapper[4943]: I0307 15:14:55.822303 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-z9dr4" Mar 07 15:14:56 crc kubenswrapper[4943]: I0307 15:14:56.455184 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-bzsv2"] Mar 07 15:14:56 crc kubenswrapper[4943]: E0307 15:14:56.455720 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dd1b036-d74a-4f01-82c4-748c0a37fad4" containerName="swift-ring-rebalance" Mar 07 15:14:56 crc kubenswrapper[4943]: I0307 15:14:56.455745 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dd1b036-d74a-4f01-82c4-748c0a37fad4" containerName="swift-ring-rebalance" Mar 07 15:14:56 crc kubenswrapper[4943]: I0307 15:14:56.456038 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="3dd1b036-d74a-4f01-82c4-748c0a37fad4" containerName="swift-ring-rebalance" Mar 07 15:14:56 crc kubenswrapper[4943]: I0307 15:14:56.456971 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-bzsv2" Mar 07 15:14:56 crc kubenswrapper[4943]: I0307 15:14:56.459760 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:14:56 crc kubenswrapper[4943]: I0307 15:14:56.469631 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:14:56 crc kubenswrapper[4943]: I0307 15:14:56.473781 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-bzsv2"] Mar 07 15:14:56 crc kubenswrapper[4943]: I0307 15:14:56.534616 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b98a1692-83da-439e-8612-7a165d35171a-swiftconf\") pod \"swift-ring-rebalance-debug-bzsv2\" (UID: \"b98a1692-83da-439e-8612-7a165d35171a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bzsv2" Mar 07 15:14:56 crc kubenswrapper[4943]: I0307 15:14:56.534656 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b98a1692-83da-439e-8612-7a165d35171a-dispersionconf\") pod \"swift-ring-rebalance-debug-bzsv2\" (UID: \"b98a1692-83da-439e-8612-7a165d35171a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bzsv2" Mar 07 15:14:56 crc kubenswrapper[4943]: I0307 15:14:56.535213 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b98a1692-83da-439e-8612-7a165d35171a-ring-data-devices\") pod \"swift-ring-rebalance-debug-bzsv2\" (UID: \"b98a1692-83da-439e-8612-7a165d35171a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bzsv2" Mar 07 15:14:56 crc kubenswrapper[4943]: I0307 15:14:56.535317 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b98a1692-83da-439e-8612-7a165d35171a-etc-swift\") pod \"swift-ring-rebalance-debug-bzsv2\" (UID: \"b98a1692-83da-439e-8612-7a165d35171a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bzsv2" Mar 07 15:14:56 crc kubenswrapper[4943]: I0307 15:14:56.535430 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b98a1692-83da-439e-8612-7a165d35171a-scripts\") pod \"swift-ring-rebalance-debug-bzsv2\" (UID: \"b98a1692-83da-439e-8612-7a165d35171a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bzsv2" Mar 07 15:14:56 crc kubenswrapper[4943]: I0307 15:14:56.535486 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqxp5\" (UniqueName: \"kubernetes.io/projected/b98a1692-83da-439e-8612-7a165d35171a-kube-api-access-jqxp5\") pod \"swift-ring-rebalance-debug-bzsv2\" (UID: \"b98a1692-83da-439e-8612-7a165d35171a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bzsv2" Mar 07 15:14:56 crc kubenswrapper[4943]: I0307 15:14:56.637425 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b98a1692-83da-439e-8612-7a165d35171a-etc-swift\") pod \"swift-ring-rebalance-debug-bzsv2\" (UID: \"b98a1692-83da-439e-8612-7a165d35171a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bzsv2" Mar 07 15:14:56 crc kubenswrapper[4943]: I0307 15:14:56.637511 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b98a1692-83da-439e-8612-7a165d35171a-scripts\") pod \"swift-ring-rebalance-debug-bzsv2\" (UID: \"b98a1692-83da-439e-8612-7a165d35171a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bzsv2" Mar 07 15:14:56 crc kubenswrapper[4943]: I0307 15:14:56.637545 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqxp5\" (UniqueName: \"kubernetes.io/projected/b98a1692-83da-439e-8612-7a165d35171a-kube-api-access-jqxp5\") pod \"swift-ring-rebalance-debug-bzsv2\" (UID: \"b98a1692-83da-439e-8612-7a165d35171a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bzsv2" Mar 07 15:14:56 crc kubenswrapper[4943]: I0307 15:14:56.637634 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b98a1692-83da-439e-8612-7a165d35171a-swiftconf\") pod \"swift-ring-rebalance-debug-bzsv2\" (UID: \"b98a1692-83da-439e-8612-7a165d35171a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bzsv2" Mar 07 15:14:56 crc kubenswrapper[4943]: I0307 15:14:56.637669 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b98a1692-83da-439e-8612-7a165d35171a-dispersionconf\") pod \"swift-ring-rebalance-debug-bzsv2\" (UID: \"b98a1692-83da-439e-8612-7a165d35171a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bzsv2" Mar 07 15:14:56 crc kubenswrapper[4943]: I0307 15:14:56.637790 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b98a1692-83da-439e-8612-7a165d35171a-ring-data-devices\") pod \"swift-ring-rebalance-debug-bzsv2\" (UID: \"b98a1692-83da-439e-8612-7a165d35171a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bzsv2" Mar 07 15:14:56 crc kubenswrapper[4943]: I0307 15:14:56.638319 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b98a1692-83da-439e-8612-7a165d35171a-etc-swift\") pod \"swift-ring-rebalance-debug-bzsv2\" (UID: \"b98a1692-83da-439e-8612-7a165d35171a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bzsv2" Mar 07 15:14:56 crc kubenswrapper[4943]: I0307 15:14:56.638383 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b98a1692-83da-439e-8612-7a165d35171a-scripts\") pod \"swift-ring-rebalance-debug-bzsv2\" (UID: \"b98a1692-83da-439e-8612-7a165d35171a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bzsv2" Mar 07 15:14:56 crc kubenswrapper[4943]: I0307 15:14:56.638775 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b98a1692-83da-439e-8612-7a165d35171a-ring-data-devices\") pod \"swift-ring-rebalance-debug-bzsv2\" (UID: \"b98a1692-83da-439e-8612-7a165d35171a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bzsv2" Mar 07 15:14:56 crc kubenswrapper[4943]: I0307 15:14:56.643397 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b98a1692-83da-439e-8612-7a165d35171a-dispersionconf\") pod \"swift-ring-rebalance-debug-bzsv2\" (UID: \"b98a1692-83da-439e-8612-7a165d35171a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bzsv2" Mar 07 15:14:56 crc kubenswrapper[4943]: I0307 15:14:56.643529 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b98a1692-83da-439e-8612-7a165d35171a-swiftconf\") pod \"swift-ring-rebalance-debug-bzsv2\" (UID: \"b98a1692-83da-439e-8612-7a165d35171a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bzsv2" Mar 07 15:14:56 crc kubenswrapper[4943]: I0307 15:14:56.655848 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqxp5\" (UniqueName: \"kubernetes.io/projected/b98a1692-83da-439e-8612-7a165d35171a-kube-api-access-jqxp5\") pod \"swift-ring-rebalance-debug-bzsv2\" (UID: \"b98a1692-83da-439e-8612-7a165d35171a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bzsv2" Mar 07 15:14:56 crc kubenswrapper[4943]: I0307 15:14:56.766316 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3dd1b036-d74a-4f01-82c4-748c0a37fad4" path="/var/lib/kubelet/pods/3dd1b036-d74a-4f01-82c4-748c0a37fad4/volumes" Mar 07 15:14:56 crc kubenswrapper[4943]: I0307 15:14:56.790904 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-bzsv2" Mar 07 15:14:57 crc kubenswrapper[4943]: I0307 15:14:57.272881 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-bzsv2"] Mar 07 15:14:57 crc kubenswrapper[4943]: I0307 15:14:57.852962 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-bzsv2" event={"ID":"b98a1692-83da-439e-8612-7a165d35171a","Type":"ContainerStarted","Data":"2ca4b462f42a1f213c6729ebe4a16c33d10eda9ae5e2e5aec8a4c07656096157"} Mar 07 15:14:57 crc kubenswrapper[4943]: I0307 15:14:57.853426 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-bzsv2" event={"ID":"b98a1692-83da-439e-8612-7a165d35171a","Type":"ContainerStarted","Data":"950bff00f2db00059ce0a30deaf68e54f299fa97c744ea290bc5541b9ff8b85b"} Mar 07 15:14:58 crc kubenswrapper[4943]: I0307 15:14:58.864873 4943 generic.go:334] "Generic (PLEG): container finished" podID="b98a1692-83da-439e-8612-7a165d35171a" containerID="2ca4b462f42a1f213c6729ebe4a16c33d10eda9ae5e2e5aec8a4c07656096157" exitCode=0 Mar 07 15:14:58 crc kubenswrapper[4943]: I0307 15:14:58.864917 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-bzsv2" event={"ID":"b98a1692-83da-439e-8612-7a165d35171a","Type":"ContainerDied","Data":"2ca4b462f42a1f213c6729ebe4a16c33d10eda9ae5e2e5aec8a4c07656096157"} Mar 07 15:15:00 crc kubenswrapper[4943]: I0307 15:15:00.137439 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29548275-tqwsv"] Mar 07 15:15:00 crc kubenswrapper[4943]: I0307 15:15:00.140000 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29548275-tqwsv" Mar 07 15:15:00 crc kubenswrapper[4943]: I0307 15:15:00.142377 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 07 15:15:00 crc kubenswrapper[4943]: I0307 15:15:00.143329 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 07 15:15:00 crc kubenswrapper[4943]: I0307 15:15:00.147815 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29548275-tqwsv"] Mar 07 15:15:00 crc kubenswrapper[4943]: I0307 15:15:00.191569 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cedf5080-bb07-4ceb-9d64-fc45d9a25f74-secret-volume\") pod \"collect-profiles-29548275-tqwsv\" (UID: \"cedf5080-bb07-4ceb-9d64-fc45d9a25f74\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29548275-tqwsv" Mar 07 15:15:00 crc kubenswrapper[4943]: I0307 15:15:00.191644 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cedf5080-bb07-4ceb-9d64-fc45d9a25f74-config-volume\") pod \"collect-profiles-29548275-tqwsv\" (UID: \"cedf5080-bb07-4ceb-9d64-fc45d9a25f74\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29548275-tqwsv" Mar 07 15:15:00 crc kubenswrapper[4943]: I0307 15:15:00.191713 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxncq\" (UniqueName: \"kubernetes.io/projected/cedf5080-bb07-4ceb-9d64-fc45d9a25f74-kube-api-access-fxncq\") pod \"collect-profiles-29548275-tqwsv\" (UID: \"cedf5080-bb07-4ceb-9d64-fc45d9a25f74\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29548275-tqwsv" Mar 07 15:15:00 crc kubenswrapper[4943]: I0307 15:15:00.231415 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-bzsv2" Mar 07 15:15:00 crc kubenswrapper[4943]: I0307 15:15:00.260453 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-bzsv2"] Mar 07 15:15:00 crc kubenswrapper[4943]: I0307 15:15:00.264886 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-bzsv2"] Mar 07 15:15:00 crc kubenswrapper[4943]: I0307 15:15:00.293039 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cedf5080-bb07-4ceb-9d64-fc45d9a25f74-config-volume\") pod \"collect-profiles-29548275-tqwsv\" (UID: \"cedf5080-bb07-4ceb-9d64-fc45d9a25f74\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29548275-tqwsv" Mar 07 15:15:00 crc kubenswrapper[4943]: I0307 15:15:00.293152 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxncq\" (UniqueName: \"kubernetes.io/projected/cedf5080-bb07-4ceb-9d64-fc45d9a25f74-kube-api-access-fxncq\") pod \"collect-profiles-29548275-tqwsv\" (UID: \"cedf5080-bb07-4ceb-9d64-fc45d9a25f74\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29548275-tqwsv" Mar 07 15:15:00 crc kubenswrapper[4943]: I0307 15:15:00.293216 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cedf5080-bb07-4ceb-9d64-fc45d9a25f74-secret-volume\") pod \"collect-profiles-29548275-tqwsv\" (UID: \"cedf5080-bb07-4ceb-9d64-fc45d9a25f74\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29548275-tqwsv" Mar 07 15:15:00 crc kubenswrapper[4943]: I0307 15:15:00.294524 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cedf5080-bb07-4ceb-9d64-fc45d9a25f74-config-volume\") pod \"collect-profiles-29548275-tqwsv\" (UID: \"cedf5080-bb07-4ceb-9d64-fc45d9a25f74\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29548275-tqwsv" Mar 07 15:15:00 crc kubenswrapper[4943]: I0307 15:15:00.301790 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cedf5080-bb07-4ceb-9d64-fc45d9a25f74-secret-volume\") pod \"collect-profiles-29548275-tqwsv\" (UID: \"cedf5080-bb07-4ceb-9d64-fc45d9a25f74\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29548275-tqwsv" Mar 07 15:15:00 crc kubenswrapper[4943]: I0307 15:15:00.312364 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxncq\" (UniqueName: \"kubernetes.io/projected/cedf5080-bb07-4ceb-9d64-fc45d9a25f74-kube-api-access-fxncq\") pod \"collect-profiles-29548275-tqwsv\" (UID: \"cedf5080-bb07-4ceb-9d64-fc45d9a25f74\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29548275-tqwsv" Mar 07 15:15:00 crc kubenswrapper[4943]: I0307 15:15:00.394606 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b98a1692-83da-439e-8612-7a165d35171a-dispersionconf\") pod \"b98a1692-83da-439e-8612-7a165d35171a\" (UID: \"b98a1692-83da-439e-8612-7a165d35171a\") " Mar 07 15:15:00 crc kubenswrapper[4943]: I0307 15:15:00.394692 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b98a1692-83da-439e-8612-7a165d35171a-ring-data-devices\") pod \"b98a1692-83da-439e-8612-7a165d35171a\" (UID: \"b98a1692-83da-439e-8612-7a165d35171a\") " Mar 07 15:15:00 crc kubenswrapper[4943]: I0307 15:15:00.394861 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b98a1692-83da-439e-8612-7a165d35171a-swiftconf\") pod \"b98a1692-83da-439e-8612-7a165d35171a\" (UID: \"b98a1692-83da-439e-8612-7a165d35171a\") " Mar 07 15:15:00 crc kubenswrapper[4943]: I0307 15:15:00.395064 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jqxp5\" (UniqueName: \"kubernetes.io/projected/b98a1692-83da-439e-8612-7a165d35171a-kube-api-access-jqxp5\") pod \"b98a1692-83da-439e-8612-7a165d35171a\" (UID: \"b98a1692-83da-439e-8612-7a165d35171a\") " Mar 07 15:15:00 crc kubenswrapper[4943]: I0307 15:15:00.395142 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b98a1692-83da-439e-8612-7a165d35171a-etc-swift\") pod \"b98a1692-83da-439e-8612-7a165d35171a\" (UID: \"b98a1692-83da-439e-8612-7a165d35171a\") " Mar 07 15:15:00 crc kubenswrapper[4943]: I0307 15:15:00.395205 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b98a1692-83da-439e-8612-7a165d35171a-scripts\") pod \"b98a1692-83da-439e-8612-7a165d35171a\" (UID: \"b98a1692-83da-439e-8612-7a165d35171a\") " Mar 07 15:15:00 crc kubenswrapper[4943]: I0307 15:15:00.395413 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b98a1692-83da-439e-8612-7a165d35171a-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "b98a1692-83da-439e-8612-7a165d35171a" (UID: "b98a1692-83da-439e-8612-7a165d35171a"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:15:00 crc kubenswrapper[4943]: I0307 15:15:00.395812 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b98a1692-83da-439e-8612-7a165d35171a-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:00 crc kubenswrapper[4943]: I0307 15:15:00.396200 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b98a1692-83da-439e-8612-7a165d35171a-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "b98a1692-83da-439e-8612-7a165d35171a" (UID: "b98a1692-83da-439e-8612-7a165d35171a"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:15:00 crc kubenswrapper[4943]: I0307 15:15:00.400971 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b98a1692-83da-439e-8612-7a165d35171a-kube-api-access-jqxp5" (OuterVolumeSpecName: "kube-api-access-jqxp5") pod "b98a1692-83da-439e-8612-7a165d35171a" (UID: "b98a1692-83da-439e-8612-7a165d35171a"). InnerVolumeSpecName "kube-api-access-jqxp5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:15:00 crc kubenswrapper[4943]: I0307 15:15:00.412733 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b98a1692-83da-439e-8612-7a165d35171a-scripts" (OuterVolumeSpecName: "scripts") pod "b98a1692-83da-439e-8612-7a165d35171a" (UID: "b98a1692-83da-439e-8612-7a165d35171a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:15:00 crc kubenswrapper[4943]: I0307 15:15:00.421627 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b98a1692-83da-439e-8612-7a165d35171a-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "b98a1692-83da-439e-8612-7a165d35171a" (UID: "b98a1692-83da-439e-8612-7a165d35171a"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:15:00 crc kubenswrapper[4943]: I0307 15:15:00.421883 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b98a1692-83da-439e-8612-7a165d35171a-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "b98a1692-83da-439e-8612-7a165d35171a" (UID: "b98a1692-83da-439e-8612-7a165d35171a"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:15:00 crc kubenswrapper[4943]: I0307 15:15:00.497051 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b98a1692-83da-439e-8612-7a165d35171a-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:00 crc kubenswrapper[4943]: I0307 15:15:00.497110 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jqxp5\" (UniqueName: \"kubernetes.io/projected/b98a1692-83da-439e-8612-7a165d35171a-kube-api-access-jqxp5\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:00 crc kubenswrapper[4943]: I0307 15:15:00.497123 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b98a1692-83da-439e-8612-7a165d35171a-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:00 crc kubenswrapper[4943]: I0307 15:15:00.497136 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b98a1692-83da-439e-8612-7a165d35171a-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:00 crc kubenswrapper[4943]: I0307 15:15:00.497148 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b98a1692-83da-439e-8612-7a165d35171a-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:00 crc kubenswrapper[4943]: I0307 15:15:00.525334 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29548275-tqwsv" Mar 07 15:15:00 crc kubenswrapper[4943]: I0307 15:15:00.768650 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b98a1692-83da-439e-8612-7a165d35171a" path="/var/lib/kubelet/pods/b98a1692-83da-439e-8612-7a165d35171a/volumes" Mar 07 15:15:00 crc kubenswrapper[4943]: I0307 15:15:00.891132 4943 scope.go:117] "RemoveContainer" containerID="2ca4b462f42a1f213c6729ebe4a16c33d10eda9ae5e2e5aec8a4c07656096157" Mar 07 15:15:00 crc kubenswrapper[4943]: I0307 15:15:00.891294 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-bzsv2" Mar 07 15:15:01 crc kubenswrapper[4943]: I0307 15:15:01.057376 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29548275-tqwsv"] Mar 07 15:15:01 crc kubenswrapper[4943]: W0307 15:15:01.064883 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcedf5080_bb07_4ceb_9d64_fc45d9a25f74.slice/crio-3fce24c4204c348672dd99a4acf621edb128a22e00053f5b8f5386ca94682edb WatchSource:0}: Error finding container 3fce24c4204c348672dd99a4acf621edb128a22e00053f5b8f5386ca94682edb: Status 404 returned error can't find the container with id 3fce24c4204c348672dd99a4acf621edb128a22e00053f5b8f5386ca94682edb Mar 07 15:15:01 crc kubenswrapper[4943]: I0307 15:15:01.426383 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-5lvp6"] Mar 07 15:15:01 crc kubenswrapper[4943]: E0307 15:15:01.427076 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b98a1692-83da-439e-8612-7a165d35171a" containerName="swift-ring-rebalance" Mar 07 15:15:01 crc kubenswrapper[4943]: I0307 15:15:01.427127 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="b98a1692-83da-439e-8612-7a165d35171a" containerName="swift-ring-rebalance" Mar 07 15:15:01 crc kubenswrapper[4943]: I0307 15:15:01.427416 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="b98a1692-83da-439e-8612-7a165d35171a" containerName="swift-ring-rebalance" Mar 07 15:15:01 crc kubenswrapper[4943]: I0307 15:15:01.427948 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5lvp6" Mar 07 15:15:01 crc kubenswrapper[4943]: I0307 15:15:01.431746 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:15:01 crc kubenswrapper[4943]: I0307 15:15:01.433575 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:15:01 crc kubenswrapper[4943]: I0307 15:15:01.442837 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-5lvp6"] Mar 07 15:15:01 crc kubenswrapper[4943]: I0307 15:15:01.511677 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f3fc562d-2186-48af-905b-ee8d099a8a02-dispersionconf\") pod \"swift-ring-rebalance-debug-5lvp6\" (UID: \"f3fc562d-2186-48af-905b-ee8d099a8a02\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5lvp6" Mar 07 15:15:01 crc kubenswrapper[4943]: I0307 15:15:01.511720 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f3fc562d-2186-48af-905b-ee8d099a8a02-swiftconf\") pod \"swift-ring-rebalance-debug-5lvp6\" (UID: \"f3fc562d-2186-48af-905b-ee8d099a8a02\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5lvp6" Mar 07 15:15:01 crc kubenswrapper[4943]: I0307 15:15:01.511788 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f3fc562d-2186-48af-905b-ee8d099a8a02-scripts\") pod \"swift-ring-rebalance-debug-5lvp6\" (UID: \"f3fc562d-2186-48af-905b-ee8d099a8a02\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5lvp6" Mar 07 15:15:01 crc kubenswrapper[4943]: I0307 15:15:01.511833 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tr9q9\" (UniqueName: \"kubernetes.io/projected/f3fc562d-2186-48af-905b-ee8d099a8a02-kube-api-access-tr9q9\") pod \"swift-ring-rebalance-debug-5lvp6\" (UID: \"f3fc562d-2186-48af-905b-ee8d099a8a02\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5lvp6" Mar 07 15:15:01 crc kubenswrapper[4943]: I0307 15:15:01.511998 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f3fc562d-2186-48af-905b-ee8d099a8a02-ring-data-devices\") pod \"swift-ring-rebalance-debug-5lvp6\" (UID: \"f3fc562d-2186-48af-905b-ee8d099a8a02\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5lvp6" Mar 07 15:15:01 crc kubenswrapper[4943]: I0307 15:15:01.512066 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f3fc562d-2186-48af-905b-ee8d099a8a02-etc-swift\") pod \"swift-ring-rebalance-debug-5lvp6\" (UID: \"f3fc562d-2186-48af-905b-ee8d099a8a02\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5lvp6" Mar 07 15:15:01 crc kubenswrapper[4943]: I0307 15:15:01.613552 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f3fc562d-2186-48af-905b-ee8d099a8a02-scripts\") pod \"swift-ring-rebalance-debug-5lvp6\" (UID: \"f3fc562d-2186-48af-905b-ee8d099a8a02\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5lvp6" Mar 07 15:15:01 crc kubenswrapper[4943]: I0307 15:15:01.613703 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tr9q9\" (UniqueName: \"kubernetes.io/projected/f3fc562d-2186-48af-905b-ee8d099a8a02-kube-api-access-tr9q9\") pod \"swift-ring-rebalance-debug-5lvp6\" (UID: \"f3fc562d-2186-48af-905b-ee8d099a8a02\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5lvp6" Mar 07 15:15:01 crc kubenswrapper[4943]: I0307 15:15:01.613881 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f3fc562d-2186-48af-905b-ee8d099a8a02-ring-data-devices\") pod \"swift-ring-rebalance-debug-5lvp6\" (UID: \"f3fc562d-2186-48af-905b-ee8d099a8a02\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5lvp6" Mar 07 15:15:01 crc kubenswrapper[4943]: I0307 15:15:01.614012 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f3fc562d-2186-48af-905b-ee8d099a8a02-etc-swift\") pod \"swift-ring-rebalance-debug-5lvp6\" (UID: \"f3fc562d-2186-48af-905b-ee8d099a8a02\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5lvp6" Mar 07 15:15:01 crc kubenswrapper[4943]: I0307 15:15:01.614105 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f3fc562d-2186-48af-905b-ee8d099a8a02-dispersionconf\") pod \"swift-ring-rebalance-debug-5lvp6\" (UID: \"f3fc562d-2186-48af-905b-ee8d099a8a02\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5lvp6" Mar 07 15:15:01 crc kubenswrapper[4943]: I0307 15:15:01.614147 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f3fc562d-2186-48af-905b-ee8d099a8a02-swiftconf\") pod \"swift-ring-rebalance-debug-5lvp6\" (UID: \"f3fc562d-2186-48af-905b-ee8d099a8a02\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5lvp6" Mar 07 15:15:01 crc kubenswrapper[4943]: I0307 15:15:01.615222 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f3fc562d-2186-48af-905b-ee8d099a8a02-etc-swift\") pod \"swift-ring-rebalance-debug-5lvp6\" (UID: \"f3fc562d-2186-48af-905b-ee8d099a8a02\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5lvp6" Mar 07 15:15:01 crc kubenswrapper[4943]: I0307 15:15:01.615313 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f3fc562d-2186-48af-905b-ee8d099a8a02-scripts\") pod \"swift-ring-rebalance-debug-5lvp6\" (UID: \"f3fc562d-2186-48af-905b-ee8d099a8a02\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5lvp6" Mar 07 15:15:01 crc kubenswrapper[4943]: I0307 15:15:01.617527 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f3fc562d-2186-48af-905b-ee8d099a8a02-ring-data-devices\") pod \"swift-ring-rebalance-debug-5lvp6\" (UID: \"f3fc562d-2186-48af-905b-ee8d099a8a02\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5lvp6" Mar 07 15:15:01 crc kubenswrapper[4943]: I0307 15:15:01.624570 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f3fc562d-2186-48af-905b-ee8d099a8a02-swiftconf\") pod \"swift-ring-rebalance-debug-5lvp6\" (UID: \"f3fc562d-2186-48af-905b-ee8d099a8a02\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5lvp6" Mar 07 15:15:01 crc kubenswrapper[4943]: I0307 15:15:01.624775 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f3fc562d-2186-48af-905b-ee8d099a8a02-dispersionconf\") pod \"swift-ring-rebalance-debug-5lvp6\" (UID: \"f3fc562d-2186-48af-905b-ee8d099a8a02\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5lvp6" Mar 07 15:15:01 crc kubenswrapper[4943]: I0307 15:15:01.641234 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tr9q9\" (UniqueName: \"kubernetes.io/projected/f3fc562d-2186-48af-905b-ee8d099a8a02-kube-api-access-tr9q9\") pod \"swift-ring-rebalance-debug-5lvp6\" (UID: \"f3fc562d-2186-48af-905b-ee8d099a8a02\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5lvp6" Mar 07 15:15:01 crc kubenswrapper[4943]: I0307 15:15:01.804497 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5lvp6" Mar 07 15:15:01 crc kubenswrapper[4943]: I0307 15:15:01.932338 4943 generic.go:334] "Generic (PLEG): container finished" podID="cedf5080-bb07-4ceb-9d64-fc45d9a25f74" containerID="fbb78df50d39638df56446fbd2543b4a8c9bab3d4dbdda513c8a5a31eecf78a2" exitCode=0 Mar 07 15:15:01 crc kubenswrapper[4943]: I0307 15:15:01.932402 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29548275-tqwsv" event={"ID":"cedf5080-bb07-4ceb-9d64-fc45d9a25f74","Type":"ContainerDied","Data":"fbb78df50d39638df56446fbd2543b4a8c9bab3d4dbdda513c8a5a31eecf78a2"} Mar 07 15:15:01 crc kubenswrapper[4943]: I0307 15:15:01.932430 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29548275-tqwsv" event={"ID":"cedf5080-bb07-4ceb-9d64-fc45d9a25f74","Type":"ContainerStarted","Data":"3fce24c4204c348672dd99a4acf621edb128a22e00053f5b8f5386ca94682edb"} Mar 07 15:15:02 crc kubenswrapper[4943]: I0307 15:15:02.326403 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-5lvp6"] Mar 07 15:15:02 crc kubenswrapper[4943]: I0307 15:15:02.953450 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5lvp6" event={"ID":"f3fc562d-2186-48af-905b-ee8d099a8a02","Type":"ContainerStarted","Data":"4a27c160db96ac0508bd3c9253afe31401289535ce5e53a860e3771b2587e6da"} Mar 07 15:15:02 crc kubenswrapper[4943]: I0307 15:15:02.953809 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5lvp6" event={"ID":"f3fc562d-2186-48af-905b-ee8d099a8a02","Type":"ContainerStarted","Data":"d59868ecce80796142a98ab9e2f638f46bb3da6378a90a57b751059378c87bc3"} Mar 07 15:15:02 crc kubenswrapper[4943]: I0307 15:15:02.978206 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5lvp6" podStartSLOduration=1.978181669 podStartE2EDuration="1.978181669s" podCreationTimestamp="2026-03-07 15:15:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:15:02.975440692 +0000 UTC m=+2144.927577200" watchObservedRunningTime="2026-03-07 15:15:02.978181669 +0000 UTC m=+2144.930318187" Mar 07 15:15:03 crc kubenswrapper[4943]: I0307 15:15:03.382315 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29548275-tqwsv" Mar 07 15:15:03 crc kubenswrapper[4943]: I0307 15:15:03.553782 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fxncq\" (UniqueName: \"kubernetes.io/projected/cedf5080-bb07-4ceb-9d64-fc45d9a25f74-kube-api-access-fxncq\") pod \"cedf5080-bb07-4ceb-9d64-fc45d9a25f74\" (UID: \"cedf5080-bb07-4ceb-9d64-fc45d9a25f74\") " Mar 07 15:15:03 crc kubenswrapper[4943]: I0307 15:15:03.555280 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cedf5080-bb07-4ceb-9d64-fc45d9a25f74-secret-volume\") pod \"cedf5080-bb07-4ceb-9d64-fc45d9a25f74\" (UID: \"cedf5080-bb07-4ceb-9d64-fc45d9a25f74\") " Mar 07 15:15:03 crc kubenswrapper[4943]: I0307 15:15:03.555427 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cedf5080-bb07-4ceb-9d64-fc45d9a25f74-config-volume\") pod \"cedf5080-bb07-4ceb-9d64-fc45d9a25f74\" (UID: \"cedf5080-bb07-4ceb-9d64-fc45d9a25f74\") " Mar 07 15:15:03 crc kubenswrapper[4943]: I0307 15:15:03.555979 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cedf5080-bb07-4ceb-9d64-fc45d9a25f74-config-volume" (OuterVolumeSpecName: "config-volume") pod "cedf5080-bb07-4ceb-9d64-fc45d9a25f74" (UID: "cedf5080-bb07-4ceb-9d64-fc45d9a25f74"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:15:03 crc kubenswrapper[4943]: I0307 15:15:03.560309 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cedf5080-bb07-4ceb-9d64-fc45d9a25f74-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "cedf5080-bb07-4ceb-9d64-fc45d9a25f74" (UID: "cedf5080-bb07-4ceb-9d64-fc45d9a25f74"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:15:03 crc kubenswrapper[4943]: I0307 15:15:03.573150 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cedf5080-bb07-4ceb-9d64-fc45d9a25f74-kube-api-access-fxncq" (OuterVolumeSpecName: "kube-api-access-fxncq") pod "cedf5080-bb07-4ceb-9d64-fc45d9a25f74" (UID: "cedf5080-bb07-4ceb-9d64-fc45d9a25f74"). InnerVolumeSpecName "kube-api-access-fxncq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:15:03 crc kubenswrapper[4943]: I0307 15:15:03.657454 4943 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cedf5080-bb07-4ceb-9d64-fc45d9a25f74-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:03 crc kubenswrapper[4943]: I0307 15:15:03.657508 4943 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cedf5080-bb07-4ceb-9d64-fc45d9a25f74-config-volume\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:03 crc kubenswrapper[4943]: I0307 15:15:03.657530 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fxncq\" (UniqueName: \"kubernetes.io/projected/cedf5080-bb07-4ceb-9d64-fc45d9a25f74-kube-api-access-fxncq\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:03 crc kubenswrapper[4943]: I0307 15:15:03.964869 4943 generic.go:334] "Generic (PLEG): container finished" podID="f3fc562d-2186-48af-905b-ee8d099a8a02" containerID="4a27c160db96ac0508bd3c9253afe31401289535ce5e53a860e3771b2587e6da" exitCode=0 Mar 07 15:15:03 crc kubenswrapper[4943]: I0307 15:15:03.964965 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5lvp6" event={"ID":"f3fc562d-2186-48af-905b-ee8d099a8a02","Type":"ContainerDied","Data":"4a27c160db96ac0508bd3c9253afe31401289535ce5e53a860e3771b2587e6da"} Mar 07 15:15:03 crc kubenswrapper[4943]: I0307 15:15:03.966990 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29548275-tqwsv" event={"ID":"cedf5080-bb07-4ceb-9d64-fc45d9a25f74","Type":"ContainerDied","Data":"3fce24c4204c348672dd99a4acf621edb128a22e00053f5b8f5386ca94682edb"} Mar 07 15:15:03 crc kubenswrapper[4943]: I0307 15:15:03.967023 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3fce24c4204c348672dd99a4acf621edb128a22e00053f5b8f5386ca94682edb" Mar 07 15:15:03 crc kubenswrapper[4943]: I0307 15:15:03.967062 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29548275-tqwsv" Mar 07 15:15:04 crc kubenswrapper[4943]: I0307 15:15:04.486372 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29548230-zr6fd"] Mar 07 15:15:04 crc kubenswrapper[4943]: I0307 15:15:04.494108 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29548230-zr6fd"] Mar 07 15:15:04 crc kubenswrapper[4943]: I0307 15:15:04.767557 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7601d311-698b-4389-a9a2-e4e24fae23ea" path="/var/lib/kubelet/pods/7601d311-698b-4389-a9a2-e4e24fae23ea/volumes" Mar 07 15:15:05 crc kubenswrapper[4943]: I0307 15:15:05.375836 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5lvp6" Mar 07 15:15:05 crc kubenswrapper[4943]: I0307 15:15:05.387096 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f3fc562d-2186-48af-905b-ee8d099a8a02-swiftconf\") pod \"f3fc562d-2186-48af-905b-ee8d099a8a02\" (UID: \"f3fc562d-2186-48af-905b-ee8d099a8a02\") " Mar 07 15:15:05 crc kubenswrapper[4943]: I0307 15:15:05.436126 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3fc562d-2186-48af-905b-ee8d099a8a02-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "f3fc562d-2186-48af-905b-ee8d099a8a02" (UID: "f3fc562d-2186-48af-905b-ee8d099a8a02"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:15:05 crc kubenswrapper[4943]: I0307 15:15:05.471855 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-5lvp6"] Mar 07 15:15:05 crc kubenswrapper[4943]: I0307 15:15:05.479053 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-5lvp6"] Mar 07 15:15:05 crc kubenswrapper[4943]: I0307 15:15:05.488056 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f3fc562d-2186-48af-905b-ee8d099a8a02-ring-data-devices\") pod \"f3fc562d-2186-48af-905b-ee8d099a8a02\" (UID: \"f3fc562d-2186-48af-905b-ee8d099a8a02\") " Mar 07 15:15:05 crc kubenswrapper[4943]: I0307 15:15:05.488138 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f3fc562d-2186-48af-905b-ee8d099a8a02-dispersionconf\") pod \"f3fc562d-2186-48af-905b-ee8d099a8a02\" (UID: \"f3fc562d-2186-48af-905b-ee8d099a8a02\") " Mar 07 15:15:05 crc kubenswrapper[4943]: I0307 15:15:05.488177 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tr9q9\" (UniqueName: \"kubernetes.io/projected/f3fc562d-2186-48af-905b-ee8d099a8a02-kube-api-access-tr9q9\") pod \"f3fc562d-2186-48af-905b-ee8d099a8a02\" (UID: \"f3fc562d-2186-48af-905b-ee8d099a8a02\") " Mar 07 15:15:05 crc kubenswrapper[4943]: I0307 15:15:05.488257 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f3fc562d-2186-48af-905b-ee8d099a8a02-etc-swift\") pod \"f3fc562d-2186-48af-905b-ee8d099a8a02\" (UID: \"f3fc562d-2186-48af-905b-ee8d099a8a02\") " Mar 07 15:15:05 crc kubenswrapper[4943]: I0307 15:15:05.488293 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f3fc562d-2186-48af-905b-ee8d099a8a02-scripts\") pod \"f3fc562d-2186-48af-905b-ee8d099a8a02\" (UID: \"f3fc562d-2186-48af-905b-ee8d099a8a02\") " Mar 07 15:15:05 crc kubenswrapper[4943]: I0307 15:15:05.488530 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f3fc562d-2186-48af-905b-ee8d099a8a02-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:05 crc kubenswrapper[4943]: I0307 15:15:05.493913 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3fc562d-2186-48af-905b-ee8d099a8a02-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "f3fc562d-2186-48af-905b-ee8d099a8a02" (UID: "f3fc562d-2186-48af-905b-ee8d099a8a02"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:15:05 crc kubenswrapper[4943]: I0307 15:15:05.494720 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f3fc562d-2186-48af-905b-ee8d099a8a02-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "f3fc562d-2186-48af-905b-ee8d099a8a02" (UID: "f3fc562d-2186-48af-905b-ee8d099a8a02"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:15:05 crc kubenswrapper[4943]: I0307 15:15:05.521336 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3fc562d-2186-48af-905b-ee8d099a8a02-kube-api-access-tr9q9" (OuterVolumeSpecName: "kube-api-access-tr9q9") pod "f3fc562d-2186-48af-905b-ee8d099a8a02" (UID: "f3fc562d-2186-48af-905b-ee8d099a8a02"). InnerVolumeSpecName "kube-api-access-tr9q9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:15:05 crc kubenswrapper[4943]: I0307 15:15:05.521406 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f3fc562d-2186-48af-905b-ee8d099a8a02-scripts" (OuterVolumeSpecName: "scripts") pod "f3fc562d-2186-48af-905b-ee8d099a8a02" (UID: "f3fc562d-2186-48af-905b-ee8d099a8a02"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:15:05 crc kubenswrapper[4943]: I0307 15:15:05.537270 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3fc562d-2186-48af-905b-ee8d099a8a02-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "f3fc562d-2186-48af-905b-ee8d099a8a02" (UID: "f3fc562d-2186-48af-905b-ee8d099a8a02"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:15:05 crc kubenswrapper[4943]: I0307 15:15:05.589448 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f3fc562d-2186-48af-905b-ee8d099a8a02-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:05 crc kubenswrapper[4943]: I0307 15:15:05.589484 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f3fc562d-2186-48af-905b-ee8d099a8a02-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:05 crc kubenswrapper[4943]: I0307 15:15:05.589493 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f3fc562d-2186-48af-905b-ee8d099a8a02-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:05 crc kubenswrapper[4943]: I0307 15:15:05.589502 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f3fc562d-2186-48af-905b-ee8d099a8a02-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:05 crc kubenswrapper[4943]: I0307 15:15:05.589510 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tr9q9\" (UniqueName: \"kubernetes.io/projected/f3fc562d-2186-48af-905b-ee8d099a8a02-kube-api-access-tr9q9\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:05 crc kubenswrapper[4943]: I0307 15:15:05.988659 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d59868ecce80796142a98ab9e2f638f46bb3da6378a90a57b751059378c87bc3" Mar 07 15:15:05 crc kubenswrapper[4943]: I0307 15:15:05.988756 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5lvp6" Mar 07 15:15:06 crc kubenswrapper[4943]: I0307 15:15:06.637764 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-pnfbr"] Mar 07 15:15:06 crc kubenswrapper[4943]: E0307 15:15:06.638505 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3fc562d-2186-48af-905b-ee8d099a8a02" containerName="swift-ring-rebalance" Mar 07 15:15:06 crc kubenswrapper[4943]: I0307 15:15:06.638526 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3fc562d-2186-48af-905b-ee8d099a8a02" containerName="swift-ring-rebalance" Mar 07 15:15:06 crc kubenswrapper[4943]: E0307 15:15:06.638552 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cedf5080-bb07-4ceb-9d64-fc45d9a25f74" containerName="collect-profiles" Mar 07 15:15:06 crc kubenswrapper[4943]: I0307 15:15:06.638564 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="cedf5080-bb07-4ceb-9d64-fc45d9a25f74" containerName="collect-profiles" Mar 07 15:15:06 crc kubenswrapper[4943]: I0307 15:15:06.638899 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="cedf5080-bb07-4ceb-9d64-fc45d9a25f74" containerName="collect-profiles" Mar 07 15:15:06 crc kubenswrapper[4943]: I0307 15:15:06.639200 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3fc562d-2186-48af-905b-ee8d099a8a02" containerName="swift-ring-rebalance" Mar 07 15:15:06 crc kubenswrapper[4943]: I0307 15:15:06.639998 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pnfbr" Mar 07 15:15:06 crc kubenswrapper[4943]: I0307 15:15:06.643244 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:15:06 crc kubenswrapper[4943]: I0307 15:15:06.643369 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:15:06 crc kubenswrapper[4943]: I0307 15:15:06.653130 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-pnfbr"] Mar 07 15:15:06 crc kubenswrapper[4943]: I0307 15:15:06.708335 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1d2774fa-2fdd-4ca6-999b-2240668baea5-scripts\") pod \"swift-ring-rebalance-debug-pnfbr\" (UID: \"1d2774fa-2fdd-4ca6-999b-2240668baea5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pnfbr" Mar 07 15:15:06 crc kubenswrapper[4943]: I0307 15:15:06.708580 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1d2774fa-2fdd-4ca6-999b-2240668baea5-swiftconf\") pod \"swift-ring-rebalance-debug-pnfbr\" (UID: \"1d2774fa-2fdd-4ca6-999b-2240668baea5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pnfbr" Mar 07 15:15:06 crc kubenswrapper[4943]: I0307 15:15:06.708814 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1d2774fa-2fdd-4ca6-999b-2240668baea5-etc-swift\") pod \"swift-ring-rebalance-debug-pnfbr\" (UID: \"1d2774fa-2fdd-4ca6-999b-2240668baea5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pnfbr" Mar 07 15:15:06 crc kubenswrapper[4943]: I0307 15:15:06.708976 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1d2774fa-2fdd-4ca6-999b-2240668baea5-ring-data-devices\") pod \"swift-ring-rebalance-debug-pnfbr\" (UID: \"1d2774fa-2fdd-4ca6-999b-2240668baea5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pnfbr" Mar 07 15:15:06 crc kubenswrapper[4943]: I0307 15:15:06.709065 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xz2j2\" (UniqueName: \"kubernetes.io/projected/1d2774fa-2fdd-4ca6-999b-2240668baea5-kube-api-access-xz2j2\") pod \"swift-ring-rebalance-debug-pnfbr\" (UID: \"1d2774fa-2fdd-4ca6-999b-2240668baea5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pnfbr" Mar 07 15:15:06 crc kubenswrapper[4943]: I0307 15:15:06.709103 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1d2774fa-2fdd-4ca6-999b-2240668baea5-dispersionconf\") pod \"swift-ring-rebalance-debug-pnfbr\" (UID: \"1d2774fa-2fdd-4ca6-999b-2240668baea5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pnfbr" Mar 07 15:15:06 crc kubenswrapper[4943]: I0307 15:15:06.771422 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3fc562d-2186-48af-905b-ee8d099a8a02" path="/var/lib/kubelet/pods/f3fc562d-2186-48af-905b-ee8d099a8a02/volumes" Mar 07 15:15:06 crc kubenswrapper[4943]: I0307 15:15:06.810639 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1d2774fa-2fdd-4ca6-999b-2240668baea5-etc-swift\") pod \"swift-ring-rebalance-debug-pnfbr\" (UID: \"1d2774fa-2fdd-4ca6-999b-2240668baea5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pnfbr" Mar 07 15:15:06 crc kubenswrapper[4943]: I0307 15:15:06.810755 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1d2774fa-2fdd-4ca6-999b-2240668baea5-ring-data-devices\") pod \"swift-ring-rebalance-debug-pnfbr\" (UID: \"1d2774fa-2fdd-4ca6-999b-2240668baea5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pnfbr" Mar 07 15:15:06 crc kubenswrapper[4943]: I0307 15:15:06.810836 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xz2j2\" (UniqueName: \"kubernetes.io/projected/1d2774fa-2fdd-4ca6-999b-2240668baea5-kube-api-access-xz2j2\") pod \"swift-ring-rebalance-debug-pnfbr\" (UID: \"1d2774fa-2fdd-4ca6-999b-2240668baea5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pnfbr" Mar 07 15:15:06 crc kubenswrapper[4943]: I0307 15:15:06.810882 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1d2774fa-2fdd-4ca6-999b-2240668baea5-dispersionconf\") pod \"swift-ring-rebalance-debug-pnfbr\" (UID: \"1d2774fa-2fdd-4ca6-999b-2240668baea5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pnfbr" Mar 07 15:15:06 crc kubenswrapper[4943]: I0307 15:15:06.811030 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1d2774fa-2fdd-4ca6-999b-2240668baea5-scripts\") pod \"swift-ring-rebalance-debug-pnfbr\" (UID: \"1d2774fa-2fdd-4ca6-999b-2240668baea5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pnfbr" Mar 07 15:15:06 crc kubenswrapper[4943]: I0307 15:15:06.811149 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1d2774fa-2fdd-4ca6-999b-2240668baea5-swiftconf\") pod \"swift-ring-rebalance-debug-pnfbr\" (UID: \"1d2774fa-2fdd-4ca6-999b-2240668baea5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pnfbr" Mar 07 15:15:06 crc kubenswrapper[4943]: I0307 15:15:06.811492 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1d2774fa-2fdd-4ca6-999b-2240668baea5-etc-swift\") pod \"swift-ring-rebalance-debug-pnfbr\" (UID: \"1d2774fa-2fdd-4ca6-999b-2240668baea5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pnfbr" Mar 07 15:15:06 crc kubenswrapper[4943]: I0307 15:15:06.812549 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1d2774fa-2fdd-4ca6-999b-2240668baea5-ring-data-devices\") pod \"swift-ring-rebalance-debug-pnfbr\" (UID: \"1d2774fa-2fdd-4ca6-999b-2240668baea5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pnfbr" Mar 07 15:15:06 crc kubenswrapper[4943]: I0307 15:15:06.812614 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1d2774fa-2fdd-4ca6-999b-2240668baea5-scripts\") pod \"swift-ring-rebalance-debug-pnfbr\" (UID: \"1d2774fa-2fdd-4ca6-999b-2240668baea5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pnfbr" Mar 07 15:15:06 crc kubenswrapper[4943]: I0307 15:15:06.817544 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1d2774fa-2fdd-4ca6-999b-2240668baea5-dispersionconf\") pod \"swift-ring-rebalance-debug-pnfbr\" (UID: \"1d2774fa-2fdd-4ca6-999b-2240668baea5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pnfbr" Mar 07 15:15:06 crc kubenswrapper[4943]: I0307 15:15:06.821518 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1d2774fa-2fdd-4ca6-999b-2240668baea5-swiftconf\") pod \"swift-ring-rebalance-debug-pnfbr\" (UID: \"1d2774fa-2fdd-4ca6-999b-2240668baea5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pnfbr" Mar 07 15:15:06 crc kubenswrapper[4943]: I0307 15:15:06.833805 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xz2j2\" (UniqueName: \"kubernetes.io/projected/1d2774fa-2fdd-4ca6-999b-2240668baea5-kube-api-access-xz2j2\") pod \"swift-ring-rebalance-debug-pnfbr\" (UID: \"1d2774fa-2fdd-4ca6-999b-2240668baea5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pnfbr" Mar 07 15:15:06 crc kubenswrapper[4943]: I0307 15:15:06.982307 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pnfbr" Mar 07 15:15:07 crc kubenswrapper[4943]: I0307 15:15:07.458433 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-pnfbr"] Mar 07 15:15:08 crc kubenswrapper[4943]: I0307 15:15:08.018633 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pnfbr" event={"ID":"1d2774fa-2fdd-4ca6-999b-2240668baea5","Type":"ContainerStarted","Data":"8c720c78343408976be1ee4238fb1f82fbc67d4eb8664274728487be5bce2bed"} Mar 07 15:15:08 crc kubenswrapper[4943]: I0307 15:15:08.019027 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pnfbr" event={"ID":"1d2774fa-2fdd-4ca6-999b-2240668baea5","Type":"ContainerStarted","Data":"b86b8b6a449294c4ed427053f799d4b5645d460e44767fe11d253cd74a6d40e1"} Mar 07 15:15:08 crc kubenswrapper[4943]: I0307 15:15:08.042509 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pnfbr" podStartSLOduration=2.04248862 podStartE2EDuration="2.04248862s" podCreationTimestamp="2026-03-07 15:15:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:15:08.037511608 +0000 UTC m=+2149.989648146" watchObservedRunningTime="2026-03-07 15:15:08.04248862 +0000 UTC m=+2149.994625148" Mar 07 15:15:09 crc kubenswrapper[4943]: I0307 15:15:09.033048 4943 generic.go:334] "Generic (PLEG): container finished" podID="1d2774fa-2fdd-4ca6-999b-2240668baea5" containerID="8c720c78343408976be1ee4238fb1f82fbc67d4eb8664274728487be5bce2bed" exitCode=0 Mar 07 15:15:09 crc kubenswrapper[4943]: I0307 15:15:09.033111 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pnfbr" event={"ID":"1d2774fa-2fdd-4ca6-999b-2240668baea5","Type":"ContainerDied","Data":"8c720c78343408976be1ee4238fb1f82fbc67d4eb8664274728487be5bce2bed"} Mar 07 15:15:10 crc kubenswrapper[4943]: I0307 15:15:10.264397 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-h57bg"] Mar 07 15:15:10 crc kubenswrapper[4943]: I0307 15:15:10.266506 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h57bg" Mar 07 15:15:10 crc kubenswrapper[4943]: I0307 15:15:10.271881 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h57bg"] Mar 07 15:15:10 crc kubenswrapper[4943]: I0307 15:15:10.358823 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pnfbr" Mar 07 15:15:10 crc kubenswrapper[4943]: I0307 15:15:10.372302 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/253f1202-6997-48d2-8312-fda66447161c-utilities\") pod \"redhat-marketplace-h57bg\" (UID: \"253f1202-6997-48d2-8312-fda66447161c\") " pod="openshift-marketplace/redhat-marketplace-h57bg" Mar 07 15:15:10 crc kubenswrapper[4943]: I0307 15:15:10.372364 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/253f1202-6997-48d2-8312-fda66447161c-catalog-content\") pod \"redhat-marketplace-h57bg\" (UID: \"253f1202-6997-48d2-8312-fda66447161c\") " pod="openshift-marketplace/redhat-marketplace-h57bg" Mar 07 15:15:10 crc kubenswrapper[4943]: I0307 15:15:10.372597 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bvs2\" (UniqueName: \"kubernetes.io/projected/253f1202-6997-48d2-8312-fda66447161c-kube-api-access-2bvs2\") pod \"redhat-marketplace-h57bg\" (UID: \"253f1202-6997-48d2-8312-fda66447161c\") " pod="openshift-marketplace/redhat-marketplace-h57bg" Mar 07 15:15:10 crc kubenswrapper[4943]: I0307 15:15:10.392680 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-pnfbr"] Mar 07 15:15:10 crc kubenswrapper[4943]: I0307 15:15:10.399024 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-pnfbr"] Mar 07 15:15:10 crc kubenswrapper[4943]: I0307 15:15:10.473682 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1d2774fa-2fdd-4ca6-999b-2240668baea5-scripts\") pod \"1d2774fa-2fdd-4ca6-999b-2240668baea5\" (UID: \"1d2774fa-2fdd-4ca6-999b-2240668baea5\") " Mar 07 15:15:10 crc kubenswrapper[4943]: I0307 15:15:10.473769 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xz2j2\" (UniqueName: \"kubernetes.io/projected/1d2774fa-2fdd-4ca6-999b-2240668baea5-kube-api-access-xz2j2\") pod \"1d2774fa-2fdd-4ca6-999b-2240668baea5\" (UID: \"1d2774fa-2fdd-4ca6-999b-2240668baea5\") " Mar 07 15:15:10 crc kubenswrapper[4943]: I0307 15:15:10.473809 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1d2774fa-2fdd-4ca6-999b-2240668baea5-etc-swift\") pod \"1d2774fa-2fdd-4ca6-999b-2240668baea5\" (UID: \"1d2774fa-2fdd-4ca6-999b-2240668baea5\") " Mar 07 15:15:10 crc kubenswrapper[4943]: I0307 15:15:10.473839 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1d2774fa-2fdd-4ca6-999b-2240668baea5-ring-data-devices\") pod \"1d2774fa-2fdd-4ca6-999b-2240668baea5\" (UID: \"1d2774fa-2fdd-4ca6-999b-2240668baea5\") " Mar 07 15:15:10 crc kubenswrapper[4943]: I0307 15:15:10.473897 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1d2774fa-2fdd-4ca6-999b-2240668baea5-dispersionconf\") pod \"1d2774fa-2fdd-4ca6-999b-2240668baea5\" (UID: \"1d2774fa-2fdd-4ca6-999b-2240668baea5\") " Mar 07 15:15:10 crc kubenswrapper[4943]: I0307 15:15:10.473978 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1d2774fa-2fdd-4ca6-999b-2240668baea5-swiftconf\") pod \"1d2774fa-2fdd-4ca6-999b-2240668baea5\" (UID: \"1d2774fa-2fdd-4ca6-999b-2240668baea5\") " Mar 07 15:15:10 crc kubenswrapper[4943]: I0307 15:15:10.474229 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bvs2\" (UniqueName: \"kubernetes.io/projected/253f1202-6997-48d2-8312-fda66447161c-kube-api-access-2bvs2\") pod \"redhat-marketplace-h57bg\" (UID: \"253f1202-6997-48d2-8312-fda66447161c\") " pod="openshift-marketplace/redhat-marketplace-h57bg" Mar 07 15:15:10 crc kubenswrapper[4943]: I0307 15:15:10.474418 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d2774fa-2fdd-4ca6-999b-2240668baea5-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "1d2774fa-2fdd-4ca6-999b-2240668baea5" (UID: "1d2774fa-2fdd-4ca6-999b-2240668baea5"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:15:10 crc kubenswrapper[4943]: I0307 15:15:10.474641 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d2774fa-2fdd-4ca6-999b-2240668baea5-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "1d2774fa-2fdd-4ca6-999b-2240668baea5" (UID: "1d2774fa-2fdd-4ca6-999b-2240668baea5"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:15:10 crc kubenswrapper[4943]: I0307 15:15:10.475677 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/253f1202-6997-48d2-8312-fda66447161c-utilities\") pod \"redhat-marketplace-h57bg\" (UID: \"253f1202-6997-48d2-8312-fda66447161c\") " pod="openshift-marketplace/redhat-marketplace-h57bg" Mar 07 15:15:10 crc kubenswrapper[4943]: I0307 15:15:10.475766 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/253f1202-6997-48d2-8312-fda66447161c-catalog-content\") pod \"redhat-marketplace-h57bg\" (UID: \"253f1202-6997-48d2-8312-fda66447161c\") " pod="openshift-marketplace/redhat-marketplace-h57bg" Mar 07 15:15:10 crc kubenswrapper[4943]: I0307 15:15:10.475859 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1d2774fa-2fdd-4ca6-999b-2240668baea5-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:10 crc kubenswrapper[4943]: I0307 15:15:10.476153 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/253f1202-6997-48d2-8312-fda66447161c-utilities\") pod \"redhat-marketplace-h57bg\" (UID: \"253f1202-6997-48d2-8312-fda66447161c\") " pod="openshift-marketplace/redhat-marketplace-h57bg" Mar 07 15:15:10 crc kubenswrapper[4943]: I0307 15:15:10.476200 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/253f1202-6997-48d2-8312-fda66447161c-catalog-content\") pod \"redhat-marketplace-h57bg\" (UID: \"253f1202-6997-48d2-8312-fda66447161c\") " pod="openshift-marketplace/redhat-marketplace-h57bg" Mar 07 15:15:10 crc kubenswrapper[4943]: I0307 15:15:10.475873 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1d2774fa-2fdd-4ca6-999b-2240668baea5-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:10 crc kubenswrapper[4943]: I0307 15:15:10.491128 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d2774fa-2fdd-4ca6-999b-2240668baea5-kube-api-access-xz2j2" (OuterVolumeSpecName: "kube-api-access-xz2j2") pod "1d2774fa-2fdd-4ca6-999b-2240668baea5" (UID: "1d2774fa-2fdd-4ca6-999b-2240668baea5"). InnerVolumeSpecName "kube-api-access-xz2j2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:15:10 crc kubenswrapper[4943]: I0307 15:15:10.492653 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d2774fa-2fdd-4ca6-999b-2240668baea5-scripts" (OuterVolumeSpecName: "scripts") pod "1d2774fa-2fdd-4ca6-999b-2240668baea5" (UID: "1d2774fa-2fdd-4ca6-999b-2240668baea5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:15:10 crc kubenswrapper[4943]: I0307 15:15:10.496704 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bvs2\" (UniqueName: \"kubernetes.io/projected/253f1202-6997-48d2-8312-fda66447161c-kube-api-access-2bvs2\") pod \"redhat-marketplace-h57bg\" (UID: \"253f1202-6997-48d2-8312-fda66447161c\") " pod="openshift-marketplace/redhat-marketplace-h57bg" Mar 07 15:15:10 crc kubenswrapper[4943]: I0307 15:15:10.497856 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d2774fa-2fdd-4ca6-999b-2240668baea5-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "1d2774fa-2fdd-4ca6-999b-2240668baea5" (UID: "1d2774fa-2fdd-4ca6-999b-2240668baea5"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:15:10 crc kubenswrapper[4943]: I0307 15:15:10.507546 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d2774fa-2fdd-4ca6-999b-2240668baea5-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "1d2774fa-2fdd-4ca6-999b-2240668baea5" (UID: "1d2774fa-2fdd-4ca6-999b-2240668baea5"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:15:10 crc kubenswrapper[4943]: I0307 15:15:10.577814 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xz2j2\" (UniqueName: \"kubernetes.io/projected/1d2774fa-2fdd-4ca6-999b-2240668baea5-kube-api-access-xz2j2\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:10 crc kubenswrapper[4943]: I0307 15:15:10.577856 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1d2774fa-2fdd-4ca6-999b-2240668baea5-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:10 crc kubenswrapper[4943]: I0307 15:15:10.577869 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1d2774fa-2fdd-4ca6-999b-2240668baea5-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:10 crc kubenswrapper[4943]: I0307 15:15:10.577880 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1d2774fa-2fdd-4ca6-999b-2240668baea5-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:10 crc kubenswrapper[4943]: I0307 15:15:10.656266 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h57bg" Mar 07 15:15:10 crc kubenswrapper[4943]: I0307 15:15:10.766863 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d2774fa-2fdd-4ca6-999b-2240668baea5" path="/var/lib/kubelet/pods/1d2774fa-2fdd-4ca6-999b-2240668baea5/volumes" Mar 07 15:15:11 crc kubenswrapper[4943]: I0307 15:15:11.053811 4943 scope.go:117] "RemoveContainer" containerID="8c720c78343408976be1ee4238fb1f82fbc67d4eb8664274728487be5bce2bed" Mar 07 15:15:11 crc kubenswrapper[4943]: I0307 15:15:11.053896 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pnfbr" Mar 07 15:15:11 crc kubenswrapper[4943]: I0307 15:15:11.157913 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h57bg"] Mar 07 15:15:11 crc kubenswrapper[4943]: W0307 15:15:11.163698 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod253f1202_6997_48d2_8312_fda66447161c.slice/crio-c7a7d8c9d2e9b754958c2e4c8929162e3ec6caab050ade1374010034f37bc4ee WatchSource:0}: Error finding container c7a7d8c9d2e9b754958c2e4c8929162e3ec6caab050ade1374010034f37bc4ee: Status 404 returned error can't find the container with id c7a7d8c9d2e9b754958c2e4c8929162e3ec6caab050ade1374010034f37bc4ee Mar 07 15:15:11 crc kubenswrapper[4943]: I0307 15:15:11.545994 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-zc4s5"] Mar 07 15:15:11 crc kubenswrapper[4943]: E0307 15:15:11.546424 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d2774fa-2fdd-4ca6-999b-2240668baea5" containerName="swift-ring-rebalance" Mar 07 15:15:11 crc kubenswrapper[4943]: I0307 15:15:11.546443 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d2774fa-2fdd-4ca6-999b-2240668baea5" containerName="swift-ring-rebalance" Mar 07 15:15:11 crc kubenswrapper[4943]: I0307 15:15:11.546632 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d2774fa-2fdd-4ca6-999b-2240668baea5" containerName="swift-ring-rebalance" Mar 07 15:15:11 crc kubenswrapper[4943]: I0307 15:15:11.547294 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-zc4s5" Mar 07 15:15:11 crc kubenswrapper[4943]: I0307 15:15:11.549029 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:15:11 crc kubenswrapper[4943]: I0307 15:15:11.550902 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:15:11 crc kubenswrapper[4943]: I0307 15:15:11.551673 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-zc4s5"] Mar 07 15:15:11 crc kubenswrapper[4943]: I0307 15:15:11.694677 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/fb620c1d-3b17-426d-9a50-b4392474b78b-etc-swift\") pod \"swift-ring-rebalance-debug-zc4s5\" (UID: \"fb620c1d-3b17-426d-9a50-b4392474b78b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zc4s5" Mar 07 15:15:11 crc kubenswrapper[4943]: I0307 15:15:11.695101 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/fb620c1d-3b17-426d-9a50-b4392474b78b-ring-data-devices\") pod \"swift-ring-rebalance-debug-zc4s5\" (UID: \"fb620c1d-3b17-426d-9a50-b4392474b78b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zc4s5" Mar 07 15:15:11 crc kubenswrapper[4943]: I0307 15:15:11.695188 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brx42\" (UniqueName: \"kubernetes.io/projected/fb620c1d-3b17-426d-9a50-b4392474b78b-kube-api-access-brx42\") pod \"swift-ring-rebalance-debug-zc4s5\" (UID: \"fb620c1d-3b17-426d-9a50-b4392474b78b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zc4s5" Mar 07 15:15:11 crc kubenswrapper[4943]: I0307 15:15:11.695262 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fb620c1d-3b17-426d-9a50-b4392474b78b-scripts\") pod \"swift-ring-rebalance-debug-zc4s5\" (UID: \"fb620c1d-3b17-426d-9a50-b4392474b78b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zc4s5" Mar 07 15:15:11 crc kubenswrapper[4943]: I0307 15:15:11.695295 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/fb620c1d-3b17-426d-9a50-b4392474b78b-swiftconf\") pod \"swift-ring-rebalance-debug-zc4s5\" (UID: \"fb620c1d-3b17-426d-9a50-b4392474b78b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zc4s5" Mar 07 15:15:11 crc kubenswrapper[4943]: I0307 15:15:11.695332 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/fb620c1d-3b17-426d-9a50-b4392474b78b-dispersionconf\") pod \"swift-ring-rebalance-debug-zc4s5\" (UID: \"fb620c1d-3b17-426d-9a50-b4392474b78b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zc4s5" Mar 07 15:15:11 crc kubenswrapper[4943]: I0307 15:15:11.796876 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/fb620c1d-3b17-426d-9a50-b4392474b78b-ring-data-devices\") pod \"swift-ring-rebalance-debug-zc4s5\" (UID: \"fb620c1d-3b17-426d-9a50-b4392474b78b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zc4s5" Mar 07 15:15:11 crc kubenswrapper[4943]: I0307 15:15:11.797025 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brx42\" (UniqueName: \"kubernetes.io/projected/fb620c1d-3b17-426d-9a50-b4392474b78b-kube-api-access-brx42\") pod \"swift-ring-rebalance-debug-zc4s5\" (UID: \"fb620c1d-3b17-426d-9a50-b4392474b78b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zc4s5" Mar 07 15:15:11 crc kubenswrapper[4943]: I0307 15:15:11.797117 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fb620c1d-3b17-426d-9a50-b4392474b78b-scripts\") pod \"swift-ring-rebalance-debug-zc4s5\" (UID: \"fb620c1d-3b17-426d-9a50-b4392474b78b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zc4s5" Mar 07 15:15:11 crc kubenswrapper[4943]: I0307 15:15:11.797149 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/fb620c1d-3b17-426d-9a50-b4392474b78b-swiftconf\") pod \"swift-ring-rebalance-debug-zc4s5\" (UID: \"fb620c1d-3b17-426d-9a50-b4392474b78b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zc4s5" Mar 07 15:15:11 crc kubenswrapper[4943]: I0307 15:15:11.797186 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/fb620c1d-3b17-426d-9a50-b4392474b78b-dispersionconf\") pod \"swift-ring-rebalance-debug-zc4s5\" (UID: \"fb620c1d-3b17-426d-9a50-b4392474b78b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zc4s5" Mar 07 15:15:11 crc kubenswrapper[4943]: I0307 15:15:11.797316 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/fb620c1d-3b17-426d-9a50-b4392474b78b-etc-swift\") pod \"swift-ring-rebalance-debug-zc4s5\" (UID: \"fb620c1d-3b17-426d-9a50-b4392474b78b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zc4s5" Mar 07 15:15:11 crc kubenswrapper[4943]: I0307 15:15:11.798183 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/fb620c1d-3b17-426d-9a50-b4392474b78b-etc-swift\") pod \"swift-ring-rebalance-debug-zc4s5\" (UID: \"fb620c1d-3b17-426d-9a50-b4392474b78b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zc4s5" Mar 07 15:15:11 crc kubenswrapper[4943]: I0307 15:15:11.798354 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/fb620c1d-3b17-426d-9a50-b4392474b78b-ring-data-devices\") pod \"swift-ring-rebalance-debug-zc4s5\" (UID: \"fb620c1d-3b17-426d-9a50-b4392474b78b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zc4s5" Mar 07 15:15:11 crc kubenswrapper[4943]: I0307 15:15:11.798476 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fb620c1d-3b17-426d-9a50-b4392474b78b-scripts\") pod \"swift-ring-rebalance-debug-zc4s5\" (UID: \"fb620c1d-3b17-426d-9a50-b4392474b78b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zc4s5" Mar 07 15:15:11 crc kubenswrapper[4943]: I0307 15:15:11.806068 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/fb620c1d-3b17-426d-9a50-b4392474b78b-swiftconf\") pod \"swift-ring-rebalance-debug-zc4s5\" (UID: \"fb620c1d-3b17-426d-9a50-b4392474b78b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zc4s5" Mar 07 15:15:11 crc kubenswrapper[4943]: I0307 15:15:11.809225 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/fb620c1d-3b17-426d-9a50-b4392474b78b-dispersionconf\") pod \"swift-ring-rebalance-debug-zc4s5\" (UID: \"fb620c1d-3b17-426d-9a50-b4392474b78b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zc4s5" Mar 07 15:15:11 crc kubenswrapper[4943]: I0307 15:15:11.822660 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brx42\" (UniqueName: \"kubernetes.io/projected/fb620c1d-3b17-426d-9a50-b4392474b78b-kube-api-access-brx42\") pod \"swift-ring-rebalance-debug-zc4s5\" (UID: \"fb620c1d-3b17-426d-9a50-b4392474b78b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zc4s5" Mar 07 15:15:11 crc kubenswrapper[4943]: I0307 15:15:11.870413 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-zc4s5" Mar 07 15:15:12 crc kubenswrapper[4943]: I0307 15:15:12.068618 4943 generic.go:334] "Generic (PLEG): container finished" podID="253f1202-6997-48d2-8312-fda66447161c" containerID="9c6128ca0ec73e2c85a0c9ac1e24ab2f61c62b4123e544345f688df9c66d76d0" exitCode=0 Mar 07 15:15:12 crc kubenswrapper[4943]: I0307 15:15:12.068663 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h57bg" event={"ID":"253f1202-6997-48d2-8312-fda66447161c","Type":"ContainerDied","Data":"9c6128ca0ec73e2c85a0c9ac1e24ab2f61c62b4123e544345f688df9c66d76d0"} Mar 07 15:15:12 crc kubenswrapper[4943]: I0307 15:15:12.068693 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h57bg" event={"ID":"253f1202-6997-48d2-8312-fda66447161c","Type":"ContainerStarted","Data":"c7a7d8c9d2e9b754958c2e4c8929162e3ec6caab050ade1374010034f37bc4ee"} Mar 07 15:15:12 crc kubenswrapper[4943]: I0307 15:15:12.305192 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-zc4s5"] Mar 07 15:15:12 crc kubenswrapper[4943]: W0307 15:15:12.307960 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfb620c1d_3b17_426d_9a50_b4392474b78b.slice/crio-af70b4b272af7f26892763748fb579e17ab3a4487f9c1a8c584cc65088f33349 WatchSource:0}: Error finding container af70b4b272af7f26892763748fb579e17ab3a4487f9c1a8c584cc65088f33349: Status 404 returned error can't find the container with id af70b4b272af7f26892763748fb579e17ab3a4487f9c1a8c584cc65088f33349 Mar 07 15:15:13 crc kubenswrapper[4943]: I0307 15:15:13.085756 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-zc4s5" event={"ID":"fb620c1d-3b17-426d-9a50-b4392474b78b","Type":"ContainerStarted","Data":"a0eaca674ddb1faa49197ac8b9583a9de6ff525edaa8faa0e36b3a34246dd60d"} Mar 07 15:15:13 crc kubenswrapper[4943]: I0307 15:15:13.086250 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-zc4s5" event={"ID":"fb620c1d-3b17-426d-9a50-b4392474b78b","Type":"ContainerStarted","Data":"af70b4b272af7f26892763748fb579e17ab3a4487f9c1a8c584cc65088f33349"} Mar 07 15:15:13 crc kubenswrapper[4943]: I0307 15:15:13.092222 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h57bg" event={"ID":"253f1202-6997-48d2-8312-fda66447161c","Type":"ContainerStarted","Data":"df6cc34f1b0db7d8086f552fd9403c5ec9657059855956c620723d8b9ce7ff31"} Mar 07 15:15:13 crc kubenswrapper[4943]: I0307 15:15:13.117473 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-zc4s5" podStartSLOduration=2.117425049 podStartE2EDuration="2.117425049s" podCreationTimestamp="2026-03-07 15:15:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:15:13.113971325 +0000 UTC m=+2155.066107883" watchObservedRunningTime="2026-03-07 15:15:13.117425049 +0000 UTC m=+2155.069561557" Mar 07 15:15:14 crc kubenswrapper[4943]: I0307 15:15:14.106922 4943 generic.go:334] "Generic (PLEG): container finished" podID="fb620c1d-3b17-426d-9a50-b4392474b78b" containerID="a0eaca674ddb1faa49197ac8b9583a9de6ff525edaa8faa0e36b3a34246dd60d" exitCode=0 Mar 07 15:15:14 crc kubenswrapper[4943]: I0307 15:15:14.107081 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-zc4s5" event={"ID":"fb620c1d-3b17-426d-9a50-b4392474b78b","Type":"ContainerDied","Data":"a0eaca674ddb1faa49197ac8b9583a9de6ff525edaa8faa0e36b3a34246dd60d"} Mar 07 15:15:14 crc kubenswrapper[4943]: I0307 15:15:14.111390 4943 generic.go:334] "Generic (PLEG): container finished" podID="253f1202-6997-48d2-8312-fda66447161c" containerID="df6cc34f1b0db7d8086f552fd9403c5ec9657059855956c620723d8b9ce7ff31" exitCode=0 Mar 07 15:15:14 crc kubenswrapper[4943]: I0307 15:15:14.111431 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h57bg" event={"ID":"253f1202-6997-48d2-8312-fda66447161c","Type":"ContainerDied","Data":"df6cc34f1b0db7d8086f552fd9403c5ec9657059855956c620723d8b9ce7ff31"} Mar 07 15:15:15 crc kubenswrapper[4943]: I0307 15:15:15.118957 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h57bg" event={"ID":"253f1202-6997-48d2-8312-fda66447161c","Type":"ContainerStarted","Data":"a03458f2c0dfbd48b5cfc5edb914541f94cc5894e42d9641b2a83e09c7eedba7"} Mar 07 15:15:15 crc kubenswrapper[4943]: I0307 15:15:15.137066 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-h57bg" podStartSLOduration=2.6484264189999998 podStartE2EDuration="5.137046225s" podCreationTimestamp="2026-03-07 15:15:10 +0000 UTC" firstStartedPulling="2026-03-07 15:15:12.073189061 +0000 UTC m=+2154.025325599" lastFinishedPulling="2026-03-07 15:15:14.561808907 +0000 UTC m=+2156.513945405" observedRunningTime="2026-03-07 15:15:15.134949023 +0000 UTC m=+2157.087085531" watchObservedRunningTime="2026-03-07 15:15:15.137046225 +0000 UTC m=+2157.089182723" Mar 07 15:15:15 crc kubenswrapper[4943]: I0307 15:15:15.422743 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-zc4s5" Mar 07 15:15:15 crc kubenswrapper[4943]: I0307 15:15:15.484762 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-zc4s5"] Mar 07 15:15:15 crc kubenswrapper[4943]: I0307 15:15:15.492211 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-zc4s5"] Mar 07 15:15:15 crc kubenswrapper[4943]: I0307 15:15:15.576870 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/fb620c1d-3b17-426d-9a50-b4392474b78b-dispersionconf\") pod \"fb620c1d-3b17-426d-9a50-b4392474b78b\" (UID: \"fb620c1d-3b17-426d-9a50-b4392474b78b\") " Mar 07 15:15:15 crc kubenswrapper[4943]: I0307 15:15:15.576966 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/fb620c1d-3b17-426d-9a50-b4392474b78b-etc-swift\") pod \"fb620c1d-3b17-426d-9a50-b4392474b78b\" (UID: \"fb620c1d-3b17-426d-9a50-b4392474b78b\") " Mar 07 15:15:15 crc kubenswrapper[4943]: I0307 15:15:15.576996 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/fb620c1d-3b17-426d-9a50-b4392474b78b-ring-data-devices\") pod \"fb620c1d-3b17-426d-9a50-b4392474b78b\" (UID: \"fb620c1d-3b17-426d-9a50-b4392474b78b\") " Mar 07 15:15:15 crc kubenswrapper[4943]: I0307 15:15:15.577038 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fb620c1d-3b17-426d-9a50-b4392474b78b-scripts\") pod \"fb620c1d-3b17-426d-9a50-b4392474b78b\" (UID: \"fb620c1d-3b17-426d-9a50-b4392474b78b\") " Mar 07 15:15:15 crc kubenswrapper[4943]: I0307 15:15:15.577061 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-brx42\" (UniqueName: \"kubernetes.io/projected/fb620c1d-3b17-426d-9a50-b4392474b78b-kube-api-access-brx42\") pod \"fb620c1d-3b17-426d-9a50-b4392474b78b\" (UID: \"fb620c1d-3b17-426d-9a50-b4392474b78b\") " Mar 07 15:15:15 crc kubenswrapper[4943]: I0307 15:15:15.577149 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/fb620c1d-3b17-426d-9a50-b4392474b78b-swiftconf\") pod \"fb620c1d-3b17-426d-9a50-b4392474b78b\" (UID: \"fb620c1d-3b17-426d-9a50-b4392474b78b\") " Mar 07 15:15:15 crc kubenswrapper[4943]: I0307 15:15:15.577871 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb620c1d-3b17-426d-9a50-b4392474b78b-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "fb620c1d-3b17-426d-9a50-b4392474b78b" (UID: "fb620c1d-3b17-426d-9a50-b4392474b78b"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:15:15 crc kubenswrapper[4943]: I0307 15:15:15.578163 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb620c1d-3b17-426d-9a50-b4392474b78b-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "fb620c1d-3b17-426d-9a50-b4392474b78b" (UID: "fb620c1d-3b17-426d-9a50-b4392474b78b"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:15:15 crc kubenswrapper[4943]: I0307 15:15:15.583813 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb620c1d-3b17-426d-9a50-b4392474b78b-kube-api-access-brx42" (OuterVolumeSpecName: "kube-api-access-brx42") pod "fb620c1d-3b17-426d-9a50-b4392474b78b" (UID: "fb620c1d-3b17-426d-9a50-b4392474b78b"). InnerVolumeSpecName "kube-api-access-brx42". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:15:15 crc kubenswrapper[4943]: I0307 15:15:15.600265 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb620c1d-3b17-426d-9a50-b4392474b78b-scripts" (OuterVolumeSpecName: "scripts") pod "fb620c1d-3b17-426d-9a50-b4392474b78b" (UID: "fb620c1d-3b17-426d-9a50-b4392474b78b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:15:15 crc kubenswrapper[4943]: I0307 15:15:15.602620 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb620c1d-3b17-426d-9a50-b4392474b78b-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "fb620c1d-3b17-426d-9a50-b4392474b78b" (UID: "fb620c1d-3b17-426d-9a50-b4392474b78b"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:15:15 crc kubenswrapper[4943]: I0307 15:15:15.608268 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb620c1d-3b17-426d-9a50-b4392474b78b-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "fb620c1d-3b17-426d-9a50-b4392474b78b" (UID: "fb620c1d-3b17-426d-9a50-b4392474b78b"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:15:15 crc kubenswrapper[4943]: I0307 15:15:15.679873 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/fb620c1d-3b17-426d-9a50-b4392474b78b-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:15 crc kubenswrapper[4943]: I0307 15:15:15.679964 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/fb620c1d-3b17-426d-9a50-b4392474b78b-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:15 crc kubenswrapper[4943]: I0307 15:15:15.679980 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/fb620c1d-3b17-426d-9a50-b4392474b78b-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:15 crc kubenswrapper[4943]: I0307 15:15:15.679999 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fb620c1d-3b17-426d-9a50-b4392474b78b-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:15 crc kubenswrapper[4943]: I0307 15:15:15.680012 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-brx42\" (UniqueName: \"kubernetes.io/projected/fb620c1d-3b17-426d-9a50-b4392474b78b-kube-api-access-brx42\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:15 crc kubenswrapper[4943]: I0307 15:15:15.680028 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/fb620c1d-3b17-426d-9a50-b4392474b78b-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:16 crc kubenswrapper[4943]: I0307 15:15:16.130043 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="af70b4b272af7f26892763748fb579e17ab3a4487f9c1a8c584cc65088f33349" Mar 07 15:15:16 crc kubenswrapper[4943]: I0307 15:15:16.130966 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-zc4s5" Mar 07 15:15:16 crc kubenswrapper[4943]: I0307 15:15:16.731536 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-t4l6p"] Mar 07 15:15:16 crc kubenswrapper[4943]: E0307 15:15:16.731984 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb620c1d-3b17-426d-9a50-b4392474b78b" containerName="swift-ring-rebalance" Mar 07 15:15:16 crc kubenswrapper[4943]: I0307 15:15:16.732008 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb620c1d-3b17-426d-9a50-b4392474b78b" containerName="swift-ring-rebalance" Mar 07 15:15:16 crc kubenswrapper[4943]: I0307 15:15:16.732284 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb620c1d-3b17-426d-9a50-b4392474b78b" containerName="swift-ring-rebalance" Mar 07 15:15:16 crc kubenswrapper[4943]: I0307 15:15:16.733063 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-t4l6p" Mar 07 15:15:16 crc kubenswrapper[4943]: I0307 15:15:16.744249 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:15:16 crc kubenswrapper[4943]: I0307 15:15:16.744647 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:15:16 crc kubenswrapper[4943]: I0307 15:15:16.750584 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-t4l6p"] Mar 07 15:15:16 crc kubenswrapper[4943]: I0307 15:15:16.778346 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb620c1d-3b17-426d-9a50-b4392474b78b" path="/var/lib/kubelet/pods/fb620c1d-3b17-426d-9a50-b4392474b78b/volumes" Mar 07 15:15:16 crc kubenswrapper[4943]: I0307 15:15:16.901797 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/caa16732-5783-4a87-99d1-b552dc3af955-ring-data-devices\") pod \"swift-ring-rebalance-debug-t4l6p\" (UID: \"caa16732-5783-4a87-99d1-b552dc3af955\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-t4l6p" Mar 07 15:15:16 crc kubenswrapper[4943]: I0307 15:15:16.901917 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77xm9\" (UniqueName: \"kubernetes.io/projected/caa16732-5783-4a87-99d1-b552dc3af955-kube-api-access-77xm9\") pod \"swift-ring-rebalance-debug-t4l6p\" (UID: \"caa16732-5783-4a87-99d1-b552dc3af955\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-t4l6p" Mar 07 15:15:16 crc kubenswrapper[4943]: I0307 15:15:16.901994 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/caa16732-5783-4a87-99d1-b552dc3af955-etc-swift\") pod \"swift-ring-rebalance-debug-t4l6p\" (UID: \"caa16732-5783-4a87-99d1-b552dc3af955\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-t4l6p" Mar 07 15:15:16 crc kubenswrapper[4943]: I0307 15:15:16.902030 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/caa16732-5783-4a87-99d1-b552dc3af955-swiftconf\") pod \"swift-ring-rebalance-debug-t4l6p\" (UID: \"caa16732-5783-4a87-99d1-b552dc3af955\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-t4l6p" Mar 07 15:15:16 crc kubenswrapper[4943]: I0307 15:15:16.902412 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/caa16732-5783-4a87-99d1-b552dc3af955-dispersionconf\") pod \"swift-ring-rebalance-debug-t4l6p\" (UID: \"caa16732-5783-4a87-99d1-b552dc3af955\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-t4l6p" Mar 07 15:15:16 crc kubenswrapper[4943]: I0307 15:15:16.902564 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/caa16732-5783-4a87-99d1-b552dc3af955-scripts\") pod \"swift-ring-rebalance-debug-t4l6p\" (UID: \"caa16732-5783-4a87-99d1-b552dc3af955\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-t4l6p" Mar 07 15:15:17 crc kubenswrapper[4943]: I0307 15:15:17.003908 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/caa16732-5783-4a87-99d1-b552dc3af955-ring-data-devices\") pod \"swift-ring-rebalance-debug-t4l6p\" (UID: \"caa16732-5783-4a87-99d1-b552dc3af955\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-t4l6p" Mar 07 15:15:17 crc kubenswrapper[4943]: I0307 15:15:17.004022 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77xm9\" (UniqueName: \"kubernetes.io/projected/caa16732-5783-4a87-99d1-b552dc3af955-kube-api-access-77xm9\") pod \"swift-ring-rebalance-debug-t4l6p\" (UID: \"caa16732-5783-4a87-99d1-b552dc3af955\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-t4l6p" Mar 07 15:15:17 crc kubenswrapper[4943]: I0307 15:15:17.004083 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/caa16732-5783-4a87-99d1-b552dc3af955-etc-swift\") pod \"swift-ring-rebalance-debug-t4l6p\" (UID: \"caa16732-5783-4a87-99d1-b552dc3af955\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-t4l6p" Mar 07 15:15:17 crc kubenswrapper[4943]: I0307 15:15:17.004145 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/caa16732-5783-4a87-99d1-b552dc3af955-swiftconf\") pod \"swift-ring-rebalance-debug-t4l6p\" (UID: \"caa16732-5783-4a87-99d1-b552dc3af955\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-t4l6p" Mar 07 15:15:17 crc kubenswrapper[4943]: I0307 15:15:17.004254 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/caa16732-5783-4a87-99d1-b552dc3af955-dispersionconf\") pod \"swift-ring-rebalance-debug-t4l6p\" (UID: \"caa16732-5783-4a87-99d1-b552dc3af955\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-t4l6p" Mar 07 15:15:17 crc kubenswrapper[4943]: I0307 15:15:17.004311 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/caa16732-5783-4a87-99d1-b552dc3af955-scripts\") pod \"swift-ring-rebalance-debug-t4l6p\" (UID: \"caa16732-5783-4a87-99d1-b552dc3af955\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-t4l6p" Mar 07 15:15:17 crc kubenswrapper[4943]: I0307 15:15:17.005073 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/caa16732-5783-4a87-99d1-b552dc3af955-etc-swift\") pod \"swift-ring-rebalance-debug-t4l6p\" (UID: \"caa16732-5783-4a87-99d1-b552dc3af955\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-t4l6p" Mar 07 15:15:17 crc kubenswrapper[4943]: I0307 15:15:17.005240 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/caa16732-5783-4a87-99d1-b552dc3af955-scripts\") pod \"swift-ring-rebalance-debug-t4l6p\" (UID: \"caa16732-5783-4a87-99d1-b552dc3af955\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-t4l6p" Mar 07 15:15:17 crc kubenswrapper[4943]: I0307 15:15:17.005283 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/caa16732-5783-4a87-99d1-b552dc3af955-ring-data-devices\") pod \"swift-ring-rebalance-debug-t4l6p\" (UID: \"caa16732-5783-4a87-99d1-b552dc3af955\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-t4l6p" Mar 07 15:15:17 crc kubenswrapper[4943]: I0307 15:15:17.008757 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/caa16732-5783-4a87-99d1-b552dc3af955-swiftconf\") pod \"swift-ring-rebalance-debug-t4l6p\" (UID: \"caa16732-5783-4a87-99d1-b552dc3af955\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-t4l6p" Mar 07 15:15:17 crc kubenswrapper[4943]: I0307 15:15:17.010262 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/caa16732-5783-4a87-99d1-b552dc3af955-dispersionconf\") pod \"swift-ring-rebalance-debug-t4l6p\" (UID: \"caa16732-5783-4a87-99d1-b552dc3af955\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-t4l6p" Mar 07 15:15:17 crc kubenswrapper[4943]: I0307 15:15:17.037080 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77xm9\" (UniqueName: \"kubernetes.io/projected/caa16732-5783-4a87-99d1-b552dc3af955-kube-api-access-77xm9\") pod \"swift-ring-rebalance-debug-t4l6p\" (UID: \"caa16732-5783-4a87-99d1-b552dc3af955\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-t4l6p" Mar 07 15:15:17 crc kubenswrapper[4943]: I0307 15:15:17.067102 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-t4l6p" Mar 07 15:15:17 crc kubenswrapper[4943]: I0307 15:15:17.552035 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-t4l6p"] Mar 07 15:15:17 crc kubenswrapper[4943]: W0307 15:15:17.557290 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcaa16732_5783_4a87_99d1_b552dc3af955.slice/crio-577bc43311f5f9c29aa4d57eb4e419b8d6c1548ac9a4a45b8e0223caa2efb718 WatchSource:0}: Error finding container 577bc43311f5f9c29aa4d57eb4e419b8d6c1548ac9a4a45b8e0223caa2efb718: Status 404 returned error can't find the container with id 577bc43311f5f9c29aa4d57eb4e419b8d6c1548ac9a4a45b8e0223caa2efb718 Mar 07 15:15:18 crc kubenswrapper[4943]: I0307 15:15:18.152624 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-t4l6p" event={"ID":"caa16732-5783-4a87-99d1-b552dc3af955","Type":"ContainerStarted","Data":"78d29fc320c6cd71a267111b2e9277dde6d0068acb5d149f78356147aa0fc254"} Mar 07 15:15:18 crc kubenswrapper[4943]: I0307 15:15:18.152971 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-t4l6p" event={"ID":"caa16732-5783-4a87-99d1-b552dc3af955","Type":"ContainerStarted","Data":"577bc43311f5f9c29aa4d57eb4e419b8d6c1548ac9a4a45b8e0223caa2efb718"} Mar 07 15:15:18 crc kubenswrapper[4943]: I0307 15:15:18.181874 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-t4l6p" podStartSLOduration=2.181842032 podStartE2EDuration="2.181842032s" podCreationTimestamp="2026-03-07 15:15:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:15:18.174412851 +0000 UTC m=+2160.126549389" watchObservedRunningTime="2026-03-07 15:15:18.181842032 +0000 UTC m=+2160.133978570" Mar 07 15:15:19 crc kubenswrapper[4943]: I0307 15:15:19.167008 4943 generic.go:334] "Generic (PLEG): container finished" podID="caa16732-5783-4a87-99d1-b552dc3af955" containerID="78d29fc320c6cd71a267111b2e9277dde6d0068acb5d149f78356147aa0fc254" exitCode=0 Mar 07 15:15:19 crc kubenswrapper[4943]: I0307 15:15:19.167124 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-t4l6p" event={"ID":"caa16732-5783-4a87-99d1-b552dc3af955","Type":"ContainerDied","Data":"78d29fc320c6cd71a267111b2e9277dde6d0068acb5d149f78356147aa0fc254"} Mar 07 15:15:20 crc kubenswrapper[4943]: I0307 15:15:20.622099 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-t4l6p" Mar 07 15:15:20 crc kubenswrapper[4943]: I0307 15:15:20.656807 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-h57bg" Mar 07 15:15:20 crc kubenswrapper[4943]: I0307 15:15:20.656984 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-h57bg" Mar 07 15:15:20 crc kubenswrapper[4943]: I0307 15:15:20.666193 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-t4l6p"] Mar 07 15:15:20 crc kubenswrapper[4943]: I0307 15:15:20.673267 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-t4l6p"] Mar 07 15:15:20 crc kubenswrapper[4943]: I0307 15:15:20.719686 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-h57bg" Mar 07 15:15:20 crc kubenswrapper[4943]: I0307 15:15:20.771614 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/caa16732-5783-4a87-99d1-b552dc3af955-swiftconf\") pod \"caa16732-5783-4a87-99d1-b552dc3af955\" (UID: \"caa16732-5783-4a87-99d1-b552dc3af955\") " Mar 07 15:15:20 crc kubenswrapper[4943]: I0307 15:15:20.771686 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-77xm9\" (UniqueName: \"kubernetes.io/projected/caa16732-5783-4a87-99d1-b552dc3af955-kube-api-access-77xm9\") pod \"caa16732-5783-4a87-99d1-b552dc3af955\" (UID: \"caa16732-5783-4a87-99d1-b552dc3af955\") " Mar 07 15:15:20 crc kubenswrapper[4943]: I0307 15:15:20.771732 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/caa16732-5783-4a87-99d1-b552dc3af955-etc-swift\") pod \"caa16732-5783-4a87-99d1-b552dc3af955\" (UID: \"caa16732-5783-4a87-99d1-b552dc3af955\") " Mar 07 15:15:20 crc kubenswrapper[4943]: I0307 15:15:20.771836 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/caa16732-5783-4a87-99d1-b552dc3af955-scripts\") pod \"caa16732-5783-4a87-99d1-b552dc3af955\" (UID: \"caa16732-5783-4a87-99d1-b552dc3af955\") " Mar 07 15:15:20 crc kubenswrapper[4943]: I0307 15:15:20.771861 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/caa16732-5783-4a87-99d1-b552dc3af955-dispersionconf\") pod \"caa16732-5783-4a87-99d1-b552dc3af955\" (UID: \"caa16732-5783-4a87-99d1-b552dc3af955\") " Mar 07 15:15:20 crc kubenswrapper[4943]: I0307 15:15:20.771953 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/caa16732-5783-4a87-99d1-b552dc3af955-ring-data-devices\") pod \"caa16732-5783-4a87-99d1-b552dc3af955\" (UID: \"caa16732-5783-4a87-99d1-b552dc3af955\") " Mar 07 15:15:20 crc kubenswrapper[4943]: I0307 15:15:20.772763 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/caa16732-5783-4a87-99d1-b552dc3af955-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "caa16732-5783-4a87-99d1-b552dc3af955" (UID: "caa16732-5783-4a87-99d1-b552dc3af955"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:15:20 crc kubenswrapper[4943]: I0307 15:15:20.773053 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/caa16732-5783-4a87-99d1-b552dc3af955-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:20 crc kubenswrapper[4943]: I0307 15:15:20.773085 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/caa16732-5783-4a87-99d1-b552dc3af955-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "caa16732-5783-4a87-99d1-b552dc3af955" (UID: "caa16732-5783-4a87-99d1-b552dc3af955"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:15:20 crc kubenswrapper[4943]: I0307 15:15:20.782175 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/caa16732-5783-4a87-99d1-b552dc3af955-kube-api-access-77xm9" (OuterVolumeSpecName: "kube-api-access-77xm9") pod "caa16732-5783-4a87-99d1-b552dc3af955" (UID: "caa16732-5783-4a87-99d1-b552dc3af955"). InnerVolumeSpecName "kube-api-access-77xm9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:15:20 crc kubenswrapper[4943]: I0307 15:15:20.796141 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/caa16732-5783-4a87-99d1-b552dc3af955-scripts" (OuterVolumeSpecName: "scripts") pod "caa16732-5783-4a87-99d1-b552dc3af955" (UID: "caa16732-5783-4a87-99d1-b552dc3af955"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:15:20 crc kubenswrapper[4943]: I0307 15:15:20.797216 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/caa16732-5783-4a87-99d1-b552dc3af955-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "caa16732-5783-4a87-99d1-b552dc3af955" (UID: "caa16732-5783-4a87-99d1-b552dc3af955"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:15:20 crc kubenswrapper[4943]: I0307 15:15:20.814775 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/caa16732-5783-4a87-99d1-b552dc3af955-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "caa16732-5783-4a87-99d1-b552dc3af955" (UID: "caa16732-5783-4a87-99d1-b552dc3af955"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:15:20 crc kubenswrapper[4943]: I0307 15:15:20.874360 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/caa16732-5783-4a87-99d1-b552dc3af955-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:20 crc kubenswrapper[4943]: I0307 15:15:20.874391 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-77xm9\" (UniqueName: \"kubernetes.io/projected/caa16732-5783-4a87-99d1-b552dc3af955-kube-api-access-77xm9\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:20 crc kubenswrapper[4943]: I0307 15:15:20.874405 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/caa16732-5783-4a87-99d1-b552dc3af955-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:20 crc kubenswrapper[4943]: I0307 15:15:20.874417 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/caa16732-5783-4a87-99d1-b552dc3af955-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:20 crc kubenswrapper[4943]: I0307 15:15:20.874429 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/caa16732-5783-4a87-99d1-b552dc3af955-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:21 crc kubenswrapper[4943]: I0307 15:15:21.198208 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-t4l6p" Mar 07 15:15:21 crc kubenswrapper[4943]: I0307 15:15:21.198215 4943 scope.go:117] "RemoveContainer" containerID="78d29fc320c6cd71a267111b2e9277dde6d0068acb5d149f78356147aa0fc254" Mar 07 15:15:21 crc kubenswrapper[4943]: I0307 15:15:21.304040 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-h57bg" Mar 07 15:15:21 crc kubenswrapper[4943]: I0307 15:15:21.357191 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-h57bg"] Mar 07 15:15:21 crc kubenswrapper[4943]: I0307 15:15:21.783162 4943 scope.go:117] "RemoveContainer" containerID="77e2176d82c678c6306ab7d0f7ca6234ef2c1b3bce73ba85d718298ef1acdc57" Mar 07 15:15:21 crc kubenswrapper[4943]: I0307 15:15:21.862044 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-hqq4h"] Mar 07 15:15:21 crc kubenswrapper[4943]: E0307 15:15:21.862493 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="caa16732-5783-4a87-99d1-b552dc3af955" containerName="swift-ring-rebalance" Mar 07 15:15:21 crc kubenswrapper[4943]: I0307 15:15:21.862518 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="caa16732-5783-4a87-99d1-b552dc3af955" containerName="swift-ring-rebalance" Mar 07 15:15:21 crc kubenswrapper[4943]: I0307 15:15:21.862752 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="caa16732-5783-4a87-99d1-b552dc3af955" containerName="swift-ring-rebalance" Mar 07 15:15:21 crc kubenswrapper[4943]: I0307 15:15:21.863361 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-hqq4h" Mar 07 15:15:21 crc kubenswrapper[4943]: I0307 15:15:21.865576 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:15:21 crc kubenswrapper[4943]: I0307 15:15:21.865701 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:15:21 crc kubenswrapper[4943]: I0307 15:15:21.886265 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-hqq4h"] Mar 07 15:15:21 crc kubenswrapper[4943]: I0307 15:15:21.991239 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9fef7925-017b-44f1-ab35-8851ce94be06-etc-swift\") pod \"swift-ring-rebalance-debug-hqq4h\" (UID: \"9fef7925-017b-44f1-ab35-8851ce94be06\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hqq4h" Mar 07 15:15:21 crc kubenswrapper[4943]: I0307 15:15:21.991297 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qpdn\" (UniqueName: \"kubernetes.io/projected/9fef7925-017b-44f1-ab35-8851ce94be06-kube-api-access-8qpdn\") pod \"swift-ring-rebalance-debug-hqq4h\" (UID: \"9fef7925-017b-44f1-ab35-8851ce94be06\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hqq4h" Mar 07 15:15:21 crc kubenswrapper[4943]: I0307 15:15:21.991362 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9fef7925-017b-44f1-ab35-8851ce94be06-scripts\") pod \"swift-ring-rebalance-debug-hqq4h\" (UID: \"9fef7925-017b-44f1-ab35-8851ce94be06\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hqq4h" Mar 07 15:15:21 crc kubenswrapper[4943]: I0307 15:15:21.991390 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9fef7925-017b-44f1-ab35-8851ce94be06-swiftconf\") pod \"swift-ring-rebalance-debug-hqq4h\" (UID: \"9fef7925-017b-44f1-ab35-8851ce94be06\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hqq4h" Mar 07 15:15:21 crc kubenswrapper[4943]: I0307 15:15:21.991748 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9fef7925-017b-44f1-ab35-8851ce94be06-ring-data-devices\") pod \"swift-ring-rebalance-debug-hqq4h\" (UID: \"9fef7925-017b-44f1-ab35-8851ce94be06\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hqq4h" Mar 07 15:15:21 crc kubenswrapper[4943]: I0307 15:15:21.992020 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9fef7925-017b-44f1-ab35-8851ce94be06-dispersionconf\") pod \"swift-ring-rebalance-debug-hqq4h\" (UID: \"9fef7925-017b-44f1-ab35-8851ce94be06\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hqq4h" Mar 07 15:15:22 crc kubenswrapper[4943]: I0307 15:15:22.093976 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9fef7925-017b-44f1-ab35-8851ce94be06-scripts\") pod \"swift-ring-rebalance-debug-hqq4h\" (UID: \"9fef7925-017b-44f1-ab35-8851ce94be06\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hqq4h" Mar 07 15:15:22 crc kubenswrapper[4943]: I0307 15:15:22.094047 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9fef7925-017b-44f1-ab35-8851ce94be06-swiftconf\") pod \"swift-ring-rebalance-debug-hqq4h\" (UID: \"9fef7925-017b-44f1-ab35-8851ce94be06\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hqq4h" Mar 07 15:15:22 crc kubenswrapper[4943]: I0307 15:15:22.094110 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9fef7925-017b-44f1-ab35-8851ce94be06-ring-data-devices\") pod \"swift-ring-rebalance-debug-hqq4h\" (UID: \"9fef7925-017b-44f1-ab35-8851ce94be06\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hqq4h" Mar 07 15:15:22 crc kubenswrapper[4943]: I0307 15:15:22.094220 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9fef7925-017b-44f1-ab35-8851ce94be06-dispersionconf\") pod \"swift-ring-rebalance-debug-hqq4h\" (UID: \"9fef7925-017b-44f1-ab35-8851ce94be06\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hqq4h" Mar 07 15:15:22 crc kubenswrapper[4943]: I0307 15:15:22.094284 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9fef7925-017b-44f1-ab35-8851ce94be06-etc-swift\") pod \"swift-ring-rebalance-debug-hqq4h\" (UID: \"9fef7925-017b-44f1-ab35-8851ce94be06\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hqq4h" Mar 07 15:15:22 crc kubenswrapper[4943]: I0307 15:15:22.094320 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qpdn\" (UniqueName: \"kubernetes.io/projected/9fef7925-017b-44f1-ab35-8851ce94be06-kube-api-access-8qpdn\") pod \"swift-ring-rebalance-debug-hqq4h\" (UID: \"9fef7925-017b-44f1-ab35-8851ce94be06\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hqq4h" Mar 07 15:15:22 crc kubenswrapper[4943]: I0307 15:15:22.095235 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9fef7925-017b-44f1-ab35-8851ce94be06-etc-swift\") pod \"swift-ring-rebalance-debug-hqq4h\" (UID: \"9fef7925-017b-44f1-ab35-8851ce94be06\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hqq4h" Mar 07 15:15:22 crc kubenswrapper[4943]: I0307 15:15:22.095607 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9fef7925-017b-44f1-ab35-8851ce94be06-scripts\") pod \"swift-ring-rebalance-debug-hqq4h\" (UID: \"9fef7925-017b-44f1-ab35-8851ce94be06\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hqq4h" Mar 07 15:15:22 crc kubenswrapper[4943]: I0307 15:15:22.097519 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9fef7925-017b-44f1-ab35-8851ce94be06-ring-data-devices\") pod \"swift-ring-rebalance-debug-hqq4h\" (UID: \"9fef7925-017b-44f1-ab35-8851ce94be06\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hqq4h" Mar 07 15:15:22 crc kubenswrapper[4943]: I0307 15:15:22.102072 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9fef7925-017b-44f1-ab35-8851ce94be06-dispersionconf\") pod \"swift-ring-rebalance-debug-hqq4h\" (UID: \"9fef7925-017b-44f1-ab35-8851ce94be06\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hqq4h" Mar 07 15:15:22 crc kubenswrapper[4943]: I0307 15:15:22.109991 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9fef7925-017b-44f1-ab35-8851ce94be06-swiftconf\") pod \"swift-ring-rebalance-debug-hqq4h\" (UID: \"9fef7925-017b-44f1-ab35-8851ce94be06\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hqq4h" Mar 07 15:15:22 crc kubenswrapper[4943]: I0307 15:15:22.118251 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qpdn\" (UniqueName: \"kubernetes.io/projected/9fef7925-017b-44f1-ab35-8851ce94be06-kube-api-access-8qpdn\") pod \"swift-ring-rebalance-debug-hqq4h\" (UID: \"9fef7925-017b-44f1-ab35-8851ce94be06\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hqq4h" Mar 07 15:15:22 crc kubenswrapper[4943]: I0307 15:15:22.199638 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-hqq4h" Mar 07 15:15:22 crc kubenswrapper[4943]: I0307 15:15:22.482431 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-hqq4h"] Mar 07 15:15:22 crc kubenswrapper[4943]: W0307 15:15:22.491126 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9fef7925_017b_44f1_ab35_8851ce94be06.slice/crio-51f26bbc3ee0c783bbf86d1b170576fa8adacbe07aba4de70d12750a16187c41 WatchSource:0}: Error finding container 51f26bbc3ee0c783bbf86d1b170576fa8adacbe07aba4de70d12750a16187c41: Status 404 returned error can't find the container with id 51f26bbc3ee0c783bbf86d1b170576fa8adacbe07aba4de70d12750a16187c41 Mar 07 15:15:22 crc kubenswrapper[4943]: I0307 15:15:22.772345 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="caa16732-5783-4a87-99d1-b552dc3af955" path="/var/lib/kubelet/pods/caa16732-5783-4a87-99d1-b552dc3af955/volumes" Mar 07 15:15:23 crc kubenswrapper[4943]: I0307 15:15:23.225088 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-hqq4h" event={"ID":"9fef7925-017b-44f1-ab35-8851ce94be06","Type":"ContainerStarted","Data":"81ad0427c89f20cf442bd3e43a1c4ec1ee825f288babfc2bff62a6c5418ed361"} Mar 07 15:15:23 crc kubenswrapper[4943]: I0307 15:15:23.225160 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-hqq4h" event={"ID":"9fef7925-017b-44f1-ab35-8851ce94be06","Type":"ContainerStarted","Data":"51f26bbc3ee0c783bbf86d1b170576fa8adacbe07aba4de70d12750a16187c41"} Mar 07 15:15:23 crc kubenswrapper[4943]: I0307 15:15:23.225241 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-h57bg" podUID="253f1202-6997-48d2-8312-fda66447161c" containerName="registry-server" containerID="cri-o://a03458f2c0dfbd48b5cfc5edb914541f94cc5894e42d9641b2a83e09c7eedba7" gracePeriod=2 Mar 07 15:15:23 crc kubenswrapper[4943]: I0307 15:15:23.271884 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-hqq4h" podStartSLOduration=2.27185398 podStartE2EDuration="2.27185398s" podCreationTimestamp="2026-03-07 15:15:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:15:23.256730011 +0000 UTC m=+2165.208866569" watchObservedRunningTime="2026-03-07 15:15:23.27185398 +0000 UTC m=+2165.223990518" Mar 07 15:15:23 crc kubenswrapper[4943]: I0307 15:15:23.728383 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h57bg" Mar 07 15:15:23 crc kubenswrapper[4943]: I0307 15:15:23.823746 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/253f1202-6997-48d2-8312-fda66447161c-catalog-content\") pod \"253f1202-6997-48d2-8312-fda66447161c\" (UID: \"253f1202-6997-48d2-8312-fda66447161c\") " Mar 07 15:15:23 crc kubenswrapper[4943]: I0307 15:15:23.823851 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/253f1202-6997-48d2-8312-fda66447161c-utilities\") pod \"253f1202-6997-48d2-8312-fda66447161c\" (UID: \"253f1202-6997-48d2-8312-fda66447161c\") " Mar 07 15:15:23 crc kubenswrapper[4943]: I0307 15:15:23.823874 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2bvs2\" (UniqueName: \"kubernetes.io/projected/253f1202-6997-48d2-8312-fda66447161c-kube-api-access-2bvs2\") pod \"253f1202-6997-48d2-8312-fda66447161c\" (UID: \"253f1202-6997-48d2-8312-fda66447161c\") " Mar 07 15:15:23 crc kubenswrapper[4943]: I0307 15:15:23.825283 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/253f1202-6997-48d2-8312-fda66447161c-utilities" (OuterVolumeSpecName: "utilities") pod "253f1202-6997-48d2-8312-fda66447161c" (UID: "253f1202-6997-48d2-8312-fda66447161c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:15:23 crc kubenswrapper[4943]: I0307 15:15:23.835329 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/253f1202-6997-48d2-8312-fda66447161c-kube-api-access-2bvs2" (OuterVolumeSpecName: "kube-api-access-2bvs2") pod "253f1202-6997-48d2-8312-fda66447161c" (UID: "253f1202-6997-48d2-8312-fda66447161c"). InnerVolumeSpecName "kube-api-access-2bvs2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:15:23 crc kubenswrapper[4943]: I0307 15:15:23.867226 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/253f1202-6997-48d2-8312-fda66447161c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "253f1202-6997-48d2-8312-fda66447161c" (UID: "253f1202-6997-48d2-8312-fda66447161c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:15:23 crc kubenswrapper[4943]: I0307 15:15:23.925583 4943 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/253f1202-6997-48d2-8312-fda66447161c-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:23 crc kubenswrapper[4943]: I0307 15:15:23.925630 4943 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/253f1202-6997-48d2-8312-fda66447161c-utilities\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:23 crc kubenswrapper[4943]: I0307 15:15:23.925649 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2bvs2\" (UniqueName: \"kubernetes.io/projected/253f1202-6997-48d2-8312-fda66447161c-kube-api-access-2bvs2\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:24 crc kubenswrapper[4943]: I0307 15:15:24.241046 4943 generic.go:334] "Generic (PLEG): container finished" podID="253f1202-6997-48d2-8312-fda66447161c" containerID="a03458f2c0dfbd48b5cfc5edb914541f94cc5894e42d9641b2a83e09c7eedba7" exitCode=0 Mar 07 15:15:24 crc kubenswrapper[4943]: I0307 15:15:24.241151 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h57bg" event={"ID":"253f1202-6997-48d2-8312-fda66447161c","Type":"ContainerDied","Data":"a03458f2c0dfbd48b5cfc5edb914541f94cc5894e42d9641b2a83e09c7eedba7"} Mar 07 15:15:24 crc kubenswrapper[4943]: I0307 15:15:24.241158 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h57bg" Mar 07 15:15:24 crc kubenswrapper[4943]: I0307 15:15:24.241195 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h57bg" event={"ID":"253f1202-6997-48d2-8312-fda66447161c","Type":"ContainerDied","Data":"c7a7d8c9d2e9b754958c2e4c8929162e3ec6caab050ade1374010034f37bc4ee"} Mar 07 15:15:24 crc kubenswrapper[4943]: I0307 15:15:24.241227 4943 scope.go:117] "RemoveContainer" containerID="a03458f2c0dfbd48b5cfc5edb914541f94cc5894e42d9641b2a83e09c7eedba7" Mar 07 15:15:24 crc kubenswrapper[4943]: I0307 15:15:24.246248 4943 generic.go:334] "Generic (PLEG): container finished" podID="9fef7925-017b-44f1-ab35-8851ce94be06" containerID="81ad0427c89f20cf442bd3e43a1c4ec1ee825f288babfc2bff62a6c5418ed361" exitCode=0 Mar 07 15:15:24 crc kubenswrapper[4943]: I0307 15:15:24.246310 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-hqq4h" event={"ID":"9fef7925-017b-44f1-ab35-8851ce94be06","Type":"ContainerDied","Data":"81ad0427c89f20cf442bd3e43a1c4ec1ee825f288babfc2bff62a6c5418ed361"} Mar 07 15:15:24 crc kubenswrapper[4943]: I0307 15:15:24.304138 4943 scope.go:117] "RemoveContainer" containerID="df6cc34f1b0db7d8086f552fd9403c5ec9657059855956c620723d8b9ce7ff31" Mar 07 15:15:24 crc kubenswrapper[4943]: I0307 15:15:24.327890 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-h57bg"] Mar 07 15:15:24 crc kubenswrapper[4943]: I0307 15:15:24.335542 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-h57bg"] Mar 07 15:15:24 crc kubenswrapper[4943]: I0307 15:15:24.336417 4943 scope.go:117] "RemoveContainer" containerID="9c6128ca0ec73e2c85a0c9ac1e24ab2f61c62b4123e544345f688df9c66d76d0" Mar 07 15:15:24 crc kubenswrapper[4943]: I0307 15:15:24.381778 4943 scope.go:117] "RemoveContainer" containerID="a03458f2c0dfbd48b5cfc5edb914541f94cc5894e42d9641b2a83e09c7eedba7" Mar 07 15:15:24 crc kubenswrapper[4943]: E0307 15:15:24.382412 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a03458f2c0dfbd48b5cfc5edb914541f94cc5894e42d9641b2a83e09c7eedba7\": container with ID starting with a03458f2c0dfbd48b5cfc5edb914541f94cc5894e42d9641b2a83e09c7eedba7 not found: ID does not exist" containerID="a03458f2c0dfbd48b5cfc5edb914541f94cc5894e42d9641b2a83e09c7eedba7" Mar 07 15:15:24 crc kubenswrapper[4943]: I0307 15:15:24.382468 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a03458f2c0dfbd48b5cfc5edb914541f94cc5894e42d9641b2a83e09c7eedba7"} err="failed to get container status \"a03458f2c0dfbd48b5cfc5edb914541f94cc5894e42d9641b2a83e09c7eedba7\": rpc error: code = NotFound desc = could not find container \"a03458f2c0dfbd48b5cfc5edb914541f94cc5894e42d9641b2a83e09c7eedba7\": container with ID starting with a03458f2c0dfbd48b5cfc5edb914541f94cc5894e42d9641b2a83e09c7eedba7 not found: ID does not exist" Mar 07 15:15:24 crc kubenswrapper[4943]: I0307 15:15:24.382503 4943 scope.go:117] "RemoveContainer" containerID="df6cc34f1b0db7d8086f552fd9403c5ec9657059855956c620723d8b9ce7ff31" Mar 07 15:15:24 crc kubenswrapper[4943]: E0307 15:15:24.382877 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df6cc34f1b0db7d8086f552fd9403c5ec9657059855956c620723d8b9ce7ff31\": container with ID starting with df6cc34f1b0db7d8086f552fd9403c5ec9657059855956c620723d8b9ce7ff31 not found: ID does not exist" containerID="df6cc34f1b0db7d8086f552fd9403c5ec9657059855956c620723d8b9ce7ff31" Mar 07 15:15:24 crc kubenswrapper[4943]: I0307 15:15:24.382923 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df6cc34f1b0db7d8086f552fd9403c5ec9657059855956c620723d8b9ce7ff31"} err="failed to get container status \"df6cc34f1b0db7d8086f552fd9403c5ec9657059855956c620723d8b9ce7ff31\": rpc error: code = NotFound desc = could not find container \"df6cc34f1b0db7d8086f552fd9403c5ec9657059855956c620723d8b9ce7ff31\": container with ID starting with df6cc34f1b0db7d8086f552fd9403c5ec9657059855956c620723d8b9ce7ff31 not found: ID does not exist" Mar 07 15:15:24 crc kubenswrapper[4943]: I0307 15:15:24.382980 4943 scope.go:117] "RemoveContainer" containerID="9c6128ca0ec73e2c85a0c9ac1e24ab2f61c62b4123e544345f688df9c66d76d0" Mar 07 15:15:24 crc kubenswrapper[4943]: E0307 15:15:24.383426 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c6128ca0ec73e2c85a0c9ac1e24ab2f61c62b4123e544345f688df9c66d76d0\": container with ID starting with 9c6128ca0ec73e2c85a0c9ac1e24ab2f61c62b4123e544345f688df9c66d76d0 not found: ID does not exist" containerID="9c6128ca0ec73e2c85a0c9ac1e24ab2f61c62b4123e544345f688df9c66d76d0" Mar 07 15:15:24 crc kubenswrapper[4943]: I0307 15:15:24.383498 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c6128ca0ec73e2c85a0c9ac1e24ab2f61c62b4123e544345f688df9c66d76d0"} err="failed to get container status \"9c6128ca0ec73e2c85a0c9ac1e24ab2f61c62b4123e544345f688df9c66d76d0\": rpc error: code = NotFound desc = could not find container \"9c6128ca0ec73e2c85a0c9ac1e24ab2f61c62b4123e544345f688df9c66d76d0\": container with ID starting with 9c6128ca0ec73e2c85a0c9ac1e24ab2f61c62b4123e544345f688df9c66d76d0 not found: ID does not exist" Mar 07 15:15:24 crc kubenswrapper[4943]: I0307 15:15:24.764999 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="253f1202-6997-48d2-8312-fda66447161c" path="/var/lib/kubelet/pods/253f1202-6997-48d2-8312-fda66447161c/volumes" Mar 07 15:15:25 crc kubenswrapper[4943]: I0307 15:15:25.722203 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-hqq4h" Mar 07 15:15:25 crc kubenswrapper[4943]: I0307 15:15:25.766437 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-hqq4h"] Mar 07 15:15:25 crc kubenswrapper[4943]: I0307 15:15:25.772033 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-hqq4h"] Mar 07 15:15:25 crc kubenswrapper[4943]: I0307 15:15:25.855998 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9fef7925-017b-44f1-ab35-8851ce94be06-etc-swift\") pod \"9fef7925-017b-44f1-ab35-8851ce94be06\" (UID: \"9fef7925-017b-44f1-ab35-8851ce94be06\") " Mar 07 15:15:25 crc kubenswrapper[4943]: I0307 15:15:25.856081 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9fef7925-017b-44f1-ab35-8851ce94be06-ring-data-devices\") pod \"9fef7925-017b-44f1-ab35-8851ce94be06\" (UID: \"9fef7925-017b-44f1-ab35-8851ce94be06\") " Mar 07 15:15:25 crc kubenswrapper[4943]: I0307 15:15:25.856126 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9fef7925-017b-44f1-ab35-8851ce94be06-dispersionconf\") pod \"9fef7925-017b-44f1-ab35-8851ce94be06\" (UID: \"9fef7925-017b-44f1-ab35-8851ce94be06\") " Mar 07 15:15:25 crc kubenswrapper[4943]: I0307 15:15:25.856999 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9fef7925-017b-44f1-ab35-8851ce94be06-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "9fef7925-017b-44f1-ab35-8851ce94be06" (UID: "9fef7925-017b-44f1-ab35-8851ce94be06"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:15:25 crc kubenswrapper[4943]: I0307 15:15:25.857164 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9fef7925-017b-44f1-ab35-8851ce94be06-scripts\") pod \"9fef7925-017b-44f1-ab35-8851ce94be06\" (UID: \"9fef7925-017b-44f1-ab35-8851ce94be06\") " Mar 07 15:15:25 crc kubenswrapper[4943]: I0307 15:15:25.857226 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9fef7925-017b-44f1-ab35-8851ce94be06-swiftconf\") pod \"9fef7925-017b-44f1-ab35-8851ce94be06\" (UID: \"9fef7925-017b-44f1-ab35-8851ce94be06\") " Mar 07 15:15:25 crc kubenswrapper[4943]: I0307 15:15:25.857262 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8qpdn\" (UniqueName: \"kubernetes.io/projected/9fef7925-017b-44f1-ab35-8851ce94be06-kube-api-access-8qpdn\") pod \"9fef7925-017b-44f1-ab35-8851ce94be06\" (UID: \"9fef7925-017b-44f1-ab35-8851ce94be06\") " Mar 07 15:15:25 crc kubenswrapper[4943]: I0307 15:15:25.857342 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9fef7925-017b-44f1-ab35-8851ce94be06-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "9fef7925-017b-44f1-ab35-8851ce94be06" (UID: "9fef7925-017b-44f1-ab35-8851ce94be06"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:15:25 crc kubenswrapper[4943]: I0307 15:15:25.857690 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9fef7925-017b-44f1-ab35-8851ce94be06-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:25 crc kubenswrapper[4943]: I0307 15:15:25.857724 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9fef7925-017b-44f1-ab35-8851ce94be06-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:25 crc kubenswrapper[4943]: I0307 15:15:25.865261 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9fef7925-017b-44f1-ab35-8851ce94be06-kube-api-access-8qpdn" (OuterVolumeSpecName: "kube-api-access-8qpdn") pod "9fef7925-017b-44f1-ab35-8851ce94be06" (UID: "9fef7925-017b-44f1-ab35-8851ce94be06"). InnerVolumeSpecName "kube-api-access-8qpdn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:15:25 crc kubenswrapper[4943]: I0307 15:15:25.886420 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fef7925-017b-44f1-ab35-8851ce94be06-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "9fef7925-017b-44f1-ab35-8851ce94be06" (UID: "9fef7925-017b-44f1-ab35-8851ce94be06"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:15:25 crc kubenswrapper[4943]: I0307 15:15:25.891506 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9fef7925-017b-44f1-ab35-8851ce94be06-scripts" (OuterVolumeSpecName: "scripts") pod "9fef7925-017b-44f1-ab35-8851ce94be06" (UID: "9fef7925-017b-44f1-ab35-8851ce94be06"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:15:25 crc kubenswrapper[4943]: I0307 15:15:25.894324 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fef7925-017b-44f1-ab35-8851ce94be06-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "9fef7925-017b-44f1-ab35-8851ce94be06" (UID: "9fef7925-017b-44f1-ab35-8851ce94be06"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:15:25 crc kubenswrapper[4943]: I0307 15:15:25.958875 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9fef7925-017b-44f1-ab35-8851ce94be06-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:25 crc kubenswrapper[4943]: I0307 15:15:25.958919 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9fef7925-017b-44f1-ab35-8851ce94be06-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:25 crc kubenswrapper[4943]: I0307 15:15:25.958994 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8qpdn\" (UniqueName: \"kubernetes.io/projected/9fef7925-017b-44f1-ab35-8851ce94be06-kube-api-access-8qpdn\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:25 crc kubenswrapper[4943]: I0307 15:15:25.959023 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9fef7925-017b-44f1-ab35-8851ce94be06-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:26 crc kubenswrapper[4943]: I0307 15:15:26.277754 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="51f26bbc3ee0c783bbf86d1b170576fa8adacbe07aba4de70d12750a16187c41" Mar 07 15:15:26 crc kubenswrapper[4943]: I0307 15:15:26.277867 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-hqq4h" Mar 07 15:15:26 crc kubenswrapper[4943]: I0307 15:15:26.771055 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9fef7925-017b-44f1-ab35-8851ce94be06" path="/var/lib/kubelet/pods/9fef7925-017b-44f1-ab35-8851ce94be06/volumes" Mar 07 15:15:26 crc kubenswrapper[4943]: I0307 15:15:26.989718 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-qhmkk"] Mar 07 15:15:26 crc kubenswrapper[4943]: E0307 15:15:26.990889 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="253f1202-6997-48d2-8312-fda66447161c" containerName="registry-server" Mar 07 15:15:26 crc kubenswrapper[4943]: I0307 15:15:26.990955 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="253f1202-6997-48d2-8312-fda66447161c" containerName="registry-server" Mar 07 15:15:26 crc kubenswrapper[4943]: E0307 15:15:26.991059 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="253f1202-6997-48d2-8312-fda66447161c" containerName="extract-utilities" Mar 07 15:15:26 crc kubenswrapper[4943]: I0307 15:15:26.991078 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="253f1202-6997-48d2-8312-fda66447161c" containerName="extract-utilities" Mar 07 15:15:26 crc kubenswrapper[4943]: E0307 15:15:26.991114 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="253f1202-6997-48d2-8312-fda66447161c" containerName="extract-content" Mar 07 15:15:26 crc kubenswrapper[4943]: I0307 15:15:26.991130 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="253f1202-6997-48d2-8312-fda66447161c" containerName="extract-content" Mar 07 15:15:26 crc kubenswrapper[4943]: E0307 15:15:26.991145 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fef7925-017b-44f1-ab35-8851ce94be06" containerName="swift-ring-rebalance" Mar 07 15:15:26 crc kubenswrapper[4943]: I0307 15:15:26.991157 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fef7925-017b-44f1-ab35-8851ce94be06" containerName="swift-ring-rebalance" Mar 07 15:15:26 crc kubenswrapper[4943]: I0307 15:15:26.991448 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fef7925-017b-44f1-ab35-8851ce94be06" containerName="swift-ring-rebalance" Mar 07 15:15:26 crc kubenswrapper[4943]: I0307 15:15:26.991493 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="253f1202-6997-48d2-8312-fda66447161c" containerName="registry-server" Mar 07 15:15:26 crc kubenswrapper[4943]: I0307 15:15:26.992716 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qhmkk" Mar 07 15:15:26 crc kubenswrapper[4943]: I0307 15:15:26.996202 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:15:26 crc kubenswrapper[4943]: I0307 15:15:26.999249 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:15:27 crc kubenswrapper[4943]: I0307 15:15:27.007073 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-qhmkk"] Mar 07 15:15:27 crc kubenswrapper[4943]: I0307 15:15:27.075389 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ad36e69d-d4e6-4c02-a340-1dfed75cf45b-scripts\") pod \"swift-ring-rebalance-debug-qhmkk\" (UID: \"ad36e69d-d4e6-4c02-a340-1dfed75cf45b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qhmkk" Mar 07 15:15:27 crc kubenswrapper[4943]: I0307 15:15:27.075722 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ad36e69d-d4e6-4c02-a340-1dfed75cf45b-etc-swift\") pod \"swift-ring-rebalance-debug-qhmkk\" (UID: \"ad36e69d-d4e6-4c02-a340-1dfed75cf45b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qhmkk" Mar 07 15:15:27 crc kubenswrapper[4943]: I0307 15:15:27.075750 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ad36e69d-d4e6-4c02-a340-1dfed75cf45b-dispersionconf\") pod \"swift-ring-rebalance-debug-qhmkk\" (UID: \"ad36e69d-d4e6-4c02-a340-1dfed75cf45b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qhmkk" Mar 07 15:15:27 crc kubenswrapper[4943]: I0307 15:15:27.075778 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rv2z\" (UniqueName: \"kubernetes.io/projected/ad36e69d-d4e6-4c02-a340-1dfed75cf45b-kube-api-access-8rv2z\") pod \"swift-ring-rebalance-debug-qhmkk\" (UID: \"ad36e69d-d4e6-4c02-a340-1dfed75cf45b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qhmkk" Mar 07 15:15:27 crc kubenswrapper[4943]: I0307 15:15:27.075803 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ad36e69d-d4e6-4c02-a340-1dfed75cf45b-swiftconf\") pod \"swift-ring-rebalance-debug-qhmkk\" (UID: \"ad36e69d-d4e6-4c02-a340-1dfed75cf45b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qhmkk" Mar 07 15:15:27 crc kubenswrapper[4943]: I0307 15:15:27.076025 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ad36e69d-d4e6-4c02-a340-1dfed75cf45b-ring-data-devices\") pod \"swift-ring-rebalance-debug-qhmkk\" (UID: \"ad36e69d-d4e6-4c02-a340-1dfed75cf45b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qhmkk" Mar 07 15:15:27 crc kubenswrapper[4943]: I0307 15:15:27.178152 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ad36e69d-d4e6-4c02-a340-1dfed75cf45b-scripts\") pod \"swift-ring-rebalance-debug-qhmkk\" (UID: \"ad36e69d-d4e6-4c02-a340-1dfed75cf45b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qhmkk" Mar 07 15:15:27 crc kubenswrapper[4943]: I0307 15:15:27.178226 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ad36e69d-d4e6-4c02-a340-1dfed75cf45b-etc-swift\") pod \"swift-ring-rebalance-debug-qhmkk\" (UID: \"ad36e69d-d4e6-4c02-a340-1dfed75cf45b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qhmkk" Mar 07 15:15:27 crc kubenswrapper[4943]: I0307 15:15:27.178266 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ad36e69d-d4e6-4c02-a340-1dfed75cf45b-dispersionconf\") pod \"swift-ring-rebalance-debug-qhmkk\" (UID: \"ad36e69d-d4e6-4c02-a340-1dfed75cf45b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qhmkk" Mar 07 15:15:27 crc kubenswrapper[4943]: I0307 15:15:27.178300 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rv2z\" (UniqueName: \"kubernetes.io/projected/ad36e69d-d4e6-4c02-a340-1dfed75cf45b-kube-api-access-8rv2z\") pod \"swift-ring-rebalance-debug-qhmkk\" (UID: \"ad36e69d-d4e6-4c02-a340-1dfed75cf45b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qhmkk" Mar 07 15:15:27 crc kubenswrapper[4943]: I0307 15:15:27.178344 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ad36e69d-d4e6-4c02-a340-1dfed75cf45b-swiftconf\") pod \"swift-ring-rebalance-debug-qhmkk\" (UID: \"ad36e69d-d4e6-4c02-a340-1dfed75cf45b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qhmkk" Mar 07 15:15:27 crc kubenswrapper[4943]: I0307 15:15:27.178423 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ad36e69d-d4e6-4c02-a340-1dfed75cf45b-ring-data-devices\") pod \"swift-ring-rebalance-debug-qhmkk\" (UID: \"ad36e69d-d4e6-4c02-a340-1dfed75cf45b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qhmkk" Mar 07 15:15:27 crc kubenswrapper[4943]: I0307 15:15:27.179559 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ad36e69d-d4e6-4c02-a340-1dfed75cf45b-scripts\") pod \"swift-ring-rebalance-debug-qhmkk\" (UID: \"ad36e69d-d4e6-4c02-a340-1dfed75cf45b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qhmkk" Mar 07 15:15:27 crc kubenswrapper[4943]: I0307 15:15:27.179642 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ad36e69d-d4e6-4c02-a340-1dfed75cf45b-ring-data-devices\") pod \"swift-ring-rebalance-debug-qhmkk\" (UID: \"ad36e69d-d4e6-4c02-a340-1dfed75cf45b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qhmkk" Mar 07 15:15:27 crc kubenswrapper[4943]: I0307 15:15:27.179825 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ad36e69d-d4e6-4c02-a340-1dfed75cf45b-etc-swift\") pod \"swift-ring-rebalance-debug-qhmkk\" (UID: \"ad36e69d-d4e6-4c02-a340-1dfed75cf45b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qhmkk" Mar 07 15:15:27 crc kubenswrapper[4943]: I0307 15:15:27.183157 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ad36e69d-d4e6-4c02-a340-1dfed75cf45b-swiftconf\") pod \"swift-ring-rebalance-debug-qhmkk\" (UID: \"ad36e69d-d4e6-4c02-a340-1dfed75cf45b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qhmkk" Mar 07 15:15:27 crc kubenswrapper[4943]: I0307 15:15:27.184079 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ad36e69d-d4e6-4c02-a340-1dfed75cf45b-dispersionconf\") pod \"swift-ring-rebalance-debug-qhmkk\" (UID: \"ad36e69d-d4e6-4c02-a340-1dfed75cf45b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qhmkk" Mar 07 15:15:27 crc kubenswrapper[4943]: I0307 15:15:27.209861 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rv2z\" (UniqueName: \"kubernetes.io/projected/ad36e69d-d4e6-4c02-a340-1dfed75cf45b-kube-api-access-8rv2z\") pod \"swift-ring-rebalance-debug-qhmkk\" (UID: \"ad36e69d-d4e6-4c02-a340-1dfed75cf45b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qhmkk" Mar 07 15:15:27 crc kubenswrapper[4943]: I0307 15:15:27.322590 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qhmkk" Mar 07 15:15:27 crc kubenswrapper[4943]: I0307 15:15:27.588516 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-qhmkk"] Mar 07 15:15:27 crc kubenswrapper[4943]: W0307 15:15:27.592162 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podad36e69d_d4e6_4c02_a340_1dfed75cf45b.slice/crio-b16f6a707f981962c7ab04acbbf8de6b8356b39821d9646deecf407887a6bcff WatchSource:0}: Error finding container b16f6a707f981962c7ab04acbbf8de6b8356b39821d9646deecf407887a6bcff: Status 404 returned error can't find the container with id b16f6a707f981962c7ab04acbbf8de6b8356b39821d9646deecf407887a6bcff Mar 07 15:15:28 crc kubenswrapper[4943]: I0307 15:15:28.303984 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qhmkk" event={"ID":"ad36e69d-d4e6-4c02-a340-1dfed75cf45b","Type":"ContainerStarted","Data":"725afe0d98faa0ad3972dd436d3ce97ef2226e2243b7710cda821e041b293b62"} Mar 07 15:15:28 crc kubenswrapper[4943]: I0307 15:15:28.304288 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qhmkk" event={"ID":"ad36e69d-d4e6-4c02-a340-1dfed75cf45b","Type":"ContainerStarted","Data":"b16f6a707f981962c7ab04acbbf8de6b8356b39821d9646deecf407887a6bcff"} Mar 07 15:15:28 crc kubenswrapper[4943]: I0307 15:15:28.352477 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qhmkk" podStartSLOduration=2.352450569 podStartE2EDuration="2.352450569s" podCreationTimestamp="2026-03-07 15:15:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:15:28.338903748 +0000 UTC m=+2170.291040286" watchObservedRunningTime="2026-03-07 15:15:28.352450569 +0000 UTC m=+2170.304587097" Mar 07 15:15:29 crc kubenswrapper[4943]: I0307 15:15:29.320263 4943 generic.go:334] "Generic (PLEG): container finished" podID="ad36e69d-d4e6-4c02-a340-1dfed75cf45b" containerID="725afe0d98faa0ad3972dd436d3ce97ef2226e2243b7710cda821e041b293b62" exitCode=0 Mar 07 15:15:29 crc kubenswrapper[4943]: I0307 15:15:29.320331 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qhmkk" event={"ID":"ad36e69d-d4e6-4c02-a340-1dfed75cf45b","Type":"ContainerDied","Data":"725afe0d98faa0ad3972dd436d3ce97ef2226e2243b7710cda821e041b293b62"} Mar 07 15:15:30 crc kubenswrapper[4943]: I0307 15:15:30.735581 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qhmkk" Mar 07 15:15:30 crc kubenswrapper[4943]: I0307 15:15:30.774970 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-qhmkk"] Mar 07 15:15:30 crc kubenswrapper[4943]: I0307 15:15:30.780989 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-qhmkk"] Mar 07 15:15:30 crc kubenswrapper[4943]: I0307 15:15:30.836028 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ad36e69d-d4e6-4c02-a340-1dfed75cf45b-swiftconf\") pod \"ad36e69d-d4e6-4c02-a340-1dfed75cf45b\" (UID: \"ad36e69d-d4e6-4c02-a340-1dfed75cf45b\") " Mar 07 15:15:30 crc kubenswrapper[4943]: I0307 15:15:30.836188 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8rv2z\" (UniqueName: \"kubernetes.io/projected/ad36e69d-d4e6-4c02-a340-1dfed75cf45b-kube-api-access-8rv2z\") pod \"ad36e69d-d4e6-4c02-a340-1dfed75cf45b\" (UID: \"ad36e69d-d4e6-4c02-a340-1dfed75cf45b\") " Mar 07 15:15:30 crc kubenswrapper[4943]: I0307 15:15:30.836231 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ad36e69d-d4e6-4c02-a340-1dfed75cf45b-ring-data-devices\") pod \"ad36e69d-d4e6-4c02-a340-1dfed75cf45b\" (UID: \"ad36e69d-d4e6-4c02-a340-1dfed75cf45b\") " Mar 07 15:15:30 crc kubenswrapper[4943]: I0307 15:15:30.836251 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ad36e69d-d4e6-4c02-a340-1dfed75cf45b-dispersionconf\") pod \"ad36e69d-d4e6-4c02-a340-1dfed75cf45b\" (UID: \"ad36e69d-d4e6-4c02-a340-1dfed75cf45b\") " Mar 07 15:15:30 crc kubenswrapper[4943]: I0307 15:15:30.836323 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ad36e69d-d4e6-4c02-a340-1dfed75cf45b-etc-swift\") pod \"ad36e69d-d4e6-4c02-a340-1dfed75cf45b\" (UID: \"ad36e69d-d4e6-4c02-a340-1dfed75cf45b\") " Mar 07 15:15:30 crc kubenswrapper[4943]: I0307 15:15:30.836343 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ad36e69d-d4e6-4c02-a340-1dfed75cf45b-scripts\") pod \"ad36e69d-d4e6-4c02-a340-1dfed75cf45b\" (UID: \"ad36e69d-d4e6-4c02-a340-1dfed75cf45b\") " Mar 07 15:15:30 crc kubenswrapper[4943]: I0307 15:15:30.837137 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad36e69d-d4e6-4c02-a340-1dfed75cf45b-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "ad36e69d-d4e6-4c02-a340-1dfed75cf45b" (UID: "ad36e69d-d4e6-4c02-a340-1dfed75cf45b"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:15:30 crc kubenswrapper[4943]: I0307 15:15:30.837548 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad36e69d-d4e6-4c02-a340-1dfed75cf45b-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "ad36e69d-d4e6-4c02-a340-1dfed75cf45b" (UID: "ad36e69d-d4e6-4c02-a340-1dfed75cf45b"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:15:30 crc kubenswrapper[4943]: I0307 15:15:30.841894 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad36e69d-d4e6-4c02-a340-1dfed75cf45b-kube-api-access-8rv2z" (OuterVolumeSpecName: "kube-api-access-8rv2z") pod "ad36e69d-d4e6-4c02-a340-1dfed75cf45b" (UID: "ad36e69d-d4e6-4c02-a340-1dfed75cf45b"). InnerVolumeSpecName "kube-api-access-8rv2z". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:15:30 crc kubenswrapper[4943]: I0307 15:15:30.856431 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad36e69d-d4e6-4c02-a340-1dfed75cf45b-scripts" (OuterVolumeSpecName: "scripts") pod "ad36e69d-d4e6-4c02-a340-1dfed75cf45b" (UID: "ad36e69d-d4e6-4c02-a340-1dfed75cf45b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:15:30 crc kubenswrapper[4943]: I0307 15:15:30.859187 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad36e69d-d4e6-4c02-a340-1dfed75cf45b-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "ad36e69d-d4e6-4c02-a340-1dfed75cf45b" (UID: "ad36e69d-d4e6-4c02-a340-1dfed75cf45b"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:15:30 crc kubenswrapper[4943]: I0307 15:15:30.866452 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad36e69d-d4e6-4c02-a340-1dfed75cf45b-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "ad36e69d-d4e6-4c02-a340-1dfed75cf45b" (UID: "ad36e69d-d4e6-4c02-a340-1dfed75cf45b"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:15:30 crc kubenswrapper[4943]: I0307 15:15:30.937876 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8rv2z\" (UniqueName: \"kubernetes.io/projected/ad36e69d-d4e6-4c02-a340-1dfed75cf45b-kube-api-access-8rv2z\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:30 crc kubenswrapper[4943]: I0307 15:15:30.938121 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ad36e69d-d4e6-4c02-a340-1dfed75cf45b-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:30 crc kubenswrapper[4943]: I0307 15:15:30.938134 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ad36e69d-d4e6-4c02-a340-1dfed75cf45b-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:30 crc kubenswrapper[4943]: I0307 15:15:30.938146 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ad36e69d-d4e6-4c02-a340-1dfed75cf45b-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:30 crc kubenswrapper[4943]: I0307 15:15:30.938157 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ad36e69d-d4e6-4c02-a340-1dfed75cf45b-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:30 crc kubenswrapper[4943]: I0307 15:15:30.938169 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ad36e69d-d4e6-4c02-a340-1dfed75cf45b-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:31 crc kubenswrapper[4943]: I0307 15:15:31.341722 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b16f6a707f981962c7ab04acbbf8de6b8356b39821d9646deecf407887a6bcff" Mar 07 15:15:31 crc kubenswrapper[4943]: I0307 15:15:31.341821 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qhmkk" Mar 07 15:15:31 crc kubenswrapper[4943]: I0307 15:15:31.938938 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-m2zzc"] Mar 07 15:15:31 crc kubenswrapper[4943]: E0307 15:15:31.939231 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad36e69d-d4e6-4c02-a340-1dfed75cf45b" containerName="swift-ring-rebalance" Mar 07 15:15:31 crc kubenswrapper[4943]: I0307 15:15:31.939245 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad36e69d-d4e6-4c02-a340-1dfed75cf45b" containerName="swift-ring-rebalance" Mar 07 15:15:31 crc kubenswrapper[4943]: I0307 15:15:31.939385 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad36e69d-d4e6-4c02-a340-1dfed75cf45b" containerName="swift-ring-rebalance" Mar 07 15:15:31 crc kubenswrapper[4943]: I0307 15:15:31.939840 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-m2zzc" Mar 07 15:15:31 crc kubenswrapper[4943]: I0307 15:15:31.946447 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:15:31 crc kubenswrapper[4943]: I0307 15:15:31.947143 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:15:31 crc kubenswrapper[4943]: I0307 15:15:31.959359 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-m2zzc"] Mar 07 15:15:32 crc kubenswrapper[4943]: I0307 15:15:32.055965 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8775d99d-e8a9-42c3-97b2-bafad3306958-scripts\") pod \"swift-ring-rebalance-debug-m2zzc\" (UID: \"8775d99d-e8a9-42c3-97b2-bafad3306958\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m2zzc" Mar 07 15:15:32 crc kubenswrapper[4943]: I0307 15:15:32.056042 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/8775d99d-e8a9-42c3-97b2-bafad3306958-dispersionconf\") pod \"swift-ring-rebalance-debug-m2zzc\" (UID: \"8775d99d-e8a9-42c3-97b2-bafad3306958\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m2zzc" Mar 07 15:15:32 crc kubenswrapper[4943]: I0307 15:15:32.056114 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/8775d99d-e8a9-42c3-97b2-bafad3306958-etc-swift\") pod \"swift-ring-rebalance-debug-m2zzc\" (UID: \"8775d99d-e8a9-42c3-97b2-bafad3306958\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m2zzc" Mar 07 15:15:32 crc kubenswrapper[4943]: I0307 15:15:32.056141 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/8775d99d-e8a9-42c3-97b2-bafad3306958-swiftconf\") pod \"swift-ring-rebalance-debug-m2zzc\" (UID: \"8775d99d-e8a9-42c3-97b2-bafad3306958\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m2zzc" Mar 07 15:15:32 crc kubenswrapper[4943]: I0307 15:15:32.056236 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/8775d99d-e8a9-42c3-97b2-bafad3306958-ring-data-devices\") pod \"swift-ring-rebalance-debug-m2zzc\" (UID: \"8775d99d-e8a9-42c3-97b2-bafad3306958\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m2zzc" Mar 07 15:15:32 crc kubenswrapper[4943]: I0307 15:15:32.056310 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxktv\" (UniqueName: \"kubernetes.io/projected/8775d99d-e8a9-42c3-97b2-bafad3306958-kube-api-access-jxktv\") pod \"swift-ring-rebalance-debug-m2zzc\" (UID: \"8775d99d-e8a9-42c3-97b2-bafad3306958\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m2zzc" Mar 07 15:15:32 crc kubenswrapper[4943]: I0307 15:15:32.158192 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8775d99d-e8a9-42c3-97b2-bafad3306958-scripts\") pod \"swift-ring-rebalance-debug-m2zzc\" (UID: \"8775d99d-e8a9-42c3-97b2-bafad3306958\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m2zzc" Mar 07 15:15:32 crc kubenswrapper[4943]: I0307 15:15:32.158243 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/8775d99d-e8a9-42c3-97b2-bafad3306958-dispersionconf\") pod \"swift-ring-rebalance-debug-m2zzc\" (UID: \"8775d99d-e8a9-42c3-97b2-bafad3306958\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m2zzc" Mar 07 15:15:32 crc kubenswrapper[4943]: I0307 15:15:32.158272 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/8775d99d-e8a9-42c3-97b2-bafad3306958-etc-swift\") pod \"swift-ring-rebalance-debug-m2zzc\" (UID: \"8775d99d-e8a9-42c3-97b2-bafad3306958\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m2zzc" Mar 07 15:15:32 crc kubenswrapper[4943]: I0307 15:15:32.158296 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/8775d99d-e8a9-42c3-97b2-bafad3306958-swiftconf\") pod \"swift-ring-rebalance-debug-m2zzc\" (UID: \"8775d99d-e8a9-42c3-97b2-bafad3306958\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m2zzc" Mar 07 15:15:32 crc kubenswrapper[4943]: I0307 15:15:32.158359 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/8775d99d-e8a9-42c3-97b2-bafad3306958-ring-data-devices\") pod \"swift-ring-rebalance-debug-m2zzc\" (UID: \"8775d99d-e8a9-42c3-97b2-bafad3306958\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m2zzc" Mar 07 15:15:32 crc kubenswrapper[4943]: I0307 15:15:32.158395 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxktv\" (UniqueName: \"kubernetes.io/projected/8775d99d-e8a9-42c3-97b2-bafad3306958-kube-api-access-jxktv\") pod \"swift-ring-rebalance-debug-m2zzc\" (UID: \"8775d99d-e8a9-42c3-97b2-bafad3306958\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m2zzc" Mar 07 15:15:32 crc kubenswrapper[4943]: I0307 15:15:32.159557 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/8775d99d-e8a9-42c3-97b2-bafad3306958-etc-swift\") pod \"swift-ring-rebalance-debug-m2zzc\" (UID: \"8775d99d-e8a9-42c3-97b2-bafad3306958\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m2zzc" Mar 07 15:15:32 crc kubenswrapper[4943]: I0307 15:15:32.159631 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8775d99d-e8a9-42c3-97b2-bafad3306958-scripts\") pod \"swift-ring-rebalance-debug-m2zzc\" (UID: \"8775d99d-e8a9-42c3-97b2-bafad3306958\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m2zzc" Mar 07 15:15:32 crc kubenswrapper[4943]: I0307 15:15:32.159638 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/8775d99d-e8a9-42c3-97b2-bafad3306958-ring-data-devices\") pod \"swift-ring-rebalance-debug-m2zzc\" (UID: \"8775d99d-e8a9-42c3-97b2-bafad3306958\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m2zzc" Mar 07 15:15:32 crc kubenswrapper[4943]: I0307 15:15:32.164430 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/8775d99d-e8a9-42c3-97b2-bafad3306958-swiftconf\") pod \"swift-ring-rebalance-debug-m2zzc\" (UID: \"8775d99d-e8a9-42c3-97b2-bafad3306958\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m2zzc" Mar 07 15:15:32 crc kubenswrapper[4943]: I0307 15:15:32.170658 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/8775d99d-e8a9-42c3-97b2-bafad3306958-dispersionconf\") pod \"swift-ring-rebalance-debug-m2zzc\" (UID: \"8775d99d-e8a9-42c3-97b2-bafad3306958\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m2zzc" Mar 07 15:15:32 crc kubenswrapper[4943]: I0307 15:15:32.179541 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxktv\" (UniqueName: \"kubernetes.io/projected/8775d99d-e8a9-42c3-97b2-bafad3306958-kube-api-access-jxktv\") pod \"swift-ring-rebalance-debug-m2zzc\" (UID: \"8775d99d-e8a9-42c3-97b2-bafad3306958\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m2zzc" Mar 07 15:15:32 crc kubenswrapper[4943]: I0307 15:15:32.275374 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-m2zzc" Mar 07 15:15:32 crc kubenswrapper[4943]: I0307 15:15:32.516021 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-m2zzc"] Mar 07 15:15:32 crc kubenswrapper[4943]: I0307 15:15:32.771895 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad36e69d-d4e6-4c02-a340-1dfed75cf45b" path="/var/lib/kubelet/pods/ad36e69d-d4e6-4c02-a340-1dfed75cf45b/volumes" Mar 07 15:15:33 crc kubenswrapper[4943]: I0307 15:15:33.363402 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-m2zzc" event={"ID":"8775d99d-e8a9-42c3-97b2-bafad3306958","Type":"ContainerStarted","Data":"b74fb102430ba96aef617e3e49ffdf5c1b424ac7bef18593d09befce2d7dd8f6"} Mar 07 15:15:33 crc kubenswrapper[4943]: I0307 15:15:33.363465 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-m2zzc" event={"ID":"8775d99d-e8a9-42c3-97b2-bafad3306958","Type":"ContainerStarted","Data":"9aec8d8e7e3498ff5b2a98c636e51d0c6674f76b767bc7909c6d49228b4183f7"} Mar 07 15:15:33 crc kubenswrapper[4943]: I0307 15:15:33.386370 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-m2zzc" podStartSLOduration=2.386348746 podStartE2EDuration="2.386348746s" podCreationTimestamp="2026-03-07 15:15:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:15:33.384845939 +0000 UTC m=+2175.336982517" watchObservedRunningTime="2026-03-07 15:15:33.386348746 +0000 UTC m=+2175.338485244" Mar 07 15:15:34 crc kubenswrapper[4943]: I0307 15:15:34.382130 4943 generic.go:334] "Generic (PLEG): container finished" podID="8775d99d-e8a9-42c3-97b2-bafad3306958" containerID="b74fb102430ba96aef617e3e49ffdf5c1b424ac7bef18593d09befce2d7dd8f6" exitCode=0 Mar 07 15:15:34 crc kubenswrapper[4943]: I0307 15:15:34.382233 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-m2zzc" event={"ID":"8775d99d-e8a9-42c3-97b2-bafad3306958","Type":"ContainerDied","Data":"b74fb102430ba96aef617e3e49ffdf5c1b424ac7bef18593d09befce2d7dd8f6"} Mar 07 15:15:35 crc kubenswrapper[4943]: I0307 15:15:35.806356 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-m2zzc" Mar 07 15:15:35 crc kubenswrapper[4943]: I0307 15:15:35.842588 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-m2zzc"] Mar 07 15:15:35 crc kubenswrapper[4943]: I0307 15:15:35.847963 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-m2zzc"] Mar 07 15:15:35 crc kubenswrapper[4943]: I0307 15:15:35.917137 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/8775d99d-e8a9-42c3-97b2-bafad3306958-ring-data-devices\") pod \"8775d99d-e8a9-42c3-97b2-bafad3306958\" (UID: \"8775d99d-e8a9-42c3-97b2-bafad3306958\") " Mar 07 15:15:35 crc kubenswrapper[4943]: I0307 15:15:35.917227 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/8775d99d-e8a9-42c3-97b2-bafad3306958-swiftconf\") pod \"8775d99d-e8a9-42c3-97b2-bafad3306958\" (UID: \"8775d99d-e8a9-42c3-97b2-bafad3306958\") " Mar 07 15:15:35 crc kubenswrapper[4943]: I0307 15:15:35.917274 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jxktv\" (UniqueName: \"kubernetes.io/projected/8775d99d-e8a9-42c3-97b2-bafad3306958-kube-api-access-jxktv\") pod \"8775d99d-e8a9-42c3-97b2-bafad3306958\" (UID: \"8775d99d-e8a9-42c3-97b2-bafad3306958\") " Mar 07 15:15:35 crc kubenswrapper[4943]: I0307 15:15:35.917385 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/8775d99d-e8a9-42c3-97b2-bafad3306958-dispersionconf\") pod \"8775d99d-e8a9-42c3-97b2-bafad3306958\" (UID: \"8775d99d-e8a9-42c3-97b2-bafad3306958\") " Mar 07 15:15:35 crc kubenswrapper[4943]: I0307 15:15:35.917525 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8775d99d-e8a9-42c3-97b2-bafad3306958-scripts\") pod \"8775d99d-e8a9-42c3-97b2-bafad3306958\" (UID: \"8775d99d-e8a9-42c3-97b2-bafad3306958\") " Mar 07 15:15:35 crc kubenswrapper[4943]: I0307 15:15:35.917618 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/8775d99d-e8a9-42c3-97b2-bafad3306958-etc-swift\") pod \"8775d99d-e8a9-42c3-97b2-bafad3306958\" (UID: \"8775d99d-e8a9-42c3-97b2-bafad3306958\") " Mar 07 15:15:35 crc kubenswrapper[4943]: I0307 15:15:35.919066 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8775d99d-e8a9-42c3-97b2-bafad3306958-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "8775d99d-e8a9-42c3-97b2-bafad3306958" (UID: "8775d99d-e8a9-42c3-97b2-bafad3306958"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:15:35 crc kubenswrapper[4943]: I0307 15:15:35.919098 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8775d99d-e8a9-42c3-97b2-bafad3306958-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "8775d99d-e8a9-42c3-97b2-bafad3306958" (UID: "8775d99d-e8a9-42c3-97b2-bafad3306958"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:15:35 crc kubenswrapper[4943]: I0307 15:15:35.925354 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8775d99d-e8a9-42c3-97b2-bafad3306958-kube-api-access-jxktv" (OuterVolumeSpecName: "kube-api-access-jxktv") pod "8775d99d-e8a9-42c3-97b2-bafad3306958" (UID: "8775d99d-e8a9-42c3-97b2-bafad3306958"). InnerVolumeSpecName "kube-api-access-jxktv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:15:35 crc kubenswrapper[4943]: I0307 15:15:35.942127 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8775d99d-e8a9-42c3-97b2-bafad3306958-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "8775d99d-e8a9-42c3-97b2-bafad3306958" (UID: "8775d99d-e8a9-42c3-97b2-bafad3306958"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:15:35 crc kubenswrapper[4943]: I0307 15:15:35.942168 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8775d99d-e8a9-42c3-97b2-bafad3306958-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "8775d99d-e8a9-42c3-97b2-bafad3306958" (UID: "8775d99d-e8a9-42c3-97b2-bafad3306958"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:15:35 crc kubenswrapper[4943]: I0307 15:15:35.959951 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8775d99d-e8a9-42c3-97b2-bafad3306958-scripts" (OuterVolumeSpecName: "scripts") pod "8775d99d-e8a9-42c3-97b2-bafad3306958" (UID: "8775d99d-e8a9-42c3-97b2-bafad3306958"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:15:36 crc kubenswrapper[4943]: I0307 15:15:36.020022 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/8775d99d-e8a9-42c3-97b2-bafad3306958-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:36 crc kubenswrapper[4943]: I0307 15:15:36.020077 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8775d99d-e8a9-42c3-97b2-bafad3306958-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:36 crc kubenswrapper[4943]: I0307 15:15:36.020095 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/8775d99d-e8a9-42c3-97b2-bafad3306958-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:36 crc kubenswrapper[4943]: I0307 15:15:36.020112 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/8775d99d-e8a9-42c3-97b2-bafad3306958-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:36 crc kubenswrapper[4943]: I0307 15:15:36.020131 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/8775d99d-e8a9-42c3-97b2-bafad3306958-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:36 crc kubenswrapper[4943]: I0307 15:15:36.020147 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jxktv\" (UniqueName: \"kubernetes.io/projected/8775d99d-e8a9-42c3-97b2-bafad3306958-kube-api-access-jxktv\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:36 crc kubenswrapper[4943]: I0307 15:15:36.073657 4943 patch_prober.go:28] interesting pod/machine-config-daemon-cpdmm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 07 15:15:36 crc kubenswrapper[4943]: I0307 15:15:36.073730 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 07 15:15:36 crc kubenswrapper[4943]: I0307 15:15:36.413072 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9aec8d8e7e3498ff5b2a98c636e51d0c6674f76b767bc7909c6d49228b4183f7" Mar 07 15:15:36 crc kubenswrapper[4943]: I0307 15:15:36.413224 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-m2zzc" Mar 07 15:15:36 crc kubenswrapper[4943]: I0307 15:15:36.769003 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8775d99d-e8a9-42c3-97b2-bafad3306958" path="/var/lib/kubelet/pods/8775d99d-e8a9-42c3-97b2-bafad3306958/volumes" Mar 07 15:15:37 crc kubenswrapper[4943]: I0307 15:15:37.057565 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-j6fcx"] Mar 07 15:15:37 crc kubenswrapper[4943]: E0307 15:15:37.057977 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8775d99d-e8a9-42c3-97b2-bafad3306958" containerName="swift-ring-rebalance" Mar 07 15:15:37 crc kubenswrapper[4943]: I0307 15:15:37.057994 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="8775d99d-e8a9-42c3-97b2-bafad3306958" containerName="swift-ring-rebalance" Mar 07 15:15:37 crc kubenswrapper[4943]: I0307 15:15:37.058386 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="8775d99d-e8a9-42c3-97b2-bafad3306958" containerName="swift-ring-rebalance" Mar 07 15:15:37 crc kubenswrapper[4943]: I0307 15:15:37.059027 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j6fcx" Mar 07 15:15:37 crc kubenswrapper[4943]: I0307 15:15:37.064133 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:15:37 crc kubenswrapper[4943]: I0307 15:15:37.066578 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-j6fcx"] Mar 07 15:15:37 crc kubenswrapper[4943]: I0307 15:15:37.072561 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:15:37 crc kubenswrapper[4943]: I0307 15:15:37.138809 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ef1098ef-d004-43f4-ada0-ebc2db5f63da-ring-data-devices\") pod \"swift-ring-rebalance-debug-j6fcx\" (UID: \"ef1098ef-d004-43f4-ada0-ebc2db5f63da\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j6fcx" Mar 07 15:15:37 crc kubenswrapper[4943]: I0307 15:15:37.139031 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ef1098ef-d004-43f4-ada0-ebc2db5f63da-swiftconf\") pod \"swift-ring-rebalance-debug-j6fcx\" (UID: \"ef1098ef-d004-43f4-ada0-ebc2db5f63da\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j6fcx" Mar 07 15:15:37 crc kubenswrapper[4943]: I0307 15:15:37.139066 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ef1098ef-d004-43f4-ada0-ebc2db5f63da-dispersionconf\") pod \"swift-ring-rebalance-debug-j6fcx\" (UID: \"ef1098ef-d004-43f4-ada0-ebc2db5f63da\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j6fcx" Mar 07 15:15:37 crc kubenswrapper[4943]: I0307 15:15:37.139092 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ef1098ef-d004-43f4-ada0-ebc2db5f63da-etc-swift\") pod \"swift-ring-rebalance-debug-j6fcx\" (UID: \"ef1098ef-d004-43f4-ada0-ebc2db5f63da\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j6fcx" Mar 07 15:15:37 crc kubenswrapper[4943]: I0307 15:15:37.139152 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ef1098ef-d004-43f4-ada0-ebc2db5f63da-scripts\") pod \"swift-ring-rebalance-debug-j6fcx\" (UID: \"ef1098ef-d004-43f4-ada0-ebc2db5f63da\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j6fcx" Mar 07 15:15:37 crc kubenswrapper[4943]: I0307 15:15:37.139549 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-782h4\" (UniqueName: \"kubernetes.io/projected/ef1098ef-d004-43f4-ada0-ebc2db5f63da-kube-api-access-782h4\") pod \"swift-ring-rebalance-debug-j6fcx\" (UID: \"ef1098ef-d004-43f4-ada0-ebc2db5f63da\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j6fcx" Mar 07 15:15:37 crc kubenswrapper[4943]: I0307 15:15:37.241265 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-782h4\" (UniqueName: \"kubernetes.io/projected/ef1098ef-d004-43f4-ada0-ebc2db5f63da-kube-api-access-782h4\") pod \"swift-ring-rebalance-debug-j6fcx\" (UID: \"ef1098ef-d004-43f4-ada0-ebc2db5f63da\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j6fcx" Mar 07 15:15:37 crc kubenswrapper[4943]: I0307 15:15:37.241336 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ef1098ef-d004-43f4-ada0-ebc2db5f63da-ring-data-devices\") pod \"swift-ring-rebalance-debug-j6fcx\" (UID: \"ef1098ef-d004-43f4-ada0-ebc2db5f63da\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j6fcx" Mar 07 15:15:37 crc kubenswrapper[4943]: I0307 15:15:37.241445 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ef1098ef-d004-43f4-ada0-ebc2db5f63da-swiftconf\") pod \"swift-ring-rebalance-debug-j6fcx\" (UID: \"ef1098ef-d004-43f4-ada0-ebc2db5f63da\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j6fcx" Mar 07 15:15:37 crc kubenswrapper[4943]: I0307 15:15:37.241475 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ef1098ef-d004-43f4-ada0-ebc2db5f63da-dispersionconf\") pod \"swift-ring-rebalance-debug-j6fcx\" (UID: \"ef1098ef-d004-43f4-ada0-ebc2db5f63da\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j6fcx" Mar 07 15:15:37 crc kubenswrapper[4943]: I0307 15:15:37.241495 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ef1098ef-d004-43f4-ada0-ebc2db5f63da-etc-swift\") pod \"swift-ring-rebalance-debug-j6fcx\" (UID: \"ef1098ef-d004-43f4-ada0-ebc2db5f63da\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j6fcx" Mar 07 15:15:37 crc kubenswrapper[4943]: I0307 15:15:37.241522 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ef1098ef-d004-43f4-ada0-ebc2db5f63da-scripts\") pod \"swift-ring-rebalance-debug-j6fcx\" (UID: \"ef1098ef-d004-43f4-ada0-ebc2db5f63da\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j6fcx" Mar 07 15:15:37 crc kubenswrapper[4943]: I0307 15:15:37.242359 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ef1098ef-d004-43f4-ada0-ebc2db5f63da-ring-data-devices\") pod \"swift-ring-rebalance-debug-j6fcx\" (UID: \"ef1098ef-d004-43f4-ada0-ebc2db5f63da\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j6fcx" Mar 07 15:15:37 crc kubenswrapper[4943]: I0307 15:15:37.242377 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ef1098ef-d004-43f4-ada0-ebc2db5f63da-scripts\") pod \"swift-ring-rebalance-debug-j6fcx\" (UID: \"ef1098ef-d004-43f4-ada0-ebc2db5f63da\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j6fcx" Mar 07 15:15:37 crc kubenswrapper[4943]: I0307 15:15:37.242458 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ef1098ef-d004-43f4-ada0-ebc2db5f63da-etc-swift\") pod \"swift-ring-rebalance-debug-j6fcx\" (UID: \"ef1098ef-d004-43f4-ada0-ebc2db5f63da\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j6fcx" Mar 07 15:15:37 crc kubenswrapper[4943]: I0307 15:15:37.247218 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ef1098ef-d004-43f4-ada0-ebc2db5f63da-dispersionconf\") pod \"swift-ring-rebalance-debug-j6fcx\" (UID: \"ef1098ef-d004-43f4-ada0-ebc2db5f63da\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j6fcx" Mar 07 15:15:37 crc kubenswrapper[4943]: I0307 15:15:37.255650 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ef1098ef-d004-43f4-ada0-ebc2db5f63da-swiftconf\") pod \"swift-ring-rebalance-debug-j6fcx\" (UID: \"ef1098ef-d004-43f4-ada0-ebc2db5f63da\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j6fcx" Mar 07 15:15:37 crc kubenswrapper[4943]: I0307 15:15:37.262630 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-782h4\" (UniqueName: \"kubernetes.io/projected/ef1098ef-d004-43f4-ada0-ebc2db5f63da-kube-api-access-782h4\") pod \"swift-ring-rebalance-debug-j6fcx\" (UID: \"ef1098ef-d004-43f4-ada0-ebc2db5f63da\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j6fcx" Mar 07 15:15:37 crc kubenswrapper[4943]: I0307 15:15:37.390241 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j6fcx" Mar 07 15:15:37 crc kubenswrapper[4943]: I0307 15:15:37.905453 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-j6fcx"] Mar 07 15:15:37 crc kubenswrapper[4943]: W0307 15:15:37.910249 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podef1098ef_d004_43f4_ada0_ebc2db5f63da.slice/crio-c06976189dd4cbf2a4f0985e74c048de539c53bff8417e1db9ad810e28e819cd WatchSource:0}: Error finding container c06976189dd4cbf2a4f0985e74c048de539c53bff8417e1db9ad810e28e819cd: Status 404 returned error can't find the container with id c06976189dd4cbf2a4f0985e74c048de539c53bff8417e1db9ad810e28e819cd Mar 07 15:15:38 crc kubenswrapper[4943]: I0307 15:15:38.444497 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j6fcx" event={"ID":"ef1098ef-d004-43f4-ada0-ebc2db5f63da","Type":"ContainerStarted","Data":"6bf668a5979e8f1613e03fabd8ad361d86548ae73aafe064ef76ccb9eddbd1fd"} Mar 07 15:15:38 crc kubenswrapper[4943]: I0307 15:15:38.445047 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j6fcx" event={"ID":"ef1098ef-d004-43f4-ada0-ebc2db5f63da","Type":"ContainerStarted","Data":"c06976189dd4cbf2a4f0985e74c048de539c53bff8417e1db9ad810e28e819cd"} Mar 07 15:15:38 crc kubenswrapper[4943]: I0307 15:15:38.470910 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j6fcx" podStartSLOduration=1.47088635 podStartE2EDuration="1.47088635s" podCreationTimestamp="2026-03-07 15:15:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:15:38.466422791 +0000 UTC m=+2180.418559299" watchObservedRunningTime="2026-03-07 15:15:38.47088635 +0000 UTC m=+2180.423022858" Mar 07 15:15:39 crc kubenswrapper[4943]: I0307 15:15:39.458714 4943 generic.go:334] "Generic (PLEG): container finished" podID="ef1098ef-d004-43f4-ada0-ebc2db5f63da" containerID="6bf668a5979e8f1613e03fabd8ad361d86548ae73aafe064ef76ccb9eddbd1fd" exitCode=0 Mar 07 15:15:39 crc kubenswrapper[4943]: I0307 15:15:39.459105 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j6fcx" event={"ID":"ef1098ef-d004-43f4-ada0-ebc2db5f63da","Type":"ContainerDied","Data":"6bf668a5979e8f1613e03fabd8ad361d86548ae73aafe064ef76ccb9eddbd1fd"} Mar 07 15:15:40 crc kubenswrapper[4943]: I0307 15:15:40.894253 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j6fcx" Mar 07 15:15:40 crc kubenswrapper[4943]: I0307 15:15:40.945107 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-j6fcx"] Mar 07 15:15:40 crc kubenswrapper[4943]: I0307 15:15:40.950992 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-j6fcx"] Mar 07 15:15:41 crc kubenswrapper[4943]: I0307 15:15:41.003390 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ef1098ef-d004-43f4-ada0-ebc2db5f63da-scripts\") pod \"ef1098ef-d004-43f4-ada0-ebc2db5f63da\" (UID: \"ef1098ef-d004-43f4-ada0-ebc2db5f63da\") " Mar 07 15:15:41 crc kubenswrapper[4943]: I0307 15:15:41.003886 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ef1098ef-d004-43f4-ada0-ebc2db5f63da-swiftconf\") pod \"ef1098ef-d004-43f4-ada0-ebc2db5f63da\" (UID: \"ef1098ef-d004-43f4-ada0-ebc2db5f63da\") " Mar 07 15:15:41 crc kubenswrapper[4943]: I0307 15:15:41.003988 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ef1098ef-d004-43f4-ada0-ebc2db5f63da-ring-data-devices\") pod \"ef1098ef-d004-43f4-ada0-ebc2db5f63da\" (UID: \"ef1098ef-d004-43f4-ada0-ebc2db5f63da\") " Mar 07 15:15:41 crc kubenswrapper[4943]: I0307 15:15:41.004013 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-782h4\" (UniqueName: \"kubernetes.io/projected/ef1098ef-d004-43f4-ada0-ebc2db5f63da-kube-api-access-782h4\") pod \"ef1098ef-d004-43f4-ada0-ebc2db5f63da\" (UID: \"ef1098ef-d004-43f4-ada0-ebc2db5f63da\") " Mar 07 15:15:41 crc kubenswrapper[4943]: I0307 15:15:41.004072 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ef1098ef-d004-43f4-ada0-ebc2db5f63da-etc-swift\") pod \"ef1098ef-d004-43f4-ada0-ebc2db5f63da\" (UID: \"ef1098ef-d004-43f4-ada0-ebc2db5f63da\") " Mar 07 15:15:41 crc kubenswrapper[4943]: I0307 15:15:41.004105 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ef1098ef-d004-43f4-ada0-ebc2db5f63da-dispersionconf\") pod \"ef1098ef-d004-43f4-ada0-ebc2db5f63da\" (UID: \"ef1098ef-d004-43f4-ada0-ebc2db5f63da\") " Mar 07 15:15:41 crc kubenswrapper[4943]: I0307 15:15:41.004780 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef1098ef-d004-43f4-ada0-ebc2db5f63da-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "ef1098ef-d004-43f4-ada0-ebc2db5f63da" (UID: "ef1098ef-d004-43f4-ada0-ebc2db5f63da"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:15:41 crc kubenswrapper[4943]: I0307 15:15:41.004862 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef1098ef-d004-43f4-ada0-ebc2db5f63da-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "ef1098ef-d004-43f4-ada0-ebc2db5f63da" (UID: "ef1098ef-d004-43f4-ada0-ebc2db5f63da"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:15:41 crc kubenswrapper[4943]: I0307 15:15:41.008905 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef1098ef-d004-43f4-ada0-ebc2db5f63da-kube-api-access-782h4" (OuterVolumeSpecName: "kube-api-access-782h4") pod "ef1098ef-d004-43f4-ada0-ebc2db5f63da" (UID: "ef1098ef-d004-43f4-ada0-ebc2db5f63da"). InnerVolumeSpecName "kube-api-access-782h4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:15:41 crc kubenswrapper[4943]: I0307 15:15:41.026565 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef1098ef-d004-43f4-ada0-ebc2db5f63da-scripts" (OuterVolumeSpecName: "scripts") pod "ef1098ef-d004-43f4-ada0-ebc2db5f63da" (UID: "ef1098ef-d004-43f4-ada0-ebc2db5f63da"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:15:41 crc kubenswrapper[4943]: I0307 15:15:41.029838 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef1098ef-d004-43f4-ada0-ebc2db5f63da-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "ef1098ef-d004-43f4-ada0-ebc2db5f63da" (UID: "ef1098ef-d004-43f4-ada0-ebc2db5f63da"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:15:41 crc kubenswrapper[4943]: I0307 15:15:41.040701 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef1098ef-d004-43f4-ada0-ebc2db5f63da-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "ef1098ef-d004-43f4-ada0-ebc2db5f63da" (UID: "ef1098ef-d004-43f4-ada0-ebc2db5f63da"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:15:41 crc kubenswrapper[4943]: I0307 15:15:41.105730 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ef1098ef-d004-43f4-ada0-ebc2db5f63da-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:41 crc kubenswrapper[4943]: I0307 15:15:41.105786 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ef1098ef-d004-43f4-ada0-ebc2db5f63da-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:41 crc kubenswrapper[4943]: I0307 15:15:41.105820 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ef1098ef-d004-43f4-ada0-ebc2db5f63da-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:41 crc kubenswrapper[4943]: I0307 15:15:41.105853 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-782h4\" (UniqueName: \"kubernetes.io/projected/ef1098ef-d004-43f4-ada0-ebc2db5f63da-kube-api-access-782h4\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:41 crc kubenswrapper[4943]: I0307 15:15:41.105881 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ef1098ef-d004-43f4-ada0-ebc2db5f63da-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:41 crc kubenswrapper[4943]: I0307 15:15:41.105906 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ef1098ef-d004-43f4-ada0-ebc2db5f63da-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:41 crc kubenswrapper[4943]: I0307 15:15:41.490004 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c06976189dd4cbf2a4f0985e74c048de539c53bff8417e1db9ad810e28e819cd" Mar 07 15:15:41 crc kubenswrapper[4943]: I0307 15:15:41.490131 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j6fcx" Mar 07 15:15:42 crc kubenswrapper[4943]: I0307 15:15:42.136780 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-tntbf"] Mar 07 15:15:42 crc kubenswrapper[4943]: E0307 15:15:42.137297 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef1098ef-d004-43f4-ada0-ebc2db5f63da" containerName="swift-ring-rebalance" Mar 07 15:15:42 crc kubenswrapper[4943]: I0307 15:15:42.137320 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef1098ef-d004-43f4-ada0-ebc2db5f63da" containerName="swift-ring-rebalance" Mar 07 15:15:42 crc kubenswrapper[4943]: I0307 15:15:42.137563 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef1098ef-d004-43f4-ada0-ebc2db5f63da" containerName="swift-ring-rebalance" Mar 07 15:15:42 crc kubenswrapper[4943]: I0307 15:15:42.138317 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-tntbf" Mar 07 15:15:42 crc kubenswrapper[4943]: I0307 15:15:42.142199 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:15:42 crc kubenswrapper[4943]: I0307 15:15:42.142270 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:15:42 crc kubenswrapper[4943]: I0307 15:15:42.160820 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-tntbf"] Mar 07 15:15:42 crc kubenswrapper[4943]: I0307 15:15:42.224131 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6f090df6-1dbf-4e85-91ba-163aded255a2-scripts\") pod \"swift-ring-rebalance-debug-tntbf\" (UID: \"6f090df6-1dbf-4e85-91ba-163aded255a2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tntbf" Mar 07 15:15:42 crc kubenswrapper[4943]: I0307 15:15:42.224557 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6f090df6-1dbf-4e85-91ba-163aded255a2-swiftconf\") pod \"swift-ring-rebalance-debug-tntbf\" (UID: \"6f090df6-1dbf-4e85-91ba-163aded255a2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tntbf" Mar 07 15:15:42 crc kubenswrapper[4943]: I0307 15:15:42.224701 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6f090df6-1dbf-4e85-91ba-163aded255a2-dispersionconf\") pod \"swift-ring-rebalance-debug-tntbf\" (UID: \"6f090df6-1dbf-4e85-91ba-163aded255a2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tntbf" Mar 07 15:15:42 crc kubenswrapper[4943]: I0307 15:15:42.224765 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6f090df6-1dbf-4e85-91ba-163aded255a2-ring-data-devices\") pod \"swift-ring-rebalance-debug-tntbf\" (UID: \"6f090df6-1dbf-4e85-91ba-163aded255a2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tntbf" Mar 07 15:15:42 crc kubenswrapper[4943]: I0307 15:15:42.224805 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8f89k\" (UniqueName: \"kubernetes.io/projected/6f090df6-1dbf-4e85-91ba-163aded255a2-kube-api-access-8f89k\") pod \"swift-ring-rebalance-debug-tntbf\" (UID: \"6f090df6-1dbf-4e85-91ba-163aded255a2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tntbf" Mar 07 15:15:42 crc kubenswrapper[4943]: I0307 15:15:42.224880 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6f090df6-1dbf-4e85-91ba-163aded255a2-etc-swift\") pod \"swift-ring-rebalance-debug-tntbf\" (UID: \"6f090df6-1dbf-4e85-91ba-163aded255a2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tntbf" Mar 07 15:15:42 crc kubenswrapper[4943]: I0307 15:15:42.326371 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6f090df6-1dbf-4e85-91ba-163aded255a2-etc-swift\") pod \"swift-ring-rebalance-debug-tntbf\" (UID: \"6f090df6-1dbf-4e85-91ba-163aded255a2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tntbf" Mar 07 15:15:42 crc kubenswrapper[4943]: I0307 15:15:42.326511 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6f090df6-1dbf-4e85-91ba-163aded255a2-scripts\") pod \"swift-ring-rebalance-debug-tntbf\" (UID: \"6f090df6-1dbf-4e85-91ba-163aded255a2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tntbf" Mar 07 15:15:42 crc kubenswrapper[4943]: I0307 15:15:42.326587 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6f090df6-1dbf-4e85-91ba-163aded255a2-swiftconf\") pod \"swift-ring-rebalance-debug-tntbf\" (UID: \"6f090df6-1dbf-4e85-91ba-163aded255a2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tntbf" Mar 07 15:15:42 crc kubenswrapper[4943]: I0307 15:15:42.326710 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6f090df6-1dbf-4e85-91ba-163aded255a2-dispersionconf\") pod \"swift-ring-rebalance-debug-tntbf\" (UID: \"6f090df6-1dbf-4e85-91ba-163aded255a2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tntbf" Mar 07 15:15:42 crc kubenswrapper[4943]: I0307 15:15:42.326751 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6f090df6-1dbf-4e85-91ba-163aded255a2-ring-data-devices\") pod \"swift-ring-rebalance-debug-tntbf\" (UID: \"6f090df6-1dbf-4e85-91ba-163aded255a2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tntbf" Mar 07 15:15:42 crc kubenswrapper[4943]: I0307 15:15:42.326786 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8f89k\" (UniqueName: \"kubernetes.io/projected/6f090df6-1dbf-4e85-91ba-163aded255a2-kube-api-access-8f89k\") pod \"swift-ring-rebalance-debug-tntbf\" (UID: \"6f090df6-1dbf-4e85-91ba-163aded255a2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tntbf" Mar 07 15:15:42 crc kubenswrapper[4943]: I0307 15:15:42.327453 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6f090df6-1dbf-4e85-91ba-163aded255a2-scripts\") pod \"swift-ring-rebalance-debug-tntbf\" (UID: \"6f090df6-1dbf-4e85-91ba-163aded255a2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tntbf" Mar 07 15:15:42 crc kubenswrapper[4943]: I0307 15:15:42.327608 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6f090df6-1dbf-4e85-91ba-163aded255a2-ring-data-devices\") pod \"swift-ring-rebalance-debug-tntbf\" (UID: \"6f090df6-1dbf-4e85-91ba-163aded255a2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tntbf" Mar 07 15:15:42 crc kubenswrapper[4943]: I0307 15:15:42.327691 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6f090df6-1dbf-4e85-91ba-163aded255a2-etc-swift\") pod \"swift-ring-rebalance-debug-tntbf\" (UID: \"6f090df6-1dbf-4e85-91ba-163aded255a2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tntbf" Mar 07 15:15:42 crc kubenswrapper[4943]: I0307 15:15:42.333372 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6f090df6-1dbf-4e85-91ba-163aded255a2-dispersionconf\") pod \"swift-ring-rebalance-debug-tntbf\" (UID: \"6f090df6-1dbf-4e85-91ba-163aded255a2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tntbf" Mar 07 15:15:42 crc kubenswrapper[4943]: I0307 15:15:42.337238 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6f090df6-1dbf-4e85-91ba-163aded255a2-swiftconf\") pod \"swift-ring-rebalance-debug-tntbf\" (UID: \"6f090df6-1dbf-4e85-91ba-163aded255a2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tntbf" Mar 07 15:15:42 crc kubenswrapper[4943]: I0307 15:15:42.360617 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8f89k\" (UniqueName: \"kubernetes.io/projected/6f090df6-1dbf-4e85-91ba-163aded255a2-kube-api-access-8f89k\") pod \"swift-ring-rebalance-debug-tntbf\" (UID: \"6f090df6-1dbf-4e85-91ba-163aded255a2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tntbf" Mar 07 15:15:42 crc kubenswrapper[4943]: I0307 15:15:42.475646 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-tntbf" Mar 07 15:15:42 crc kubenswrapper[4943]: I0307 15:15:42.767350 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef1098ef-d004-43f4-ada0-ebc2db5f63da" path="/var/lib/kubelet/pods/ef1098ef-d004-43f4-ada0-ebc2db5f63da/volumes" Mar 07 15:15:42 crc kubenswrapper[4943]: I0307 15:15:42.985702 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-tntbf"] Mar 07 15:15:42 crc kubenswrapper[4943]: W0307 15:15:42.994816 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6f090df6_1dbf_4e85_91ba_163aded255a2.slice/crio-10442179d6983b41bb5426ec9cf18d26e5a48d9f6367dc000e9760976bf134f1 WatchSource:0}: Error finding container 10442179d6983b41bb5426ec9cf18d26e5a48d9f6367dc000e9760976bf134f1: Status 404 returned error can't find the container with id 10442179d6983b41bb5426ec9cf18d26e5a48d9f6367dc000e9760976bf134f1 Mar 07 15:15:43 crc kubenswrapper[4943]: I0307 15:15:43.528693 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-tntbf" event={"ID":"6f090df6-1dbf-4e85-91ba-163aded255a2","Type":"ContainerStarted","Data":"4f6fa82d7c8fbcb5ffbe12a1b3c1454a3713c428d9888a404d38e918c2e10b37"} Mar 07 15:15:43 crc kubenswrapper[4943]: I0307 15:15:43.529184 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-tntbf" event={"ID":"6f090df6-1dbf-4e85-91ba-163aded255a2","Type":"ContainerStarted","Data":"10442179d6983b41bb5426ec9cf18d26e5a48d9f6367dc000e9760976bf134f1"} Mar 07 15:15:43 crc kubenswrapper[4943]: I0307 15:15:43.568014 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-tntbf" podStartSLOduration=1.5679733420000002 podStartE2EDuration="1.567973342s" podCreationTimestamp="2026-03-07 15:15:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:15:43.553110108 +0000 UTC m=+2185.505246666" watchObservedRunningTime="2026-03-07 15:15:43.567973342 +0000 UTC m=+2185.520109860" Mar 07 15:15:44 crc kubenswrapper[4943]: I0307 15:15:44.541827 4943 generic.go:334] "Generic (PLEG): container finished" podID="6f090df6-1dbf-4e85-91ba-163aded255a2" containerID="4f6fa82d7c8fbcb5ffbe12a1b3c1454a3713c428d9888a404d38e918c2e10b37" exitCode=0 Mar 07 15:15:44 crc kubenswrapper[4943]: I0307 15:15:44.542005 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-tntbf" event={"ID":"6f090df6-1dbf-4e85-91ba-163aded255a2","Type":"ContainerDied","Data":"4f6fa82d7c8fbcb5ffbe12a1b3c1454a3713c428d9888a404d38e918c2e10b37"} Mar 07 15:15:45 crc kubenswrapper[4943]: I0307 15:15:45.945475 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-tntbf" Mar 07 15:15:45 crc kubenswrapper[4943]: I0307 15:15:45.982314 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-tntbf"] Mar 07 15:15:45 crc kubenswrapper[4943]: I0307 15:15:45.986633 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-tntbf"] Mar 07 15:15:46 crc kubenswrapper[4943]: I0307 15:15:46.090904 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6f090df6-1dbf-4e85-91ba-163aded255a2-swiftconf\") pod \"6f090df6-1dbf-4e85-91ba-163aded255a2\" (UID: \"6f090df6-1dbf-4e85-91ba-163aded255a2\") " Mar 07 15:15:46 crc kubenswrapper[4943]: I0307 15:15:46.091028 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6f090df6-1dbf-4e85-91ba-163aded255a2-dispersionconf\") pod \"6f090df6-1dbf-4e85-91ba-163aded255a2\" (UID: \"6f090df6-1dbf-4e85-91ba-163aded255a2\") " Mar 07 15:15:46 crc kubenswrapper[4943]: I0307 15:15:46.091177 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6f090df6-1dbf-4e85-91ba-163aded255a2-scripts\") pod \"6f090df6-1dbf-4e85-91ba-163aded255a2\" (UID: \"6f090df6-1dbf-4e85-91ba-163aded255a2\") " Mar 07 15:15:46 crc kubenswrapper[4943]: I0307 15:15:46.091376 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8f89k\" (UniqueName: \"kubernetes.io/projected/6f090df6-1dbf-4e85-91ba-163aded255a2-kube-api-access-8f89k\") pod \"6f090df6-1dbf-4e85-91ba-163aded255a2\" (UID: \"6f090df6-1dbf-4e85-91ba-163aded255a2\") " Mar 07 15:15:46 crc kubenswrapper[4943]: I0307 15:15:46.091426 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6f090df6-1dbf-4e85-91ba-163aded255a2-etc-swift\") pod \"6f090df6-1dbf-4e85-91ba-163aded255a2\" (UID: \"6f090df6-1dbf-4e85-91ba-163aded255a2\") " Mar 07 15:15:46 crc kubenswrapper[4943]: I0307 15:15:46.091480 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6f090df6-1dbf-4e85-91ba-163aded255a2-ring-data-devices\") pod \"6f090df6-1dbf-4e85-91ba-163aded255a2\" (UID: \"6f090df6-1dbf-4e85-91ba-163aded255a2\") " Mar 07 15:15:46 crc kubenswrapper[4943]: I0307 15:15:46.092708 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f090df6-1dbf-4e85-91ba-163aded255a2-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "6f090df6-1dbf-4e85-91ba-163aded255a2" (UID: "6f090df6-1dbf-4e85-91ba-163aded255a2"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:15:46 crc kubenswrapper[4943]: I0307 15:15:46.093250 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6f090df6-1dbf-4e85-91ba-163aded255a2-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "6f090df6-1dbf-4e85-91ba-163aded255a2" (UID: "6f090df6-1dbf-4e85-91ba-163aded255a2"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:15:46 crc kubenswrapper[4943]: I0307 15:15:46.105207 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f090df6-1dbf-4e85-91ba-163aded255a2-kube-api-access-8f89k" (OuterVolumeSpecName: "kube-api-access-8f89k") pod "6f090df6-1dbf-4e85-91ba-163aded255a2" (UID: "6f090df6-1dbf-4e85-91ba-163aded255a2"). InnerVolumeSpecName "kube-api-access-8f89k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:15:46 crc kubenswrapper[4943]: I0307 15:15:46.131072 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f090df6-1dbf-4e85-91ba-163aded255a2-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "6f090df6-1dbf-4e85-91ba-163aded255a2" (UID: "6f090df6-1dbf-4e85-91ba-163aded255a2"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:15:46 crc kubenswrapper[4943]: I0307 15:15:46.135860 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f090df6-1dbf-4e85-91ba-163aded255a2-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "6f090df6-1dbf-4e85-91ba-163aded255a2" (UID: "6f090df6-1dbf-4e85-91ba-163aded255a2"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:15:46 crc kubenswrapper[4943]: I0307 15:15:46.139093 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f090df6-1dbf-4e85-91ba-163aded255a2-scripts" (OuterVolumeSpecName: "scripts") pod "6f090df6-1dbf-4e85-91ba-163aded255a2" (UID: "6f090df6-1dbf-4e85-91ba-163aded255a2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:15:46 crc kubenswrapper[4943]: I0307 15:15:46.193704 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6f090df6-1dbf-4e85-91ba-163aded255a2-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:46 crc kubenswrapper[4943]: I0307 15:15:46.193769 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6f090df6-1dbf-4e85-91ba-163aded255a2-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:46 crc kubenswrapper[4943]: I0307 15:15:46.193793 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6f090df6-1dbf-4e85-91ba-163aded255a2-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:46 crc kubenswrapper[4943]: I0307 15:15:46.193815 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8f89k\" (UniqueName: \"kubernetes.io/projected/6f090df6-1dbf-4e85-91ba-163aded255a2-kube-api-access-8f89k\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:46 crc kubenswrapper[4943]: I0307 15:15:46.193835 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6f090df6-1dbf-4e85-91ba-163aded255a2-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:46 crc kubenswrapper[4943]: I0307 15:15:46.193851 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6f090df6-1dbf-4e85-91ba-163aded255a2-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:46 crc kubenswrapper[4943]: I0307 15:15:46.567916 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="10442179d6983b41bb5426ec9cf18d26e5a48d9f6367dc000e9760976bf134f1" Mar 07 15:15:46 crc kubenswrapper[4943]: I0307 15:15:46.568001 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-tntbf" Mar 07 15:15:46 crc kubenswrapper[4943]: I0307 15:15:46.773655 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f090df6-1dbf-4e85-91ba-163aded255a2" path="/var/lib/kubelet/pods/6f090df6-1dbf-4e85-91ba-163aded255a2/volumes" Mar 07 15:15:47 crc kubenswrapper[4943]: I0307 15:15:47.140745 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-njf9l"] Mar 07 15:15:47 crc kubenswrapper[4943]: E0307 15:15:47.141211 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f090df6-1dbf-4e85-91ba-163aded255a2" containerName="swift-ring-rebalance" Mar 07 15:15:47 crc kubenswrapper[4943]: I0307 15:15:47.141233 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f090df6-1dbf-4e85-91ba-163aded255a2" containerName="swift-ring-rebalance" Mar 07 15:15:47 crc kubenswrapper[4943]: I0307 15:15:47.141509 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f090df6-1dbf-4e85-91ba-163aded255a2" containerName="swift-ring-rebalance" Mar 07 15:15:47 crc kubenswrapper[4943]: I0307 15:15:47.142964 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-njf9l" Mar 07 15:15:47 crc kubenswrapper[4943]: I0307 15:15:47.146277 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:15:47 crc kubenswrapper[4943]: I0307 15:15:47.146641 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:15:47 crc kubenswrapper[4943]: I0307 15:15:47.159068 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-njf9l"] Mar 07 15:15:47 crc kubenswrapper[4943]: I0307 15:15:47.211639 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0fe46184-7453-4cc6-8e87-6abd5f05dd68-scripts\") pod \"swift-ring-rebalance-debug-njf9l\" (UID: \"0fe46184-7453-4cc6-8e87-6abd5f05dd68\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-njf9l" Mar 07 15:15:47 crc kubenswrapper[4943]: I0307 15:15:47.211698 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/0fe46184-7453-4cc6-8e87-6abd5f05dd68-swiftconf\") pod \"swift-ring-rebalance-debug-njf9l\" (UID: \"0fe46184-7453-4cc6-8e87-6abd5f05dd68\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-njf9l" Mar 07 15:15:47 crc kubenswrapper[4943]: I0307 15:15:47.211715 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/0fe46184-7453-4cc6-8e87-6abd5f05dd68-etc-swift\") pod \"swift-ring-rebalance-debug-njf9l\" (UID: \"0fe46184-7453-4cc6-8e87-6abd5f05dd68\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-njf9l" Mar 07 15:15:47 crc kubenswrapper[4943]: I0307 15:15:47.211890 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/0fe46184-7453-4cc6-8e87-6abd5f05dd68-dispersionconf\") pod \"swift-ring-rebalance-debug-njf9l\" (UID: \"0fe46184-7453-4cc6-8e87-6abd5f05dd68\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-njf9l" Mar 07 15:15:47 crc kubenswrapper[4943]: I0307 15:15:47.212187 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhcn5\" (UniqueName: \"kubernetes.io/projected/0fe46184-7453-4cc6-8e87-6abd5f05dd68-kube-api-access-fhcn5\") pod \"swift-ring-rebalance-debug-njf9l\" (UID: \"0fe46184-7453-4cc6-8e87-6abd5f05dd68\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-njf9l" Mar 07 15:15:47 crc kubenswrapper[4943]: I0307 15:15:47.212755 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/0fe46184-7453-4cc6-8e87-6abd5f05dd68-ring-data-devices\") pod \"swift-ring-rebalance-debug-njf9l\" (UID: \"0fe46184-7453-4cc6-8e87-6abd5f05dd68\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-njf9l" Mar 07 15:15:47 crc kubenswrapper[4943]: I0307 15:15:47.314048 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0fe46184-7453-4cc6-8e87-6abd5f05dd68-scripts\") pod \"swift-ring-rebalance-debug-njf9l\" (UID: \"0fe46184-7453-4cc6-8e87-6abd5f05dd68\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-njf9l" Mar 07 15:15:47 crc kubenswrapper[4943]: I0307 15:15:47.314164 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/0fe46184-7453-4cc6-8e87-6abd5f05dd68-swiftconf\") pod \"swift-ring-rebalance-debug-njf9l\" (UID: \"0fe46184-7453-4cc6-8e87-6abd5f05dd68\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-njf9l" Mar 07 15:15:47 crc kubenswrapper[4943]: I0307 15:15:47.314197 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/0fe46184-7453-4cc6-8e87-6abd5f05dd68-etc-swift\") pod \"swift-ring-rebalance-debug-njf9l\" (UID: \"0fe46184-7453-4cc6-8e87-6abd5f05dd68\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-njf9l" Mar 07 15:15:47 crc kubenswrapper[4943]: I0307 15:15:47.314252 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/0fe46184-7453-4cc6-8e87-6abd5f05dd68-dispersionconf\") pod \"swift-ring-rebalance-debug-njf9l\" (UID: \"0fe46184-7453-4cc6-8e87-6abd5f05dd68\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-njf9l" Mar 07 15:15:47 crc kubenswrapper[4943]: I0307 15:15:47.314323 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhcn5\" (UniqueName: \"kubernetes.io/projected/0fe46184-7453-4cc6-8e87-6abd5f05dd68-kube-api-access-fhcn5\") pod \"swift-ring-rebalance-debug-njf9l\" (UID: \"0fe46184-7453-4cc6-8e87-6abd5f05dd68\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-njf9l" Mar 07 15:15:47 crc kubenswrapper[4943]: I0307 15:15:47.314373 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/0fe46184-7453-4cc6-8e87-6abd5f05dd68-ring-data-devices\") pod \"swift-ring-rebalance-debug-njf9l\" (UID: \"0fe46184-7453-4cc6-8e87-6abd5f05dd68\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-njf9l" Mar 07 15:15:47 crc kubenswrapper[4943]: I0307 15:15:47.315284 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/0fe46184-7453-4cc6-8e87-6abd5f05dd68-etc-swift\") pod \"swift-ring-rebalance-debug-njf9l\" (UID: \"0fe46184-7453-4cc6-8e87-6abd5f05dd68\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-njf9l" Mar 07 15:15:47 crc kubenswrapper[4943]: I0307 15:15:47.315791 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/0fe46184-7453-4cc6-8e87-6abd5f05dd68-ring-data-devices\") pod \"swift-ring-rebalance-debug-njf9l\" (UID: \"0fe46184-7453-4cc6-8e87-6abd5f05dd68\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-njf9l" Mar 07 15:15:47 crc kubenswrapper[4943]: I0307 15:15:47.316840 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0fe46184-7453-4cc6-8e87-6abd5f05dd68-scripts\") pod \"swift-ring-rebalance-debug-njf9l\" (UID: \"0fe46184-7453-4cc6-8e87-6abd5f05dd68\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-njf9l" Mar 07 15:15:47 crc kubenswrapper[4943]: I0307 15:15:47.321863 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/0fe46184-7453-4cc6-8e87-6abd5f05dd68-swiftconf\") pod \"swift-ring-rebalance-debug-njf9l\" (UID: \"0fe46184-7453-4cc6-8e87-6abd5f05dd68\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-njf9l" Mar 07 15:15:47 crc kubenswrapper[4943]: I0307 15:15:47.323696 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/0fe46184-7453-4cc6-8e87-6abd5f05dd68-dispersionconf\") pod \"swift-ring-rebalance-debug-njf9l\" (UID: \"0fe46184-7453-4cc6-8e87-6abd5f05dd68\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-njf9l" Mar 07 15:15:47 crc kubenswrapper[4943]: I0307 15:15:47.339338 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhcn5\" (UniqueName: \"kubernetes.io/projected/0fe46184-7453-4cc6-8e87-6abd5f05dd68-kube-api-access-fhcn5\") pod \"swift-ring-rebalance-debug-njf9l\" (UID: \"0fe46184-7453-4cc6-8e87-6abd5f05dd68\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-njf9l" Mar 07 15:15:47 crc kubenswrapper[4943]: I0307 15:15:47.480532 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-njf9l" Mar 07 15:15:47 crc kubenswrapper[4943]: I0307 15:15:47.795758 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-njf9l"] Mar 07 15:15:48 crc kubenswrapper[4943]: I0307 15:15:48.601569 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-njf9l" event={"ID":"0fe46184-7453-4cc6-8e87-6abd5f05dd68","Type":"ContainerStarted","Data":"0427dcf71f86ffa745eea10704d4fb1aacf6e0c845aa3d0cf085fb10961aedf2"} Mar 07 15:15:48 crc kubenswrapper[4943]: I0307 15:15:48.604114 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-njf9l" event={"ID":"0fe46184-7453-4cc6-8e87-6abd5f05dd68","Type":"ContainerStarted","Data":"a67902ef7a79df36058714d34c81f72a146616ed2191e707865638efa1e136ba"} Mar 07 15:15:48 crc kubenswrapper[4943]: I0307 15:15:48.633553 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-njf9l" podStartSLOduration=1.633519272 podStartE2EDuration="1.633519272s" podCreationTimestamp="2026-03-07 15:15:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:15:48.631332928 +0000 UTC m=+2190.583469486" watchObservedRunningTime="2026-03-07 15:15:48.633519272 +0000 UTC m=+2190.585655810" Mar 07 15:15:49 crc kubenswrapper[4943]: I0307 15:15:49.612494 4943 generic.go:334] "Generic (PLEG): container finished" podID="0fe46184-7453-4cc6-8e87-6abd5f05dd68" containerID="0427dcf71f86ffa745eea10704d4fb1aacf6e0c845aa3d0cf085fb10961aedf2" exitCode=0 Mar 07 15:15:49 crc kubenswrapper[4943]: I0307 15:15:49.612610 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-njf9l" event={"ID":"0fe46184-7453-4cc6-8e87-6abd5f05dd68","Type":"ContainerDied","Data":"0427dcf71f86ffa745eea10704d4fb1aacf6e0c845aa3d0cf085fb10961aedf2"} Mar 07 15:15:50 crc kubenswrapper[4943]: I0307 15:15:50.921402 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-njf9l" Mar 07 15:15:50 crc kubenswrapper[4943]: I0307 15:15:50.953612 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-njf9l"] Mar 07 15:15:50 crc kubenswrapper[4943]: I0307 15:15:50.959510 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-njf9l"] Mar 07 15:15:51 crc kubenswrapper[4943]: I0307 15:15:51.076195 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/0fe46184-7453-4cc6-8e87-6abd5f05dd68-swiftconf\") pod \"0fe46184-7453-4cc6-8e87-6abd5f05dd68\" (UID: \"0fe46184-7453-4cc6-8e87-6abd5f05dd68\") " Mar 07 15:15:51 crc kubenswrapper[4943]: I0307 15:15:51.076320 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fhcn5\" (UniqueName: \"kubernetes.io/projected/0fe46184-7453-4cc6-8e87-6abd5f05dd68-kube-api-access-fhcn5\") pod \"0fe46184-7453-4cc6-8e87-6abd5f05dd68\" (UID: \"0fe46184-7453-4cc6-8e87-6abd5f05dd68\") " Mar 07 15:15:51 crc kubenswrapper[4943]: I0307 15:15:51.076399 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/0fe46184-7453-4cc6-8e87-6abd5f05dd68-dispersionconf\") pod \"0fe46184-7453-4cc6-8e87-6abd5f05dd68\" (UID: \"0fe46184-7453-4cc6-8e87-6abd5f05dd68\") " Mar 07 15:15:51 crc kubenswrapper[4943]: I0307 15:15:51.076542 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0fe46184-7453-4cc6-8e87-6abd5f05dd68-scripts\") pod \"0fe46184-7453-4cc6-8e87-6abd5f05dd68\" (UID: \"0fe46184-7453-4cc6-8e87-6abd5f05dd68\") " Mar 07 15:15:51 crc kubenswrapper[4943]: I0307 15:15:51.076621 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/0fe46184-7453-4cc6-8e87-6abd5f05dd68-etc-swift\") pod \"0fe46184-7453-4cc6-8e87-6abd5f05dd68\" (UID: \"0fe46184-7453-4cc6-8e87-6abd5f05dd68\") " Mar 07 15:15:51 crc kubenswrapper[4943]: I0307 15:15:51.076743 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/0fe46184-7453-4cc6-8e87-6abd5f05dd68-ring-data-devices\") pod \"0fe46184-7453-4cc6-8e87-6abd5f05dd68\" (UID: \"0fe46184-7453-4cc6-8e87-6abd5f05dd68\") " Mar 07 15:15:51 crc kubenswrapper[4943]: I0307 15:15:51.077608 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0fe46184-7453-4cc6-8e87-6abd5f05dd68-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "0fe46184-7453-4cc6-8e87-6abd5f05dd68" (UID: "0fe46184-7453-4cc6-8e87-6abd5f05dd68"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:15:51 crc kubenswrapper[4943]: I0307 15:15:51.077865 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0fe46184-7453-4cc6-8e87-6abd5f05dd68-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "0fe46184-7453-4cc6-8e87-6abd5f05dd68" (UID: "0fe46184-7453-4cc6-8e87-6abd5f05dd68"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:15:51 crc kubenswrapper[4943]: I0307 15:15:51.084107 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0fe46184-7453-4cc6-8e87-6abd5f05dd68-kube-api-access-fhcn5" (OuterVolumeSpecName: "kube-api-access-fhcn5") pod "0fe46184-7453-4cc6-8e87-6abd5f05dd68" (UID: "0fe46184-7453-4cc6-8e87-6abd5f05dd68"). InnerVolumeSpecName "kube-api-access-fhcn5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:15:51 crc kubenswrapper[4943]: I0307 15:15:51.100266 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0fe46184-7453-4cc6-8e87-6abd5f05dd68-scripts" (OuterVolumeSpecName: "scripts") pod "0fe46184-7453-4cc6-8e87-6abd5f05dd68" (UID: "0fe46184-7453-4cc6-8e87-6abd5f05dd68"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:15:51 crc kubenswrapper[4943]: I0307 15:15:51.102068 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fe46184-7453-4cc6-8e87-6abd5f05dd68-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "0fe46184-7453-4cc6-8e87-6abd5f05dd68" (UID: "0fe46184-7453-4cc6-8e87-6abd5f05dd68"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:15:51 crc kubenswrapper[4943]: I0307 15:15:51.120522 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fe46184-7453-4cc6-8e87-6abd5f05dd68-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "0fe46184-7453-4cc6-8e87-6abd5f05dd68" (UID: "0fe46184-7453-4cc6-8e87-6abd5f05dd68"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:15:51 crc kubenswrapper[4943]: I0307 15:15:51.179347 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0fe46184-7453-4cc6-8e87-6abd5f05dd68-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:51 crc kubenswrapper[4943]: I0307 15:15:51.179393 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/0fe46184-7453-4cc6-8e87-6abd5f05dd68-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:51 crc kubenswrapper[4943]: I0307 15:15:51.179427 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/0fe46184-7453-4cc6-8e87-6abd5f05dd68-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:51 crc kubenswrapper[4943]: I0307 15:15:51.179479 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/0fe46184-7453-4cc6-8e87-6abd5f05dd68-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:51 crc kubenswrapper[4943]: I0307 15:15:51.179496 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fhcn5\" (UniqueName: \"kubernetes.io/projected/0fe46184-7453-4cc6-8e87-6abd5f05dd68-kube-api-access-fhcn5\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:51 crc kubenswrapper[4943]: I0307 15:15:51.179513 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/0fe46184-7453-4cc6-8e87-6abd5f05dd68-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:51 crc kubenswrapper[4943]: I0307 15:15:51.637969 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a67902ef7a79df36058714d34c81f72a146616ed2191e707865638efa1e136ba" Mar 07 15:15:51 crc kubenswrapper[4943]: I0307 15:15:51.638051 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-njf9l" Mar 07 15:15:52 crc kubenswrapper[4943]: I0307 15:15:52.142716 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-ntvzv"] Mar 07 15:15:52 crc kubenswrapper[4943]: E0307 15:15:52.144012 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fe46184-7453-4cc6-8e87-6abd5f05dd68" containerName="swift-ring-rebalance" Mar 07 15:15:52 crc kubenswrapper[4943]: I0307 15:15:52.144126 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fe46184-7453-4cc6-8e87-6abd5f05dd68" containerName="swift-ring-rebalance" Mar 07 15:15:52 crc kubenswrapper[4943]: I0307 15:15:52.144377 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fe46184-7453-4cc6-8e87-6abd5f05dd68" containerName="swift-ring-rebalance" Mar 07 15:15:52 crc kubenswrapper[4943]: I0307 15:15:52.145078 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ntvzv" Mar 07 15:15:52 crc kubenswrapper[4943]: I0307 15:15:52.149673 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:15:52 crc kubenswrapper[4943]: I0307 15:15:52.150072 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:15:52 crc kubenswrapper[4943]: I0307 15:15:52.158407 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-ntvzv"] Mar 07 15:15:52 crc kubenswrapper[4943]: I0307 15:15:52.299447 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1d615eb2-5e1c-46aa-8775-9432362d2bad-scripts\") pod \"swift-ring-rebalance-debug-ntvzv\" (UID: \"1d615eb2-5e1c-46aa-8775-9432362d2bad\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ntvzv" Mar 07 15:15:52 crc kubenswrapper[4943]: I0307 15:15:52.299559 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1d615eb2-5e1c-46aa-8775-9432362d2bad-dispersionconf\") pod \"swift-ring-rebalance-debug-ntvzv\" (UID: \"1d615eb2-5e1c-46aa-8775-9432362d2bad\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ntvzv" Mar 07 15:15:52 crc kubenswrapper[4943]: I0307 15:15:52.299654 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nzlv\" (UniqueName: \"kubernetes.io/projected/1d615eb2-5e1c-46aa-8775-9432362d2bad-kube-api-access-2nzlv\") pod \"swift-ring-rebalance-debug-ntvzv\" (UID: \"1d615eb2-5e1c-46aa-8775-9432362d2bad\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ntvzv" Mar 07 15:15:52 crc kubenswrapper[4943]: I0307 15:15:52.299808 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1d615eb2-5e1c-46aa-8775-9432362d2bad-ring-data-devices\") pod \"swift-ring-rebalance-debug-ntvzv\" (UID: \"1d615eb2-5e1c-46aa-8775-9432362d2bad\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ntvzv" Mar 07 15:15:52 crc kubenswrapper[4943]: I0307 15:15:52.300074 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1d615eb2-5e1c-46aa-8775-9432362d2bad-swiftconf\") pod \"swift-ring-rebalance-debug-ntvzv\" (UID: \"1d615eb2-5e1c-46aa-8775-9432362d2bad\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ntvzv" Mar 07 15:15:52 crc kubenswrapper[4943]: I0307 15:15:52.300242 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1d615eb2-5e1c-46aa-8775-9432362d2bad-etc-swift\") pod \"swift-ring-rebalance-debug-ntvzv\" (UID: \"1d615eb2-5e1c-46aa-8775-9432362d2bad\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ntvzv" Mar 07 15:15:52 crc kubenswrapper[4943]: I0307 15:15:52.401780 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1d615eb2-5e1c-46aa-8775-9432362d2bad-dispersionconf\") pod \"swift-ring-rebalance-debug-ntvzv\" (UID: \"1d615eb2-5e1c-46aa-8775-9432362d2bad\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ntvzv" Mar 07 15:15:52 crc kubenswrapper[4943]: I0307 15:15:52.402558 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nzlv\" (UniqueName: \"kubernetes.io/projected/1d615eb2-5e1c-46aa-8775-9432362d2bad-kube-api-access-2nzlv\") pod \"swift-ring-rebalance-debug-ntvzv\" (UID: \"1d615eb2-5e1c-46aa-8775-9432362d2bad\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ntvzv" Mar 07 15:15:52 crc kubenswrapper[4943]: I0307 15:15:52.402743 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1d615eb2-5e1c-46aa-8775-9432362d2bad-ring-data-devices\") pod \"swift-ring-rebalance-debug-ntvzv\" (UID: \"1d615eb2-5e1c-46aa-8775-9432362d2bad\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ntvzv" Mar 07 15:15:52 crc kubenswrapper[4943]: I0307 15:15:52.402973 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1d615eb2-5e1c-46aa-8775-9432362d2bad-swiftconf\") pod \"swift-ring-rebalance-debug-ntvzv\" (UID: \"1d615eb2-5e1c-46aa-8775-9432362d2bad\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ntvzv" Mar 07 15:15:52 crc kubenswrapper[4943]: I0307 15:15:52.403163 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1d615eb2-5e1c-46aa-8775-9432362d2bad-etc-swift\") pod \"swift-ring-rebalance-debug-ntvzv\" (UID: \"1d615eb2-5e1c-46aa-8775-9432362d2bad\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ntvzv" Mar 07 15:15:52 crc kubenswrapper[4943]: I0307 15:15:52.403757 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1d615eb2-5e1c-46aa-8775-9432362d2bad-scripts\") pod \"swift-ring-rebalance-debug-ntvzv\" (UID: \"1d615eb2-5e1c-46aa-8775-9432362d2bad\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ntvzv" Mar 07 15:15:52 crc kubenswrapper[4943]: I0307 15:15:52.405663 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1d615eb2-5e1c-46aa-8775-9432362d2bad-etc-swift\") pod \"swift-ring-rebalance-debug-ntvzv\" (UID: \"1d615eb2-5e1c-46aa-8775-9432362d2bad\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ntvzv" Mar 07 15:15:52 crc kubenswrapper[4943]: I0307 15:15:52.405885 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1d615eb2-5e1c-46aa-8775-9432362d2bad-scripts\") pod \"swift-ring-rebalance-debug-ntvzv\" (UID: \"1d615eb2-5e1c-46aa-8775-9432362d2bad\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ntvzv" Mar 07 15:15:52 crc kubenswrapper[4943]: I0307 15:15:52.406002 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1d615eb2-5e1c-46aa-8775-9432362d2bad-ring-data-devices\") pod \"swift-ring-rebalance-debug-ntvzv\" (UID: \"1d615eb2-5e1c-46aa-8775-9432362d2bad\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ntvzv" Mar 07 15:15:52 crc kubenswrapper[4943]: I0307 15:15:52.408043 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1d615eb2-5e1c-46aa-8775-9432362d2bad-swiftconf\") pod \"swift-ring-rebalance-debug-ntvzv\" (UID: \"1d615eb2-5e1c-46aa-8775-9432362d2bad\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ntvzv" Mar 07 15:15:52 crc kubenswrapper[4943]: I0307 15:15:52.409455 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1d615eb2-5e1c-46aa-8775-9432362d2bad-dispersionconf\") pod \"swift-ring-rebalance-debug-ntvzv\" (UID: \"1d615eb2-5e1c-46aa-8775-9432362d2bad\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ntvzv" Mar 07 15:15:52 crc kubenswrapper[4943]: I0307 15:15:52.434706 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nzlv\" (UniqueName: \"kubernetes.io/projected/1d615eb2-5e1c-46aa-8775-9432362d2bad-kube-api-access-2nzlv\") pod \"swift-ring-rebalance-debug-ntvzv\" (UID: \"1d615eb2-5e1c-46aa-8775-9432362d2bad\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ntvzv" Mar 07 15:15:52 crc kubenswrapper[4943]: I0307 15:15:52.497481 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ntvzv" Mar 07 15:15:52 crc kubenswrapper[4943]: I0307 15:15:52.769816 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0fe46184-7453-4cc6-8e87-6abd5f05dd68" path="/var/lib/kubelet/pods/0fe46184-7453-4cc6-8e87-6abd5f05dd68/volumes" Mar 07 15:15:52 crc kubenswrapper[4943]: I0307 15:15:52.975793 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-ntvzv"] Mar 07 15:15:53 crc kubenswrapper[4943]: I0307 15:15:53.661270 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ntvzv" event={"ID":"1d615eb2-5e1c-46aa-8775-9432362d2bad","Type":"ContainerStarted","Data":"37428c462b8d4fed4c4147a76f6eff119d556b1a17f576bdc1717aa7b58b2d7a"} Mar 07 15:15:53 crc kubenswrapper[4943]: I0307 15:15:53.661677 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ntvzv" event={"ID":"1d615eb2-5e1c-46aa-8775-9432362d2bad","Type":"ContainerStarted","Data":"d685dca765cee8f00ae1c83dc7d9de043b2f318778f904f1ff701bcef875a45e"} Mar 07 15:15:53 crc kubenswrapper[4943]: I0307 15:15:53.690970 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ntvzv" podStartSLOduration=1.690918133 podStartE2EDuration="1.690918133s" podCreationTimestamp="2026-03-07 15:15:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:15:53.677533316 +0000 UTC m=+2195.629669844" watchObservedRunningTime="2026-03-07 15:15:53.690918133 +0000 UTC m=+2195.643054671" Mar 07 15:15:54 crc kubenswrapper[4943]: I0307 15:15:54.671668 4943 generic.go:334] "Generic (PLEG): container finished" podID="1d615eb2-5e1c-46aa-8775-9432362d2bad" containerID="37428c462b8d4fed4c4147a76f6eff119d556b1a17f576bdc1717aa7b58b2d7a" exitCode=0 Mar 07 15:15:54 crc kubenswrapper[4943]: I0307 15:15:54.671728 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ntvzv" event={"ID":"1d615eb2-5e1c-46aa-8775-9432362d2bad","Type":"ContainerDied","Data":"37428c462b8d4fed4c4147a76f6eff119d556b1a17f576bdc1717aa7b58b2d7a"} Mar 07 15:15:56 crc kubenswrapper[4943]: I0307 15:15:56.067680 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ntvzv" Mar 07 15:15:56 crc kubenswrapper[4943]: I0307 15:15:56.127166 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-ntvzv"] Mar 07 15:15:56 crc kubenswrapper[4943]: I0307 15:15:56.141228 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-ntvzv"] Mar 07 15:15:56 crc kubenswrapper[4943]: I0307 15:15:56.175347 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1d615eb2-5e1c-46aa-8775-9432362d2bad-scripts\") pod \"1d615eb2-5e1c-46aa-8775-9432362d2bad\" (UID: \"1d615eb2-5e1c-46aa-8775-9432362d2bad\") " Mar 07 15:15:56 crc kubenswrapper[4943]: I0307 15:15:56.175461 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1d615eb2-5e1c-46aa-8775-9432362d2bad-ring-data-devices\") pod \"1d615eb2-5e1c-46aa-8775-9432362d2bad\" (UID: \"1d615eb2-5e1c-46aa-8775-9432362d2bad\") " Mar 07 15:15:56 crc kubenswrapper[4943]: I0307 15:15:56.175609 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1d615eb2-5e1c-46aa-8775-9432362d2bad-etc-swift\") pod \"1d615eb2-5e1c-46aa-8775-9432362d2bad\" (UID: \"1d615eb2-5e1c-46aa-8775-9432362d2bad\") " Mar 07 15:15:56 crc kubenswrapper[4943]: I0307 15:15:56.176036 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1d615eb2-5e1c-46aa-8775-9432362d2bad-dispersionconf\") pod \"1d615eb2-5e1c-46aa-8775-9432362d2bad\" (UID: \"1d615eb2-5e1c-46aa-8775-9432362d2bad\") " Mar 07 15:15:56 crc kubenswrapper[4943]: I0307 15:15:56.176137 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2nzlv\" (UniqueName: \"kubernetes.io/projected/1d615eb2-5e1c-46aa-8775-9432362d2bad-kube-api-access-2nzlv\") pod \"1d615eb2-5e1c-46aa-8775-9432362d2bad\" (UID: \"1d615eb2-5e1c-46aa-8775-9432362d2bad\") " Mar 07 15:15:56 crc kubenswrapper[4943]: I0307 15:15:56.176237 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1d615eb2-5e1c-46aa-8775-9432362d2bad-swiftconf\") pod \"1d615eb2-5e1c-46aa-8775-9432362d2bad\" (UID: \"1d615eb2-5e1c-46aa-8775-9432362d2bad\") " Mar 07 15:15:56 crc kubenswrapper[4943]: I0307 15:15:56.176820 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d615eb2-5e1c-46aa-8775-9432362d2bad-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "1d615eb2-5e1c-46aa-8775-9432362d2bad" (UID: "1d615eb2-5e1c-46aa-8775-9432362d2bad"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:15:56 crc kubenswrapper[4943]: I0307 15:15:56.177145 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d615eb2-5e1c-46aa-8775-9432362d2bad-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "1d615eb2-5e1c-46aa-8775-9432362d2bad" (UID: "1d615eb2-5e1c-46aa-8775-9432362d2bad"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:15:56 crc kubenswrapper[4943]: I0307 15:15:56.183685 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d615eb2-5e1c-46aa-8775-9432362d2bad-kube-api-access-2nzlv" (OuterVolumeSpecName: "kube-api-access-2nzlv") pod "1d615eb2-5e1c-46aa-8775-9432362d2bad" (UID: "1d615eb2-5e1c-46aa-8775-9432362d2bad"). InnerVolumeSpecName "kube-api-access-2nzlv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:15:56 crc kubenswrapper[4943]: I0307 15:15:56.197147 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d615eb2-5e1c-46aa-8775-9432362d2bad-scripts" (OuterVolumeSpecName: "scripts") pod "1d615eb2-5e1c-46aa-8775-9432362d2bad" (UID: "1d615eb2-5e1c-46aa-8775-9432362d2bad"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:15:56 crc kubenswrapper[4943]: I0307 15:15:56.208345 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d615eb2-5e1c-46aa-8775-9432362d2bad-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "1d615eb2-5e1c-46aa-8775-9432362d2bad" (UID: "1d615eb2-5e1c-46aa-8775-9432362d2bad"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:15:56 crc kubenswrapper[4943]: I0307 15:15:56.208639 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d615eb2-5e1c-46aa-8775-9432362d2bad-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "1d615eb2-5e1c-46aa-8775-9432362d2bad" (UID: "1d615eb2-5e1c-46aa-8775-9432362d2bad"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:15:56 crc kubenswrapper[4943]: I0307 15:15:56.278006 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1d615eb2-5e1c-46aa-8775-9432362d2bad-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:56 crc kubenswrapper[4943]: I0307 15:15:56.278084 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1d615eb2-5e1c-46aa-8775-9432362d2bad-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:56 crc kubenswrapper[4943]: I0307 15:15:56.278102 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1d615eb2-5e1c-46aa-8775-9432362d2bad-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:56 crc kubenswrapper[4943]: I0307 15:15:56.278117 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1d615eb2-5e1c-46aa-8775-9432362d2bad-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:56 crc kubenswrapper[4943]: I0307 15:15:56.278128 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1d615eb2-5e1c-46aa-8775-9432362d2bad-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:56 crc kubenswrapper[4943]: I0307 15:15:56.278138 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2nzlv\" (UniqueName: \"kubernetes.io/projected/1d615eb2-5e1c-46aa-8775-9432362d2bad-kube-api-access-2nzlv\") on node \"crc\" DevicePath \"\"" Mar 07 15:15:56 crc kubenswrapper[4943]: I0307 15:15:56.699447 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d685dca765cee8f00ae1c83dc7d9de043b2f318778f904f1ff701bcef875a45e" Mar 07 15:15:56 crc kubenswrapper[4943]: I0307 15:15:56.699598 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ntvzv" Mar 07 15:15:56 crc kubenswrapper[4943]: I0307 15:15:56.772152 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d615eb2-5e1c-46aa-8775-9432362d2bad" path="/var/lib/kubelet/pods/1d615eb2-5e1c-46aa-8775-9432362d2bad/volumes" Mar 07 15:15:57 crc kubenswrapper[4943]: I0307 15:15:57.320996 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-fr6j2"] Mar 07 15:15:57 crc kubenswrapper[4943]: E0307 15:15:57.321407 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d615eb2-5e1c-46aa-8775-9432362d2bad" containerName="swift-ring-rebalance" Mar 07 15:15:57 crc kubenswrapper[4943]: I0307 15:15:57.321429 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d615eb2-5e1c-46aa-8775-9432362d2bad" containerName="swift-ring-rebalance" Mar 07 15:15:57 crc kubenswrapper[4943]: I0307 15:15:57.321726 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d615eb2-5e1c-46aa-8775-9432362d2bad" containerName="swift-ring-rebalance" Mar 07 15:15:57 crc kubenswrapper[4943]: I0307 15:15:57.322536 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-fr6j2" Mar 07 15:15:57 crc kubenswrapper[4943]: I0307 15:15:57.325367 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:15:57 crc kubenswrapper[4943]: I0307 15:15:57.325580 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:15:57 crc kubenswrapper[4943]: I0307 15:15:57.338068 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-fr6j2"] Mar 07 15:15:57 crc kubenswrapper[4943]: I0307 15:15:57.395231 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/08e82bc5-1462-440c-9be0-f27f342e5794-etc-swift\") pod \"swift-ring-rebalance-debug-fr6j2\" (UID: \"08e82bc5-1462-440c-9be0-f27f342e5794\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fr6j2" Mar 07 15:15:57 crc kubenswrapper[4943]: I0307 15:15:57.395342 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/08e82bc5-1462-440c-9be0-f27f342e5794-ring-data-devices\") pod \"swift-ring-rebalance-debug-fr6j2\" (UID: \"08e82bc5-1462-440c-9be0-f27f342e5794\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fr6j2" Mar 07 15:15:57 crc kubenswrapper[4943]: I0307 15:15:57.395403 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/08e82bc5-1462-440c-9be0-f27f342e5794-swiftconf\") pod \"swift-ring-rebalance-debug-fr6j2\" (UID: \"08e82bc5-1462-440c-9be0-f27f342e5794\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fr6j2" Mar 07 15:15:57 crc kubenswrapper[4943]: I0307 15:15:57.395441 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fp9t\" (UniqueName: \"kubernetes.io/projected/08e82bc5-1462-440c-9be0-f27f342e5794-kube-api-access-4fp9t\") pod \"swift-ring-rebalance-debug-fr6j2\" (UID: \"08e82bc5-1462-440c-9be0-f27f342e5794\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fr6j2" Mar 07 15:15:57 crc kubenswrapper[4943]: I0307 15:15:57.395472 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/08e82bc5-1462-440c-9be0-f27f342e5794-dispersionconf\") pod \"swift-ring-rebalance-debug-fr6j2\" (UID: \"08e82bc5-1462-440c-9be0-f27f342e5794\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fr6j2" Mar 07 15:15:57 crc kubenswrapper[4943]: I0307 15:15:57.395500 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/08e82bc5-1462-440c-9be0-f27f342e5794-scripts\") pod \"swift-ring-rebalance-debug-fr6j2\" (UID: \"08e82bc5-1462-440c-9be0-f27f342e5794\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fr6j2" Mar 07 15:15:57 crc kubenswrapper[4943]: I0307 15:15:57.497525 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/08e82bc5-1462-440c-9be0-f27f342e5794-scripts\") pod \"swift-ring-rebalance-debug-fr6j2\" (UID: \"08e82bc5-1462-440c-9be0-f27f342e5794\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fr6j2" Mar 07 15:15:57 crc kubenswrapper[4943]: I0307 15:15:57.497723 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/08e82bc5-1462-440c-9be0-f27f342e5794-etc-swift\") pod \"swift-ring-rebalance-debug-fr6j2\" (UID: \"08e82bc5-1462-440c-9be0-f27f342e5794\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fr6j2" Mar 07 15:15:57 crc kubenswrapper[4943]: I0307 15:15:57.497845 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/08e82bc5-1462-440c-9be0-f27f342e5794-ring-data-devices\") pod \"swift-ring-rebalance-debug-fr6j2\" (UID: \"08e82bc5-1462-440c-9be0-f27f342e5794\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fr6j2" Mar 07 15:15:57 crc kubenswrapper[4943]: I0307 15:15:57.497985 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/08e82bc5-1462-440c-9be0-f27f342e5794-swiftconf\") pod \"swift-ring-rebalance-debug-fr6j2\" (UID: \"08e82bc5-1462-440c-9be0-f27f342e5794\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fr6j2" Mar 07 15:15:57 crc kubenswrapper[4943]: I0307 15:15:57.498055 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fp9t\" (UniqueName: \"kubernetes.io/projected/08e82bc5-1462-440c-9be0-f27f342e5794-kube-api-access-4fp9t\") pod \"swift-ring-rebalance-debug-fr6j2\" (UID: \"08e82bc5-1462-440c-9be0-f27f342e5794\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fr6j2" Mar 07 15:15:57 crc kubenswrapper[4943]: I0307 15:15:57.498107 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/08e82bc5-1462-440c-9be0-f27f342e5794-dispersionconf\") pod \"swift-ring-rebalance-debug-fr6j2\" (UID: \"08e82bc5-1462-440c-9be0-f27f342e5794\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fr6j2" Mar 07 15:15:57 crc kubenswrapper[4943]: I0307 15:15:57.498500 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/08e82bc5-1462-440c-9be0-f27f342e5794-etc-swift\") pod \"swift-ring-rebalance-debug-fr6j2\" (UID: \"08e82bc5-1462-440c-9be0-f27f342e5794\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fr6j2" Mar 07 15:15:57 crc kubenswrapper[4943]: I0307 15:15:57.499296 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/08e82bc5-1462-440c-9be0-f27f342e5794-ring-data-devices\") pod \"swift-ring-rebalance-debug-fr6j2\" (UID: \"08e82bc5-1462-440c-9be0-f27f342e5794\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fr6j2" Mar 07 15:15:57 crc kubenswrapper[4943]: I0307 15:15:57.499347 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/08e82bc5-1462-440c-9be0-f27f342e5794-scripts\") pod \"swift-ring-rebalance-debug-fr6j2\" (UID: \"08e82bc5-1462-440c-9be0-f27f342e5794\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fr6j2" Mar 07 15:15:57 crc kubenswrapper[4943]: I0307 15:15:57.509197 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/08e82bc5-1462-440c-9be0-f27f342e5794-swiftconf\") pod \"swift-ring-rebalance-debug-fr6j2\" (UID: \"08e82bc5-1462-440c-9be0-f27f342e5794\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fr6j2" Mar 07 15:15:57 crc kubenswrapper[4943]: I0307 15:15:57.510487 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/08e82bc5-1462-440c-9be0-f27f342e5794-dispersionconf\") pod \"swift-ring-rebalance-debug-fr6j2\" (UID: \"08e82bc5-1462-440c-9be0-f27f342e5794\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fr6j2" Mar 07 15:15:57 crc kubenswrapper[4943]: I0307 15:15:57.520222 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fp9t\" (UniqueName: \"kubernetes.io/projected/08e82bc5-1462-440c-9be0-f27f342e5794-kube-api-access-4fp9t\") pod \"swift-ring-rebalance-debug-fr6j2\" (UID: \"08e82bc5-1462-440c-9be0-f27f342e5794\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fr6j2" Mar 07 15:15:57 crc kubenswrapper[4943]: I0307 15:15:57.660497 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-fr6j2" Mar 07 15:15:57 crc kubenswrapper[4943]: I0307 15:15:57.927756 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-fr6j2"] Mar 07 15:15:58 crc kubenswrapper[4943]: I0307 15:15:58.739542 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-fr6j2" event={"ID":"08e82bc5-1462-440c-9be0-f27f342e5794","Type":"ContainerStarted","Data":"2691dce5a061b11d167965c7e2049be3d1bf60151c7e2349125fc3120468c7db"} Mar 07 15:15:58 crc kubenswrapper[4943]: I0307 15:15:58.739882 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-fr6j2" event={"ID":"08e82bc5-1462-440c-9be0-f27f342e5794","Type":"ContainerStarted","Data":"413300b76204838592935c57600c40af865f7a63590843cb52da8834ec4ab6be"} Mar 07 15:15:58 crc kubenswrapper[4943]: I0307 15:15:58.772040 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-fr6j2" podStartSLOduration=1.772019174 podStartE2EDuration="1.772019174s" podCreationTimestamp="2026-03-07 15:15:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:15:58.768758615 +0000 UTC m=+2200.720895113" watchObservedRunningTime="2026-03-07 15:15:58.772019174 +0000 UTC m=+2200.724155692" Mar 07 15:15:59 crc kubenswrapper[4943]: I0307 15:15:59.757876 4943 generic.go:334] "Generic (PLEG): container finished" podID="08e82bc5-1462-440c-9be0-f27f342e5794" containerID="2691dce5a061b11d167965c7e2049be3d1bf60151c7e2349125fc3120468c7db" exitCode=0 Mar 07 15:15:59 crc kubenswrapper[4943]: I0307 15:15:59.757980 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-fr6j2" event={"ID":"08e82bc5-1462-440c-9be0-f27f342e5794","Type":"ContainerDied","Data":"2691dce5a061b11d167965c7e2049be3d1bf60151c7e2349125fc3120468c7db"} Mar 07 15:16:00 crc kubenswrapper[4943]: I0307 15:16:00.155008 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29548276-5qggb"] Mar 07 15:16:00 crc kubenswrapper[4943]: I0307 15:16:00.156640 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548276-5qggb" Mar 07 15:16:00 crc kubenswrapper[4943]: I0307 15:16:00.166870 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 07 15:16:00 crc kubenswrapper[4943]: I0307 15:16:00.167161 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-qqlk5" Mar 07 15:16:00 crc kubenswrapper[4943]: I0307 15:16:00.168171 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 07 15:16:00 crc kubenswrapper[4943]: I0307 15:16:00.187880 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29548276-5qggb"] Mar 07 15:16:00 crc kubenswrapper[4943]: I0307 15:16:00.246757 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xxh4\" (UniqueName: \"kubernetes.io/projected/c7215a3b-6a23-4b8d-bf59-de153d613605-kube-api-access-2xxh4\") pod \"auto-csr-approver-29548276-5qggb\" (UID: \"c7215a3b-6a23-4b8d-bf59-de153d613605\") " pod="openshift-infra/auto-csr-approver-29548276-5qggb" Mar 07 15:16:00 crc kubenswrapper[4943]: I0307 15:16:00.349082 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xxh4\" (UniqueName: \"kubernetes.io/projected/c7215a3b-6a23-4b8d-bf59-de153d613605-kube-api-access-2xxh4\") pod \"auto-csr-approver-29548276-5qggb\" (UID: \"c7215a3b-6a23-4b8d-bf59-de153d613605\") " pod="openshift-infra/auto-csr-approver-29548276-5qggb" Mar 07 15:16:00 crc kubenswrapper[4943]: I0307 15:16:00.375709 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xxh4\" (UniqueName: \"kubernetes.io/projected/c7215a3b-6a23-4b8d-bf59-de153d613605-kube-api-access-2xxh4\") pod \"auto-csr-approver-29548276-5qggb\" (UID: \"c7215a3b-6a23-4b8d-bf59-de153d613605\") " pod="openshift-infra/auto-csr-approver-29548276-5qggb" Mar 07 15:16:00 crc kubenswrapper[4943]: I0307 15:16:00.500285 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548276-5qggb" Mar 07 15:16:00 crc kubenswrapper[4943]: I0307 15:16:00.966852 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29548276-5qggb"] Mar 07 15:16:00 crc kubenswrapper[4943]: W0307 15:16:00.973776 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc7215a3b_6a23_4b8d_bf59_de153d613605.slice/crio-9957916c1ccd1667eff82b0be3b50feb4e65b1f0bd9b0148e5b6772ba780bab9 WatchSource:0}: Error finding container 9957916c1ccd1667eff82b0be3b50feb4e65b1f0bd9b0148e5b6772ba780bab9: Status 404 returned error can't find the container with id 9957916c1ccd1667eff82b0be3b50feb4e65b1f0bd9b0148e5b6772ba780bab9 Mar 07 15:16:01 crc kubenswrapper[4943]: I0307 15:16:01.068251 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-fr6j2" Mar 07 15:16:01 crc kubenswrapper[4943]: I0307 15:16:01.108757 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-fr6j2"] Mar 07 15:16:01 crc kubenswrapper[4943]: I0307 15:16:01.115748 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-fr6j2"] Mar 07 15:16:01 crc kubenswrapper[4943]: I0307 15:16:01.164405 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/08e82bc5-1462-440c-9be0-f27f342e5794-dispersionconf\") pod \"08e82bc5-1462-440c-9be0-f27f342e5794\" (UID: \"08e82bc5-1462-440c-9be0-f27f342e5794\") " Mar 07 15:16:01 crc kubenswrapper[4943]: I0307 15:16:01.164450 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/08e82bc5-1462-440c-9be0-f27f342e5794-scripts\") pod \"08e82bc5-1462-440c-9be0-f27f342e5794\" (UID: \"08e82bc5-1462-440c-9be0-f27f342e5794\") " Mar 07 15:16:01 crc kubenswrapper[4943]: I0307 15:16:01.164484 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/08e82bc5-1462-440c-9be0-f27f342e5794-ring-data-devices\") pod \"08e82bc5-1462-440c-9be0-f27f342e5794\" (UID: \"08e82bc5-1462-440c-9be0-f27f342e5794\") " Mar 07 15:16:01 crc kubenswrapper[4943]: I0307 15:16:01.164514 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4fp9t\" (UniqueName: \"kubernetes.io/projected/08e82bc5-1462-440c-9be0-f27f342e5794-kube-api-access-4fp9t\") pod \"08e82bc5-1462-440c-9be0-f27f342e5794\" (UID: \"08e82bc5-1462-440c-9be0-f27f342e5794\") " Mar 07 15:16:01 crc kubenswrapper[4943]: I0307 15:16:01.164552 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/08e82bc5-1462-440c-9be0-f27f342e5794-swiftconf\") pod \"08e82bc5-1462-440c-9be0-f27f342e5794\" (UID: \"08e82bc5-1462-440c-9be0-f27f342e5794\") " Mar 07 15:16:01 crc kubenswrapper[4943]: I0307 15:16:01.164659 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/08e82bc5-1462-440c-9be0-f27f342e5794-etc-swift\") pod \"08e82bc5-1462-440c-9be0-f27f342e5794\" (UID: \"08e82bc5-1462-440c-9be0-f27f342e5794\") " Mar 07 15:16:01 crc kubenswrapper[4943]: I0307 15:16:01.165976 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08e82bc5-1462-440c-9be0-f27f342e5794-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "08e82bc5-1462-440c-9be0-f27f342e5794" (UID: "08e82bc5-1462-440c-9be0-f27f342e5794"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:16:01 crc kubenswrapper[4943]: I0307 15:16:01.166022 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08e82bc5-1462-440c-9be0-f27f342e5794-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "08e82bc5-1462-440c-9be0-f27f342e5794" (UID: "08e82bc5-1462-440c-9be0-f27f342e5794"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:16:01 crc kubenswrapper[4943]: I0307 15:16:01.166507 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/08e82bc5-1462-440c-9be0-f27f342e5794-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:01 crc kubenswrapper[4943]: I0307 15:16:01.166523 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/08e82bc5-1462-440c-9be0-f27f342e5794-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:01 crc kubenswrapper[4943]: I0307 15:16:01.170659 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08e82bc5-1462-440c-9be0-f27f342e5794-kube-api-access-4fp9t" (OuterVolumeSpecName: "kube-api-access-4fp9t") pod "08e82bc5-1462-440c-9be0-f27f342e5794" (UID: "08e82bc5-1462-440c-9be0-f27f342e5794"). InnerVolumeSpecName "kube-api-access-4fp9t". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:16:01 crc kubenswrapper[4943]: I0307 15:16:01.185720 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08e82bc5-1462-440c-9be0-f27f342e5794-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "08e82bc5-1462-440c-9be0-f27f342e5794" (UID: "08e82bc5-1462-440c-9be0-f27f342e5794"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:16:01 crc kubenswrapper[4943]: I0307 15:16:01.189233 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08e82bc5-1462-440c-9be0-f27f342e5794-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "08e82bc5-1462-440c-9be0-f27f342e5794" (UID: "08e82bc5-1462-440c-9be0-f27f342e5794"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:16:01 crc kubenswrapper[4943]: I0307 15:16:01.207390 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08e82bc5-1462-440c-9be0-f27f342e5794-scripts" (OuterVolumeSpecName: "scripts") pod "08e82bc5-1462-440c-9be0-f27f342e5794" (UID: "08e82bc5-1462-440c-9be0-f27f342e5794"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:16:01 crc kubenswrapper[4943]: I0307 15:16:01.268541 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/08e82bc5-1462-440c-9be0-f27f342e5794-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:01 crc kubenswrapper[4943]: I0307 15:16:01.268593 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/08e82bc5-1462-440c-9be0-f27f342e5794-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:01 crc kubenswrapper[4943]: I0307 15:16:01.268612 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4fp9t\" (UniqueName: \"kubernetes.io/projected/08e82bc5-1462-440c-9be0-f27f342e5794-kube-api-access-4fp9t\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:01 crc kubenswrapper[4943]: I0307 15:16:01.268635 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/08e82bc5-1462-440c-9be0-f27f342e5794-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:01 crc kubenswrapper[4943]: I0307 15:16:01.781871 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548276-5qggb" event={"ID":"c7215a3b-6a23-4b8d-bf59-de153d613605","Type":"ContainerStarted","Data":"9957916c1ccd1667eff82b0be3b50feb4e65b1f0bd9b0148e5b6772ba780bab9"} Mar 07 15:16:01 crc kubenswrapper[4943]: I0307 15:16:01.784355 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="413300b76204838592935c57600c40af865f7a63590843cb52da8834ec4ab6be" Mar 07 15:16:01 crc kubenswrapper[4943]: I0307 15:16:01.784434 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-fr6j2" Mar 07 15:16:02 crc kubenswrapper[4943]: I0307 15:16:02.307613 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7zw5f"] Mar 07 15:16:02 crc kubenswrapper[4943]: E0307 15:16:02.309486 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08e82bc5-1462-440c-9be0-f27f342e5794" containerName="swift-ring-rebalance" Mar 07 15:16:02 crc kubenswrapper[4943]: I0307 15:16:02.309587 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="08e82bc5-1462-440c-9be0-f27f342e5794" containerName="swift-ring-rebalance" Mar 07 15:16:02 crc kubenswrapper[4943]: I0307 15:16:02.309800 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="08e82bc5-1462-440c-9be0-f27f342e5794" containerName="swift-ring-rebalance" Mar 07 15:16:02 crc kubenswrapper[4943]: I0307 15:16:02.310860 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7zw5f" Mar 07 15:16:02 crc kubenswrapper[4943]: I0307 15:16:02.313633 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:16:02 crc kubenswrapper[4943]: I0307 15:16:02.313874 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:16:02 crc kubenswrapper[4943]: I0307 15:16:02.316305 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7zw5f"] Mar 07 15:16:02 crc kubenswrapper[4943]: I0307 15:16:02.387839 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e65e756f-edb1-4414-b8f7-3122694c6c10-etc-swift\") pod \"swift-ring-rebalance-debug-7zw5f\" (UID: \"e65e756f-edb1-4414-b8f7-3122694c6c10\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7zw5f" Mar 07 15:16:02 crc kubenswrapper[4943]: I0307 15:16:02.387902 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e65e756f-edb1-4414-b8f7-3122694c6c10-dispersionconf\") pod \"swift-ring-rebalance-debug-7zw5f\" (UID: \"e65e756f-edb1-4414-b8f7-3122694c6c10\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7zw5f" Mar 07 15:16:02 crc kubenswrapper[4943]: I0307 15:16:02.387969 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e65e756f-edb1-4414-b8f7-3122694c6c10-ring-data-devices\") pod \"swift-ring-rebalance-debug-7zw5f\" (UID: \"e65e756f-edb1-4414-b8f7-3122694c6c10\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7zw5f" Mar 07 15:16:02 crc kubenswrapper[4943]: I0307 15:16:02.388033 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmvmh\" (UniqueName: \"kubernetes.io/projected/e65e756f-edb1-4414-b8f7-3122694c6c10-kube-api-access-fmvmh\") pod \"swift-ring-rebalance-debug-7zw5f\" (UID: \"e65e756f-edb1-4414-b8f7-3122694c6c10\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7zw5f" Mar 07 15:16:02 crc kubenswrapper[4943]: I0307 15:16:02.388070 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e65e756f-edb1-4414-b8f7-3122694c6c10-swiftconf\") pod \"swift-ring-rebalance-debug-7zw5f\" (UID: \"e65e756f-edb1-4414-b8f7-3122694c6c10\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7zw5f" Mar 07 15:16:02 crc kubenswrapper[4943]: I0307 15:16:02.388105 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e65e756f-edb1-4414-b8f7-3122694c6c10-scripts\") pod \"swift-ring-rebalance-debug-7zw5f\" (UID: \"e65e756f-edb1-4414-b8f7-3122694c6c10\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7zw5f" Mar 07 15:16:02 crc kubenswrapper[4943]: I0307 15:16:02.489425 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e65e756f-edb1-4414-b8f7-3122694c6c10-ring-data-devices\") pod \"swift-ring-rebalance-debug-7zw5f\" (UID: \"e65e756f-edb1-4414-b8f7-3122694c6c10\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7zw5f" Mar 07 15:16:02 crc kubenswrapper[4943]: I0307 15:16:02.489507 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmvmh\" (UniqueName: \"kubernetes.io/projected/e65e756f-edb1-4414-b8f7-3122694c6c10-kube-api-access-fmvmh\") pod \"swift-ring-rebalance-debug-7zw5f\" (UID: \"e65e756f-edb1-4414-b8f7-3122694c6c10\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7zw5f" Mar 07 15:16:02 crc kubenswrapper[4943]: I0307 15:16:02.489535 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e65e756f-edb1-4414-b8f7-3122694c6c10-swiftconf\") pod \"swift-ring-rebalance-debug-7zw5f\" (UID: \"e65e756f-edb1-4414-b8f7-3122694c6c10\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7zw5f" Mar 07 15:16:02 crc kubenswrapper[4943]: I0307 15:16:02.489561 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e65e756f-edb1-4414-b8f7-3122694c6c10-scripts\") pod \"swift-ring-rebalance-debug-7zw5f\" (UID: \"e65e756f-edb1-4414-b8f7-3122694c6c10\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7zw5f" Mar 07 15:16:02 crc kubenswrapper[4943]: I0307 15:16:02.489602 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e65e756f-edb1-4414-b8f7-3122694c6c10-etc-swift\") pod \"swift-ring-rebalance-debug-7zw5f\" (UID: \"e65e756f-edb1-4414-b8f7-3122694c6c10\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7zw5f" Mar 07 15:16:02 crc kubenswrapper[4943]: I0307 15:16:02.489624 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e65e756f-edb1-4414-b8f7-3122694c6c10-dispersionconf\") pod \"swift-ring-rebalance-debug-7zw5f\" (UID: \"e65e756f-edb1-4414-b8f7-3122694c6c10\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7zw5f" Mar 07 15:16:02 crc kubenswrapper[4943]: I0307 15:16:02.490188 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e65e756f-edb1-4414-b8f7-3122694c6c10-etc-swift\") pod \"swift-ring-rebalance-debug-7zw5f\" (UID: \"e65e756f-edb1-4414-b8f7-3122694c6c10\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7zw5f" Mar 07 15:16:02 crc kubenswrapper[4943]: I0307 15:16:02.490584 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e65e756f-edb1-4414-b8f7-3122694c6c10-ring-data-devices\") pod \"swift-ring-rebalance-debug-7zw5f\" (UID: \"e65e756f-edb1-4414-b8f7-3122694c6c10\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7zw5f" Mar 07 15:16:02 crc kubenswrapper[4943]: I0307 15:16:02.494114 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e65e756f-edb1-4414-b8f7-3122694c6c10-scripts\") pod \"swift-ring-rebalance-debug-7zw5f\" (UID: \"e65e756f-edb1-4414-b8f7-3122694c6c10\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7zw5f" Mar 07 15:16:02 crc kubenswrapper[4943]: I0307 15:16:02.494873 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e65e756f-edb1-4414-b8f7-3122694c6c10-swiftconf\") pod \"swift-ring-rebalance-debug-7zw5f\" (UID: \"e65e756f-edb1-4414-b8f7-3122694c6c10\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7zw5f" Mar 07 15:16:02 crc kubenswrapper[4943]: I0307 15:16:02.496417 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e65e756f-edb1-4414-b8f7-3122694c6c10-dispersionconf\") pod \"swift-ring-rebalance-debug-7zw5f\" (UID: \"e65e756f-edb1-4414-b8f7-3122694c6c10\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7zw5f" Mar 07 15:16:02 crc kubenswrapper[4943]: I0307 15:16:02.506159 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmvmh\" (UniqueName: \"kubernetes.io/projected/e65e756f-edb1-4414-b8f7-3122694c6c10-kube-api-access-fmvmh\") pod \"swift-ring-rebalance-debug-7zw5f\" (UID: \"e65e756f-edb1-4414-b8f7-3122694c6c10\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7zw5f" Mar 07 15:16:02 crc kubenswrapper[4943]: I0307 15:16:02.629982 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7zw5f" Mar 07 15:16:02 crc kubenswrapper[4943]: I0307 15:16:02.764589 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08e82bc5-1462-440c-9be0-f27f342e5794" path="/var/lib/kubelet/pods/08e82bc5-1462-440c-9be0-f27f342e5794/volumes" Mar 07 15:16:02 crc kubenswrapper[4943]: I0307 15:16:02.797879 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548276-5qggb" event={"ID":"c7215a3b-6a23-4b8d-bf59-de153d613605","Type":"ContainerStarted","Data":"a736868f2888b612a98b97aaeb4cf05314e6e0f58df2907209bfa17f89dc7b4a"} Mar 07 15:16:02 crc kubenswrapper[4943]: I0307 15:16:02.812726 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29548276-5qggb" podStartSLOduration=1.470593831 podStartE2EDuration="2.812710899s" podCreationTimestamp="2026-03-07 15:16:00 +0000 UTC" firstStartedPulling="2026-03-07 15:16:00.976377874 +0000 UTC m=+2202.928514412" lastFinishedPulling="2026-03-07 15:16:02.318494962 +0000 UTC m=+2204.270631480" observedRunningTime="2026-03-07 15:16:02.810953966 +0000 UTC m=+2204.763090484" watchObservedRunningTime="2026-03-07 15:16:02.812710899 +0000 UTC m=+2204.764847397" Mar 07 15:16:02 crc kubenswrapper[4943]: I0307 15:16:02.885129 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7zw5f"] Mar 07 15:16:02 crc kubenswrapper[4943]: W0307 15:16:02.890094 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode65e756f_edb1_4414_b8f7_3122694c6c10.slice/crio-be89aa3758df91319080f9a1446aaebf0b6bed47db4f39e3b47e9b244d295f6b WatchSource:0}: Error finding container be89aa3758df91319080f9a1446aaebf0b6bed47db4f39e3b47e9b244d295f6b: Status 404 returned error can't find the container with id be89aa3758df91319080f9a1446aaebf0b6bed47db4f39e3b47e9b244d295f6b Mar 07 15:16:03 crc kubenswrapper[4943]: I0307 15:16:03.816073 4943 generic.go:334] "Generic (PLEG): container finished" podID="c7215a3b-6a23-4b8d-bf59-de153d613605" containerID="a736868f2888b612a98b97aaeb4cf05314e6e0f58df2907209bfa17f89dc7b4a" exitCode=0 Mar 07 15:16:03 crc kubenswrapper[4943]: I0307 15:16:03.816166 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548276-5qggb" event={"ID":"c7215a3b-6a23-4b8d-bf59-de153d613605","Type":"ContainerDied","Data":"a736868f2888b612a98b97aaeb4cf05314e6e0f58df2907209bfa17f89dc7b4a"} Mar 07 15:16:03 crc kubenswrapper[4943]: I0307 15:16:03.822618 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7zw5f" event={"ID":"e65e756f-edb1-4414-b8f7-3122694c6c10","Type":"ContainerStarted","Data":"96c5d176d1e4797a95a65693243a9b75a0486096c1ea64487538635fcd6f3a37"} Mar 07 15:16:03 crc kubenswrapper[4943]: I0307 15:16:03.822754 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7zw5f" event={"ID":"e65e756f-edb1-4414-b8f7-3122694c6c10","Type":"ContainerStarted","Data":"be89aa3758df91319080f9a1446aaebf0b6bed47db4f39e3b47e9b244d295f6b"} Mar 07 15:16:03 crc kubenswrapper[4943]: I0307 15:16:03.873518 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7zw5f" podStartSLOduration=1.873489102 podStartE2EDuration="1.873489102s" podCreationTimestamp="2026-03-07 15:16:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:16:03.866589714 +0000 UTC m=+2205.818726272" watchObservedRunningTime="2026-03-07 15:16:03.873489102 +0000 UTC m=+2205.825625640" Mar 07 15:16:04 crc kubenswrapper[4943]: I0307 15:16:04.838215 4943 generic.go:334] "Generic (PLEG): container finished" podID="e65e756f-edb1-4414-b8f7-3122694c6c10" containerID="96c5d176d1e4797a95a65693243a9b75a0486096c1ea64487538635fcd6f3a37" exitCode=0 Mar 07 15:16:04 crc kubenswrapper[4943]: I0307 15:16:04.838322 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7zw5f" event={"ID":"e65e756f-edb1-4414-b8f7-3122694c6c10","Type":"ContainerDied","Data":"96c5d176d1e4797a95a65693243a9b75a0486096c1ea64487538635fcd6f3a37"} Mar 07 15:16:05 crc kubenswrapper[4943]: I0307 15:16:05.218030 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548276-5qggb" Mar 07 15:16:05 crc kubenswrapper[4943]: I0307 15:16:05.336116 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2xxh4\" (UniqueName: \"kubernetes.io/projected/c7215a3b-6a23-4b8d-bf59-de153d613605-kube-api-access-2xxh4\") pod \"c7215a3b-6a23-4b8d-bf59-de153d613605\" (UID: \"c7215a3b-6a23-4b8d-bf59-de153d613605\") " Mar 07 15:16:05 crc kubenswrapper[4943]: I0307 15:16:05.348274 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7215a3b-6a23-4b8d-bf59-de153d613605-kube-api-access-2xxh4" (OuterVolumeSpecName: "kube-api-access-2xxh4") pod "c7215a3b-6a23-4b8d-bf59-de153d613605" (UID: "c7215a3b-6a23-4b8d-bf59-de153d613605"). InnerVolumeSpecName "kube-api-access-2xxh4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:16:05 crc kubenswrapper[4943]: I0307 15:16:05.438198 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2xxh4\" (UniqueName: \"kubernetes.io/projected/c7215a3b-6a23-4b8d-bf59-de153d613605-kube-api-access-2xxh4\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:05 crc kubenswrapper[4943]: I0307 15:16:05.856267 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548276-5qggb" event={"ID":"c7215a3b-6a23-4b8d-bf59-de153d613605","Type":"ContainerDied","Data":"9957916c1ccd1667eff82b0be3b50feb4e65b1f0bd9b0148e5b6772ba780bab9"} Mar 07 15:16:05 crc kubenswrapper[4943]: I0307 15:16:05.856327 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548276-5qggb" Mar 07 15:16:05 crc kubenswrapper[4943]: I0307 15:16:05.856349 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9957916c1ccd1667eff82b0be3b50feb4e65b1f0bd9b0148e5b6772ba780bab9" Mar 07 15:16:05 crc kubenswrapper[4943]: I0307 15:16:05.903578 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29548270-v9j8n"] Mar 07 15:16:05 crc kubenswrapper[4943]: I0307 15:16:05.911375 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29548270-v9j8n"] Mar 07 15:16:06 crc kubenswrapper[4943]: I0307 15:16:06.073965 4943 patch_prober.go:28] interesting pod/machine-config-daemon-cpdmm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 07 15:16:06 crc kubenswrapper[4943]: I0307 15:16:06.074037 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 07 15:16:06 crc kubenswrapper[4943]: I0307 15:16:06.164669 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7zw5f" Mar 07 15:16:06 crc kubenswrapper[4943]: I0307 15:16:06.212798 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7zw5f"] Mar 07 15:16:06 crc kubenswrapper[4943]: I0307 15:16:06.218193 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7zw5f"] Mar 07 15:16:06 crc kubenswrapper[4943]: I0307 15:16:06.249291 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e65e756f-edb1-4414-b8f7-3122694c6c10-ring-data-devices\") pod \"e65e756f-edb1-4414-b8f7-3122694c6c10\" (UID: \"e65e756f-edb1-4414-b8f7-3122694c6c10\") " Mar 07 15:16:06 crc kubenswrapper[4943]: I0307 15:16:06.249364 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fmvmh\" (UniqueName: \"kubernetes.io/projected/e65e756f-edb1-4414-b8f7-3122694c6c10-kube-api-access-fmvmh\") pod \"e65e756f-edb1-4414-b8f7-3122694c6c10\" (UID: \"e65e756f-edb1-4414-b8f7-3122694c6c10\") " Mar 07 15:16:06 crc kubenswrapper[4943]: I0307 15:16:06.249433 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e65e756f-edb1-4414-b8f7-3122694c6c10-swiftconf\") pod \"e65e756f-edb1-4414-b8f7-3122694c6c10\" (UID: \"e65e756f-edb1-4414-b8f7-3122694c6c10\") " Mar 07 15:16:06 crc kubenswrapper[4943]: I0307 15:16:06.249530 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e65e756f-edb1-4414-b8f7-3122694c6c10-dispersionconf\") pod \"e65e756f-edb1-4414-b8f7-3122694c6c10\" (UID: \"e65e756f-edb1-4414-b8f7-3122694c6c10\") " Mar 07 15:16:06 crc kubenswrapper[4943]: I0307 15:16:06.249568 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e65e756f-edb1-4414-b8f7-3122694c6c10-scripts\") pod \"e65e756f-edb1-4414-b8f7-3122694c6c10\" (UID: \"e65e756f-edb1-4414-b8f7-3122694c6c10\") " Mar 07 15:16:06 crc kubenswrapper[4943]: I0307 15:16:06.249590 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e65e756f-edb1-4414-b8f7-3122694c6c10-etc-swift\") pod \"e65e756f-edb1-4414-b8f7-3122694c6c10\" (UID: \"e65e756f-edb1-4414-b8f7-3122694c6c10\") " Mar 07 15:16:06 crc kubenswrapper[4943]: I0307 15:16:06.251247 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e65e756f-edb1-4414-b8f7-3122694c6c10-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "e65e756f-edb1-4414-b8f7-3122694c6c10" (UID: "e65e756f-edb1-4414-b8f7-3122694c6c10"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:16:06 crc kubenswrapper[4943]: I0307 15:16:06.251408 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e65e756f-edb1-4414-b8f7-3122694c6c10-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "e65e756f-edb1-4414-b8f7-3122694c6c10" (UID: "e65e756f-edb1-4414-b8f7-3122694c6c10"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:16:06 crc kubenswrapper[4943]: I0307 15:16:06.262143 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e65e756f-edb1-4414-b8f7-3122694c6c10-kube-api-access-fmvmh" (OuterVolumeSpecName: "kube-api-access-fmvmh") pod "e65e756f-edb1-4414-b8f7-3122694c6c10" (UID: "e65e756f-edb1-4414-b8f7-3122694c6c10"). InnerVolumeSpecName "kube-api-access-fmvmh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:16:06 crc kubenswrapper[4943]: I0307 15:16:06.270533 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e65e756f-edb1-4414-b8f7-3122694c6c10-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "e65e756f-edb1-4414-b8f7-3122694c6c10" (UID: "e65e756f-edb1-4414-b8f7-3122694c6c10"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:16:06 crc kubenswrapper[4943]: I0307 15:16:06.273183 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e65e756f-edb1-4414-b8f7-3122694c6c10-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "e65e756f-edb1-4414-b8f7-3122694c6c10" (UID: "e65e756f-edb1-4414-b8f7-3122694c6c10"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:16:06 crc kubenswrapper[4943]: I0307 15:16:06.275543 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e65e756f-edb1-4414-b8f7-3122694c6c10-scripts" (OuterVolumeSpecName: "scripts") pod "e65e756f-edb1-4414-b8f7-3122694c6c10" (UID: "e65e756f-edb1-4414-b8f7-3122694c6c10"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:16:06 crc kubenswrapper[4943]: I0307 15:16:06.352027 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e65e756f-edb1-4414-b8f7-3122694c6c10-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:06 crc kubenswrapper[4943]: I0307 15:16:06.352073 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e65e756f-edb1-4414-b8f7-3122694c6c10-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:06 crc kubenswrapper[4943]: I0307 15:16:06.352092 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e65e756f-edb1-4414-b8f7-3122694c6c10-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:06 crc kubenswrapper[4943]: I0307 15:16:06.352110 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e65e756f-edb1-4414-b8f7-3122694c6c10-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:06 crc kubenswrapper[4943]: I0307 15:16:06.352130 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fmvmh\" (UniqueName: \"kubernetes.io/projected/e65e756f-edb1-4414-b8f7-3122694c6c10-kube-api-access-fmvmh\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:06 crc kubenswrapper[4943]: I0307 15:16:06.352193 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e65e756f-edb1-4414-b8f7-3122694c6c10-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:06 crc kubenswrapper[4943]: I0307 15:16:06.774679 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62e0e72a-7faf-4dbb-86d7-cbb756a90d63" path="/var/lib/kubelet/pods/62e0e72a-7faf-4dbb-86d7-cbb756a90d63/volumes" Mar 07 15:16:06 crc kubenswrapper[4943]: I0307 15:16:06.776374 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e65e756f-edb1-4414-b8f7-3122694c6c10" path="/var/lib/kubelet/pods/e65e756f-edb1-4414-b8f7-3122694c6c10/volumes" Mar 07 15:16:06 crc kubenswrapper[4943]: I0307 15:16:06.870961 4943 scope.go:117] "RemoveContainer" containerID="96c5d176d1e4797a95a65693243a9b75a0486096c1ea64487538635fcd6f3a37" Mar 07 15:16:06 crc kubenswrapper[4943]: I0307 15:16:06.871010 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7zw5f" Mar 07 15:16:07 crc kubenswrapper[4943]: I0307 15:16:07.385582 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7xw87"] Mar 07 15:16:07 crc kubenswrapper[4943]: E0307 15:16:07.386026 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e65e756f-edb1-4414-b8f7-3122694c6c10" containerName="swift-ring-rebalance" Mar 07 15:16:07 crc kubenswrapper[4943]: I0307 15:16:07.386051 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="e65e756f-edb1-4414-b8f7-3122694c6c10" containerName="swift-ring-rebalance" Mar 07 15:16:07 crc kubenswrapper[4943]: E0307 15:16:07.386089 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7215a3b-6a23-4b8d-bf59-de153d613605" containerName="oc" Mar 07 15:16:07 crc kubenswrapper[4943]: I0307 15:16:07.386101 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7215a3b-6a23-4b8d-bf59-de153d613605" containerName="oc" Mar 07 15:16:07 crc kubenswrapper[4943]: I0307 15:16:07.386323 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="e65e756f-edb1-4414-b8f7-3122694c6c10" containerName="swift-ring-rebalance" Mar 07 15:16:07 crc kubenswrapper[4943]: I0307 15:16:07.386343 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7215a3b-6a23-4b8d-bf59-de153d613605" containerName="oc" Mar 07 15:16:07 crc kubenswrapper[4943]: I0307 15:16:07.386977 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7xw87" Mar 07 15:16:07 crc kubenswrapper[4943]: I0307 15:16:07.390245 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:16:07 crc kubenswrapper[4943]: I0307 15:16:07.391038 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:16:07 crc kubenswrapper[4943]: I0307 15:16:07.401646 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7xw87"] Mar 07 15:16:07 crc kubenswrapper[4943]: I0307 15:16:07.473464 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54lcx\" (UniqueName: \"kubernetes.io/projected/96e83e83-5c00-48a4-b180-d170d4cff072-kube-api-access-54lcx\") pod \"swift-ring-rebalance-debug-7xw87\" (UID: \"96e83e83-5c00-48a4-b180-d170d4cff072\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7xw87" Mar 07 15:16:07 crc kubenswrapper[4943]: I0307 15:16:07.473764 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/96e83e83-5c00-48a4-b180-d170d4cff072-etc-swift\") pod \"swift-ring-rebalance-debug-7xw87\" (UID: \"96e83e83-5c00-48a4-b180-d170d4cff072\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7xw87" Mar 07 15:16:07 crc kubenswrapper[4943]: I0307 15:16:07.473816 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/96e83e83-5c00-48a4-b180-d170d4cff072-dispersionconf\") pod \"swift-ring-rebalance-debug-7xw87\" (UID: \"96e83e83-5c00-48a4-b180-d170d4cff072\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7xw87" Mar 07 15:16:07 crc kubenswrapper[4943]: I0307 15:16:07.473866 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/96e83e83-5c00-48a4-b180-d170d4cff072-scripts\") pod \"swift-ring-rebalance-debug-7xw87\" (UID: \"96e83e83-5c00-48a4-b180-d170d4cff072\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7xw87" Mar 07 15:16:07 crc kubenswrapper[4943]: I0307 15:16:07.473938 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/96e83e83-5c00-48a4-b180-d170d4cff072-swiftconf\") pod \"swift-ring-rebalance-debug-7xw87\" (UID: \"96e83e83-5c00-48a4-b180-d170d4cff072\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7xw87" Mar 07 15:16:07 crc kubenswrapper[4943]: I0307 15:16:07.474215 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/96e83e83-5c00-48a4-b180-d170d4cff072-ring-data-devices\") pod \"swift-ring-rebalance-debug-7xw87\" (UID: \"96e83e83-5c00-48a4-b180-d170d4cff072\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7xw87" Mar 07 15:16:07 crc kubenswrapper[4943]: I0307 15:16:07.575688 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/96e83e83-5c00-48a4-b180-d170d4cff072-dispersionconf\") pod \"swift-ring-rebalance-debug-7xw87\" (UID: \"96e83e83-5c00-48a4-b180-d170d4cff072\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7xw87" Mar 07 15:16:07 crc kubenswrapper[4943]: I0307 15:16:07.575824 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/96e83e83-5c00-48a4-b180-d170d4cff072-scripts\") pod \"swift-ring-rebalance-debug-7xw87\" (UID: \"96e83e83-5c00-48a4-b180-d170d4cff072\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7xw87" Mar 07 15:16:07 crc kubenswrapper[4943]: I0307 15:16:07.575911 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/96e83e83-5c00-48a4-b180-d170d4cff072-swiftconf\") pod \"swift-ring-rebalance-debug-7xw87\" (UID: \"96e83e83-5c00-48a4-b180-d170d4cff072\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7xw87" Mar 07 15:16:07 crc kubenswrapper[4943]: I0307 15:16:07.576033 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/96e83e83-5c00-48a4-b180-d170d4cff072-ring-data-devices\") pod \"swift-ring-rebalance-debug-7xw87\" (UID: \"96e83e83-5c00-48a4-b180-d170d4cff072\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7xw87" Mar 07 15:16:07 crc kubenswrapper[4943]: I0307 15:16:07.576080 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54lcx\" (UniqueName: \"kubernetes.io/projected/96e83e83-5c00-48a4-b180-d170d4cff072-kube-api-access-54lcx\") pod \"swift-ring-rebalance-debug-7xw87\" (UID: \"96e83e83-5c00-48a4-b180-d170d4cff072\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7xw87" Mar 07 15:16:07 crc kubenswrapper[4943]: I0307 15:16:07.576130 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/96e83e83-5c00-48a4-b180-d170d4cff072-etc-swift\") pod \"swift-ring-rebalance-debug-7xw87\" (UID: \"96e83e83-5c00-48a4-b180-d170d4cff072\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7xw87" Mar 07 15:16:07 crc kubenswrapper[4943]: I0307 15:16:07.576975 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/96e83e83-5c00-48a4-b180-d170d4cff072-etc-swift\") pod \"swift-ring-rebalance-debug-7xw87\" (UID: \"96e83e83-5c00-48a4-b180-d170d4cff072\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7xw87" Mar 07 15:16:07 crc kubenswrapper[4943]: I0307 15:16:07.577543 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/96e83e83-5c00-48a4-b180-d170d4cff072-ring-data-devices\") pod \"swift-ring-rebalance-debug-7xw87\" (UID: \"96e83e83-5c00-48a4-b180-d170d4cff072\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7xw87" Mar 07 15:16:07 crc kubenswrapper[4943]: I0307 15:16:07.577897 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/96e83e83-5c00-48a4-b180-d170d4cff072-scripts\") pod \"swift-ring-rebalance-debug-7xw87\" (UID: \"96e83e83-5c00-48a4-b180-d170d4cff072\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7xw87" Mar 07 15:16:07 crc kubenswrapper[4943]: I0307 15:16:07.582735 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/96e83e83-5c00-48a4-b180-d170d4cff072-dispersionconf\") pod \"swift-ring-rebalance-debug-7xw87\" (UID: \"96e83e83-5c00-48a4-b180-d170d4cff072\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7xw87" Mar 07 15:16:07 crc kubenswrapper[4943]: I0307 15:16:07.583849 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/96e83e83-5c00-48a4-b180-d170d4cff072-swiftconf\") pod \"swift-ring-rebalance-debug-7xw87\" (UID: \"96e83e83-5c00-48a4-b180-d170d4cff072\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7xw87" Mar 07 15:16:07 crc kubenswrapper[4943]: I0307 15:16:07.593071 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54lcx\" (UniqueName: \"kubernetes.io/projected/96e83e83-5c00-48a4-b180-d170d4cff072-kube-api-access-54lcx\") pod \"swift-ring-rebalance-debug-7xw87\" (UID: \"96e83e83-5c00-48a4-b180-d170d4cff072\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7xw87" Mar 07 15:16:07 crc kubenswrapper[4943]: I0307 15:16:07.714906 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7xw87" Mar 07 15:16:08 crc kubenswrapper[4943]: I0307 15:16:08.006701 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7xw87"] Mar 07 15:16:08 crc kubenswrapper[4943]: W0307 15:16:08.013203 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod96e83e83_5c00_48a4_b180_d170d4cff072.slice/crio-6290507bcdd8f672072e6806a84ed090d4bd26eaf946da43460f0a4d43ca7e1d WatchSource:0}: Error finding container 6290507bcdd8f672072e6806a84ed090d4bd26eaf946da43460f0a4d43ca7e1d: Status 404 returned error can't find the container with id 6290507bcdd8f672072e6806a84ed090d4bd26eaf946da43460f0a4d43ca7e1d Mar 07 15:16:08 crc kubenswrapper[4943]: I0307 15:16:08.899754 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7xw87" event={"ID":"96e83e83-5c00-48a4-b180-d170d4cff072","Type":"ContainerStarted","Data":"8769f5fcfacebac78a23380cbd6f9ea0f53a3990a721ef60a5f080181a664f5f"} Mar 07 15:16:08 crc kubenswrapper[4943]: I0307 15:16:08.899823 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7xw87" event={"ID":"96e83e83-5c00-48a4-b180-d170d4cff072","Type":"ContainerStarted","Data":"6290507bcdd8f672072e6806a84ed090d4bd26eaf946da43460f0a4d43ca7e1d"} Mar 07 15:16:08 crc kubenswrapper[4943]: I0307 15:16:08.927130 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7xw87" podStartSLOduration=1.9270793510000002 podStartE2EDuration="1.927079351s" podCreationTimestamp="2026-03-07 15:16:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:16:08.923321169 +0000 UTC m=+2210.875457677" watchObservedRunningTime="2026-03-07 15:16:08.927079351 +0000 UTC m=+2210.879215869" Mar 07 15:16:09 crc kubenswrapper[4943]: I0307 15:16:09.912848 4943 generic.go:334] "Generic (PLEG): container finished" podID="96e83e83-5c00-48a4-b180-d170d4cff072" containerID="8769f5fcfacebac78a23380cbd6f9ea0f53a3990a721ef60a5f080181a664f5f" exitCode=0 Mar 07 15:16:09 crc kubenswrapper[4943]: I0307 15:16:09.912899 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7xw87" event={"ID":"96e83e83-5c00-48a4-b180-d170d4cff072","Type":"ContainerDied","Data":"8769f5fcfacebac78a23380cbd6f9ea0f53a3990a721ef60a5f080181a664f5f"} Mar 07 15:16:11 crc kubenswrapper[4943]: I0307 15:16:11.339572 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7xw87" Mar 07 15:16:11 crc kubenswrapper[4943]: I0307 15:16:11.390438 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7xw87"] Mar 07 15:16:11 crc kubenswrapper[4943]: I0307 15:16:11.402308 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7xw87"] Mar 07 15:16:11 crc kubenswrapper[4943]: I0307 15:16:11.450963 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-54lcx\" (UniqueName: \"kubernetes.io/projected/96e83e83-5c00-48a4-b180-d170d4cff072-kube-api-access-54lcx\") pod \"96e83e83-5c00-48a4-b180-d170d4cff072\" (UID: \"96e83e83-5c00-48a4-b180-d170d4cff072\") " Mar 07 15:16:11 crc kubenswrapper[4943]: I0307 15:16:11.451064 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/96e83e83-5c00-48a4-b180-d170d4cff072-ring-data-devices\") pod \"96e83e83-5c00-48a4-b180-d170d4cff072\" (UID: \"96e83e83-5c00-48a4-b180-d170d4cff072\") " Mar 07 15:16:11 crc kubenswrapper[4943]: I0307 15:16:11.451139 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/96e83e83-5c00-48a4-b180-d170d4cff072-etc-swift\") pod \"96e83e83-5c00-48a4-b180-d170d4cff072\" (UID: \"96e83e83-5c00-48a4-b180-d170d4cff072\") " Mar 07 15:16:11 crc kubenswrapper[4943]: I0307 15:16:11.451183 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/96e83e83-5c00-48a4-b180-d170d4cff072-dispersionconf\") pod \"96e83e83-5c00-48a4-b180-d170d4cff072\" (UID: \"96e83e83-5c00-48a4-b180-d170d4cff072\") " Mar 07 15:16:11 crc kubenswrapper[4943]: I0307 15:16:11.451247 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/96e83e83-5c00-48a4-b180-d170d4cff072-swiftconf\") pod \"96e83e83-5c00-48a4-b180-d170d4cff072\" (UID: \"96e83e83-5c00-48a4-b180-d170d4cff072\") " Mar 07 15:16:11 crc kubenswrapper[4943]: I0307 15:16:11.451293 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/96e83e83-5c00-48a4-b180-d170d4cff072-scripts\") pod \"96e83e83-5c00-48a4-b180-d170d4cff072\" (UID: \"96e83e83-5c00-48a4-b180-d170d4cff072\") " Mar 07 15:16:11 crc kubenswrapper[4943]: I0307 15:16:11.452030 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96e83e83-5c00-48a4-b180-d170d4cff072-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "96e83e83-5c00-48a4-b180-d170d4cff072" (UID: "96e83e83-5c00-48a4-b180-d170d4cff072"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:16:11 crc kubenswrapper[4943]: I0307 15:16:11.452382 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96e83e83-5c00-48a4-b180-d170d4cff072-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "96e83e83-5c00-48a4-b180-d170d4cff072" (UID: "96e83e83-5c00-48a4-b180-d170d4cff072"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:16:11 crc kubenswrapper[4943]: I0307 15:16:11.456676 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96e83e83-5c00-48a4-b180-d170d4cff072-kube-api-access-54lcx" (OuterVolumeSpecName: "kube-api-access-54lcx") pod "96e83e83-5c00-48a4-b180-d170d4cff072" (UID: "96e83e83-5c00-48a4-b180-d170d4cff072"). InnerVolumeSpecName "kube-api-access-54lcx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:16:11 crc kubenswrapper[4943]: I0307 15:16:11.478998 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96e83e83-5c00-48a4-b180-d170d4cff072-scripts" (OuterVolumeSpecName: "scripts") pod "96e83e83-5c00-48a4-b180-d170d4cff072" (UID: "96e83e83-5c00-48a4-b180-d170d4cff072"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:16:11 crc kubenswrapper[4943]: I0307 15:16:11.497276 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96e83e83-5c00-48a4-b180-d170d4cff072-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "96e83e83-5c00-48a4-b180-d170d4cff072" (UID: "96e83e83-5c00-48a4-b180-d170d4cff072"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:16:11 crc kubenswrapper[4943]: I0307 15:16:11.497672 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96e83e83-5c00-48a4-b180-d170d4cff072-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "96e83e83-5c00-48a4-b180-d170d4cff072" (UID: "96e83e83-5c00-48a4-b180-d170d4cff072"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:16:11 crc kubenswrapper[4943]: I0307 15:16:11.553321 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/96e83e83-5c00-48a4-b180-d170d4cff072-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:11 crc kubenswrapper[4943]: I0307 15:16:11.553372 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-54lcx\" (UniqueName: \"kubernetes.io/projected/96e83e83-5c00-48a4-b180-d170d4cff072-kube-api-access-54lcx\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:11 crc kubenswrapper[4943]: I0307 15:16:11.553385 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/96e83e83-5c00-48a4-b180-d170d4cff072-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:11 crc kubenswrapper[4943]: I0307 15:16:11.553394 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/96e83e83-5c00-48a4-b180-d170d4cff072-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:11 crc kubenswrapper[4943]: I0307 15:16:11.553403 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/96e83e83-5c00-48a4-b180-d170d4cff072-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:11 crc kubenswrapper[4943]: I0307 15:16:11.553410 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/96e83e83-5c00-48a4-b180-d170d4cff072-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:11 crc kubenswrapper[4943]: I0307 15:16:11.938339 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6290507bcdd8f672072e6806a84ed090d4bd26eaf946da43460f0a4d43ca7e1d" Mar 07 15:16:11 crc kubenswrapper[4943]: I0307 15:16:11.938473 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7xw87" Mar 07 15:16:12 crc kubenswrapper[4943]: I0307 15:16:12.535062 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-672r7"] Mar 07 15:16:12 crc kubenswrapper[4943]: E0307 15:16:12.535782 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96e83e83-5c00-48a4-b180-d170d4cff072" containerName="swift-ring-rebalance" Mar 07 15:16:12 crc kubenswrapper[4943]: I0307 15:16:12.535805 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="96e83e83-5c00-48a4-b180-d170d4cff072" containerName="swift-ring-rebalance" Mar 07 15:16:12 crc kubenswrapper[4943]: I0307 15:16:12.536170 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="96e83e83-5c00-48a4-b180-d170d4cff072" containerName="swift-ring-rebalance" Mar 07 15:16:12 crc kubenswrapper[4943]: I0307 15:16:12.536975 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-672r7" Mar 07 15:16:12 crc kubenswrapper[4943]: I0307 15:16:12.542723 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:16:12 crc kubenswrapper[4943]: I0307 15:16:12.542824 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:16:12 crc kubenswrapper[4943]: I0307 15:16:12.551348 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-672r7"] Mar 07 15:16:12 crc kubenswrapper[4943]: I0307 15:16:12.676206 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d0c0f330-a8a6-4f08-a3ab-f23a067e72cf-swiftconf\") pod \"swift-ring-rebalance-debug-672r7\" (UID: \"d0c0f330-a8a6-4f08-a3ab-f23a067e72cf\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-672r7" Mar 07 15:16:12 crc kubenswrapper[4943]: I0307 15:16:12.676475 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d0c0f330-a8a6-4f08-a3ab-f23a067e72cf-scripts\") pod \"swift-ring-rebalance-debug-672r7\" (UID: \"d0c0f330-a8a6-4f08-a3ab-f23a067e72cf\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-672r7" Mar 07 15:16:12 crc kubenswrapper[4943]: I0307 15:16:12.676562 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d0c0f330-a8a6-4f08-a3ab-f23a067e72cf-dispersionconf\") pod \"swift-ring-rebalance-debug-672r7\" (UID: \"d0c0f330-a8a6-4f08-a3ab-f23a067e72cf\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-672r7" Mar 07 15:16:12 crc kubenswrapper[4943]: I0307 15:16:12.676685 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d0c0f330-a8a6-4f08-a3ab-f23a067e72cf-etc-swift\") pod \"swift-ring-rebalance-debug-672r7\" (UID: \"d0c0f330-a8a6-4f08-a3ab-f23a067e72cf\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-672r7" Mar 07 15:16:12 crc kubenswrapper[4943]: I0307 15:16:12.676786 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d0c0f330-a8a6-4f08-a3ab-f23a067e72cf-ring-data-devices\") pod \"swift-ring-rebalance-debug-672r7\" (UID: \"d0c0f330-a8a6-4f08-a3ab-f23a067e72cf\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-672r7" Mar 07 15:16:12 crc kubenswrapper[4943]: I0307 15:16:12.676877 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhrp9\" (UniqueName: \"kubernetes.io/projected/d0c0f330-a8a6-4f08-a3ab-f23a067e72cf-kube-api-access-mhrp9\") pod \"swift-ring-rebalance-debug-672r7\" (UID: \"d0c0f330-a8a6-4f08-a3ab-f23a067e72cf\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-672r7" Mar 07 15:16:12 crc kubenswrapper[4943]: I0307 15:16:12.770415 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96e83e83-5c00-48a4-b180-d170d4cff072" path="/var/lib/kubelet/pods/96e83e83-5c00-48a4-b180-d170d4cff072/volumes" Mar 07 15:16:12 crc kubenswrapper[4943]: I0307 15:16:12.778244 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d0c0f330-a8a6-4f08-a3ab-f23a067e72cf-scripts\") pod \"swift-ring-rebalance-debug-672r7\" (UID: \"d0c0f330-a8a6-4f08-a3ab-f23a067e72cf\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-672r7" Mar 07 15:16:12 crc kubenswrapper[4943]: I0307 15:16:12.778283 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d0c0f330-a8a6-4f08-a3ab-f23a067e72cf-dispersionconf\") pod \"swift-ring-rebalance-debug-672r7\" (UID: \"d0c0f330-a8a6-4f08-a3ab-f23a067e72cf\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-672r7" Mar 07 15:16:12 crc kubenswrapper[4943]: I0307 15:16:12.778393 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d0c0f330-a8a6-4f08-a3ab-f23a067e72cf-etc-swift\") pod \"swift-ring-rebalance-debug-672r7\" (UID: \"d0c0f330-a8a6-4f08-a3ab-f23a067e72cf\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-672r7" Mar 07 15:16:12 crc kubenswrapper[4943]: I0307 15:16:12.778436 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d0c0f330-a8a6-4f08-a3ab-f23a067e72cf-ring-data-devices\") pod \"swift-ring-rebalance-debug-672r7\" (UID: \"d0c0f330-a8a6-4f08-a3ab-f23a067e72cf\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-672r7" Mar 07 15:16:12 crc kubenswrapper[4943]: I0307 15:16:12.778484 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhrp9\" (UniqueName: \"kubernetes.io/projected/d0c0f330-a8a6-4f08-a3ab-f23a067e72cf-kube-api-access-mhrp9\") pod \"swift-ring-rebalance-debug-672r7\" (UID: \"d0c0f330-a8a6-4f08-a3ab-f23a067e72cf\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-672r7" Mar 07 15:16:12 crc kubenswrapper[4943]: I0307 15:16:12.778528 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d0c0f330-a8a6-4f08-a3ab-f23a067e72cf-swiftconf\") pod \"swift-ring-rebalance-debug-672r7\" (UID: \"d0c0f330-a8a6-4f08-a3ab-f23a067e72cf\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-672r7" Mar 07 15:16:12 crc kubenswrapper[4943]: I0307 15:16:12.779053 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d0c0f330-a8a6-4f08-a3ab-f23a067e72cf-etc-swift\") pod \"swift-ring-rebalance-debug-672r7\" (UID: \"d0c0f330-a8a6-4f08-a3ab-f23a067e72cf\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-672r7" Mar 07 15:16:12 crc kubenswrapper[4943]: I0307 15:16:12.779252 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d0c0f330-a8a6-4f08-a3ab-f23a067e72cf-scripts\") pod \"swift-ring-rebalance-debug-672r7\" (UID: \"d0c0f330-a8a6-4f08-a3ab-f23a067e72cf\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-672r7" Mar 07 15:16:12 crc kubenswrapper[4943]: I0307 15:16:12.784231 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d0c0f330-a8a6-4f08-a3ab-f23a067e72cf-swiftconf\") pod \"swift-ring-rebalance-debug-672r7\" (UID: \"d0c0f330-a8a6-4f08-a3ab-f23a067e72cf\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-672r7" Mar 07 15:16:12 crc kubenswrapper[4943]: I0307 15:16:12.784638 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d0c0f330-a8a6-4f08-a3ab-f23a067e72cf-ring-data-devices\") pod \"swift-ring-rebalance-debug-672r7\" (UID: \"d0c0f330-a8a6-4f08-a3ab-f23a067e72cf\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-672r7" Mar 07 15:16:12 crc kubenswrapper[4943]: I0307 15:16:12.794508 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d0c0f330-a8a6-4f08-a3ab-f23a067e72cf-dispersionconf\") pod \"swift-ring-rebalance-debug-672r7\" (UID: \"d0c0f330-a8a6-4f08-a3ab-f23a067e72cf\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-672r7" Mar 07 15:16:12 crc kubenswrapper[4943]: I0307 15:16:12.812170 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhrp9\" (UniqueName: \"kubernetes.io/projected/d0c0f330-a8a6-4f08-a3ab-f23a067e72cf-kube-api-access-mhrp9\") pod \"swift-ring-rebalance-debug-672r7\" (UID: \"d0c0f330-a8a6-4f08-a3ab-f23a067e72cf\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-672r7" Mar 07 15:16:12 crc kubenswrapper[4943]: I0307 15:16:12.871287 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-672r7" Mar 07 15:16:13 crc kubenswrapper[4943]: I0307 15:16:13.105279 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-672r7"] Mar 07 15:16:13 crc kubenswrapper[4943]: W0307 15:16:13.105400 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd0c0f330_a8a6_4f08_a3ab_f23a067e72cf.slice/crio-9ba2f498ba6ee52701dc1d03851fc74b044b2fb587dbfbb530ea0d46b65bf0f0 WatchSource:0}: Error finding container 9ba2f498ba6ee52701dc1d03851fc74b044b2fb587dbfbb530ea0d46b65bf0f0: Status 404 returned error can't find the container with id 9ba2f498ba6ee52701dc1d03851fc74b044b2fb587dbfbb530ea0d46b65bf0f0 Mar 07 15:16:13 crc kubenswrapper[4943]: I0307 15:16:13.962444 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-672r7" event={"ID":"d0c0f330-a8a6-4f08-a3ab-f23a067e72cf","Type":"ContainerStarted","Data":"5adf1f4239f1d966c4d0d9db20cd790c0cbacc018ac371dcc37d3bac94d45c71"} Mar 07 15:16:13 crc kubenswrapper[4943]: I0307 15:16:13.962849 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-672r7" event={"ID":"d0c0f330-a8a6-4f08-a3ab-f23a067e72cf","Type":"ContainerStarted","Data":"9ba2f498ba6ee52701dc1d03851fc74b044b2fb587dbfbb530ea0d46b65bf0f0"} Mar 07 15:16:14 crc kubenswrapper[4943]: I0307 15:16:14.002323 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-672r7" podStartSLOduration=2.002291797 podStartE2EDuration="2.002291797s" podCreationTimestamp="2026-03-07 15:16:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:16:13.987538116 +0000 UTC m=+2215.939674654" watchObservedRunningTime="2026-03-07 15:16:14.002291797 +0000 UTC m=+2215.954428335" Mar 07 15:16:14 crc kubenswrapper[4943]: I0307 15:16:14.977615 4943 generic.go:334] "Generic (PLEG): container finished" podID="d0c0f330-a8a6-4f08-a3ab-f23a067e72cf" containerID="5adf1f4239f1d966c4d0d9db20cd790c0cbacc018ac371dcc37d3bac94d45c71" exitCode=0 Mar 07 15:16:14 crc kubenswrapper[4943]: I0307 15:16:14.978087 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-672r7" event={"ID":"d0c0f330-a8a6-4f08-a3ab-f23a067e72cf","Type":"ContainerDied","Data":"5adf1f4239f1d966c4d0d9db20cd790c0cbacc018ac371dcc37d3bac94d45c71"} Mar 07 15:16:16 crc kubenswrapper[4943]: I0307 15:16:16.384418 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-672r7" Mar 07 15:16:16 crc kubenswrapper[4943]: I0307 15:16:16.432693 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-672r7"] Mar 07 15:16:16 crc kubenswrapper[4943]: I0307 15:16:16.439111 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-672r7"] Mar 07 15:16:16 crc kubenswrapper[4943]: I0307 15:16:16.547346 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d0c0f330-a8a6-4f08-a3ab-f23a067e72cf-dispersionconf\") pod \"d0c0f330-a8a6-4f08-a3ab-f23a067e72cf\" (UID: \"d0c0f330-a8a6-4f08-a3ab-f23a067e72cf\") " Mar 07 15:16:16 crc kubenswrapper[4943]: I0307 15:16:16.547484 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mhrp9\" (UniqueName: \"kubernetes.io/projected/d0c0f330-a8a6-4f08-a3ab-f23a067e72cf-kube-api-access-mhrp9\") pod \"d0c0f330-a8a6-4f08-a3ab-f23a067e72cf\" (UID: \"d0c0f330-a8a6-4f08-a3ab-f23a067e72cf\") " Mar 07 15:16:16 crc kubenswrapper[4943]: I0307 15:16:16.547544 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d0c0f330-a8a6-4f08-a3ab-f23a067e72cf-swiftconf\") pod \"d0c0f330-a8a6-4f08-a3ab-f23a067e72cf\" (UID: \"d0c0f330-a8a6-4f08-a3ab-f23a067e72cf\") " Mar 07 15:16:16 crc kubenswrapper[4943]: I0307 15:16:16.547589 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d0c0f330-a8a6-4f08-a3ab-f23a067e72cf-etc-swift\") pod \"d0c0f330-a8a6-4f08-a3ab-f23a067e72cf\" (UID: \"d0c0f330-a8a6-4f08-a3ab-f23a067e72cf\") " Mar 07 15:16:16 crc kubenswrapper[4943]: I0307 15:16:16.547675 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d0c0f330-a8a6-4f08-a3ab-f23a067e72cf-ring-data-devices\") pod \"d0c0f330-a8a6-4f08-a3ab-f23a067e72cf\" (UID: \"d0c0f330-a8a6-4f08-a3ab-f23a067e72cf\") " Mar 07 15:16:16 crc kubenswrapper[4943]: I0307 15:16:16.547760 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d0c0f330-a8a6-4f08-a3ab-f23a067e72cf-scripts\") pod \"d0c0f330-a8a6-4f08-a3ab-f23a067e72cf\" (UID: \"d0c0f330-a8a6-4f08-a3ab-f23a067e72cf\") " Mar 07 15:16:16 crc kubenswrapper[4943]: I0307 15:16:16.549005 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0c0f330-a8a6-4f08-a3ab-f23a067e72cf-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "d0c0f330-a8a6-4f08-a3ab-f23a067e72cf" (UID: "d0c0f330-a8a6-4f08-a3ab-f23a067e72cf"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:16:16 crc kubenswrapper[4943]: I0307 15:16:16.549040 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0c0f330-a8a6-4f08-a3ab-f23a067e72cf-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "d0c0f330-a8a6-4f08-a3ab-f23a067e72cf" (UID: "d0c0f330-a8a6-4f08-a3ab-f23a067e72cf"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:16:16 crc kubenswrapper[4943]: I0307 15:16:16.556031 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0c0f330-a8a6-4f08-a3ab-f23a067e72cf-kube-api-access-mhrp9" (OuterVolumeSpecName: "kube-api-access-mhrp9") pod "d0c0f330-a8a6-4f08-a3ab-f23a067e72cf" (UID: "d0c0f330-a8a6-4f08-a3ab-f23a067e72cf"). InnerVolumeSpecName "kube-api-access-mhrp9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:16:16 crc kubenswrapper[4943]: I0307 15:16:16.581541 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0c0f330-a8a6-4f08-a3ab-f23a067e72cf-scripts" (OuterVolumeSpecName: "scripts") pod "d0c0f330-a8a6-4f08-a3ab-f23a067e72cf" (UID: "d0c0f330-a8a6-4f08-a3ab-f23a067e72cf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:16:16 crc kubenswrapper[4943]: I0307 15:16:16.586756 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0c0f330-a8a6-4f08-a3ab-f23a067e72cf-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "d0c0f330-a8a6-4f08-a3ab-f23a067e72cf" (UID: "d0c0f330-a8a6-4f08-a3ab-f23a067e72cf"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:16:16 crc kubenswrapper[4943]: I0307 15:16:16.591032 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0c0f330-a8a6-4f08-a3ab-f23a067e72cf-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "d0c0f330-a8a6-4f08-a3ab-f23a067e72cf" (UID: "d0c0f330-a8a6-4f08-a3ab-f23a067e72cf"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:16:16 crc kubenswrapper[4943]: I0307 15:16:16.650462 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d0c0f330-a8a6-4f08-a3ab-f23a067e72cf-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:16 crc kubenswrapper[4943]: I0307 15:16:16.650701 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mhrp9\" (UniqueName: \"kubernetes.io/projected/d0c0f330-a8a6-4f08-a3ab-f23a067e72cf-kube-api-access-mhrp9\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:16 crc kubenswrapper[4943]: I0307 15:16:16.650867 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d0c0f330-a8a6-4f08-a3ab-f23a067e72cf-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:16 crc kubenswrapper[4943]: I0307 15:16:16.651052 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d0c0f330-a8a6-4f08-a3ab-f23a067e72cf-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:16 crc kubenswrapper[4943]: I0307 15:16:16.651185 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d0c0f330-a8a6-4f08-a3ab-f23a067e72cf-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:16 crc kubenswrapper[4943]: I0307 15:16:16.651321 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d0c0f330-a8a6-4f08-a3ab-f23a067e72cf-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:16 crc kubenswrapper[4943]: I0307 15:16:16.772079 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0c0f330-a8a6-4f08-a3ab-f23a067e72cf" path="/var/lib/kubelet/pods/d0c0f330-a8a6-4f08-a3ab-f23a067e72cf/volumes" Mar 07 15:16:17 crc kubenswrapper[4943]: I0307 15:16:17.004751 4943 scope.go:117] "RemoveContainer" containerID="5adf1f4239f1d966c4d0d9db20cd790c0cbacc018ac371dcc37d3bac94d45c71" Mar 07 15:16:17 crc kubenswrapper[4943]: I0307 15:16:17.004857 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-672r7" Mar 07 15:16:17 crc kubenswrapper[4943]: I0307 15:16:17.643312 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-9qnkj"] Mar 07 15:16:17 crc kubenswrapper[4943]: E0307 15:16:17.644384 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0c0f330-a8a6-4f08-a3ab-f23a067e72cf" containerName="swift-ring-rebalance" Mar 07 15:16:17 crc kubenswrapper[4943]: I0307 15:16:17.644417 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0c0f330-a8a6-4f08-a3ab-f23a067e72cf" containerName="swift-ring-rebalance" Mar 07 15:16:17 crc kubenswrapper[4943]: I0307 15:16:17.644861 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0c0f330-a8a6-4f08-a3ab-f23a067e72cf" containerName="swift-ring-rebalance" Mar 07 15:16:17 crc kubenswrapper[4943]: I0307 15:16:17.645867 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-9qnkj" Mar 07 15:16:17 crc kubenswrapper[4943]: I0307 15:16:17.649273 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:16:17 crc kubenswrapper[4943]: I0307 15:16:17.649560 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:16:17 crc kubenswrapper[4943]: I0307 15:16:17.652240 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-9qnkj"] Mar 07 15:16:17 crc kubenswrapper[4943]: I0307 15:16:17.770583 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/11e09613-e4be-4478-a435-dc66b29c61bd-ring-data-devices\") pod \"swift-ring-rebalance-debug-9qnkj\" (UID: \"11e09613-e4be-4478-a435-dc66b29c61bd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-9qnkj" Mar 07 15:16:17 crc kubenswrapper[4943]: I0307 15:16:17.770682 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/11e09613-e4be-4478-a435-dc66b29c61bd-scripts\") pod \"swift-ring-rebalance-debug-9qnkj\" (UID: \"11e09613-e4be-4478-a435-dc66b29c61bd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-9qnkj" Mar 07 15:16:17 crc kubenswrapper[4943]: I0307 15:16:17.770813 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/11e09613-e4be-4478-a435-dc66b29c61bd-dispersionconf\") pod \"swift-ring-rebalance-debug-9qnkj\" (UID: \"11e09613-e4be-4478-a435-dc66b29c61bd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-9qnkj" Mar 07 15:16:17 crc kubenswrapper[4943]: I0307 15:16:17.770868 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgdlz\" (UniqueName: \"kubernetes.io/projected/11e09613-e4be-4478-a435-dc66b29c61bd-kube-api-access-tgdlz\") pod \"swift-ring-rebalance-debug-9qnkj\" (UID: \"11e09613-e4be-4478-a435-dc66b29c61bd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-9qnkj" Mar 07 15:16:17 crc kubenswrapper[4943]: I0307 15:16:17.770917 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/11e09613-e4be-4478-a435-dc66b29c61bd-swiftconf\") pod \"swift-ring-rebalance-debug-9qnkj\" (UID: \"11e09613-e4be-4478-a435-dc66b29c61bd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-9qnkj" Mar 07 15:16:17 crc kubenswrapper[4943]: I0307 15:16:17.770982 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/11e09613-e4be-4478-a435-dc66b29c61bd-etc-swift\") pod \"swift-ring-rebalance-debug-9qnkj\" (UID: \"11e09613-e4be-4478-a435-dc66b29c61bd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-9qnkj" Mar 07 15:16:17 crc kubenswrapper[4943]: I0307 15:16:17.872700 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/11e09613-e4be-4478-a435-dc66b29c61bd-ring-data-devices\") pod \"swift-ring-rebalance-debug-9qnkj\" (UID: \"11e09613-e4be-4478-a435-dc66b29c61bd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-9qnkj" Mar 07 15:16:17 crc kubenswrapper[4943]: I0307 15:16:17.872827 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/11e09613-e4be-4478-a435-dc66b29c61bd-scripts\") pod \"swift-ring-rebalance-debug-9qnkj\" (UID: \"11e09613-e4be-4478-a435-dc66b29c61bd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-9qnkj" Mar 07 15:16:17 crc kubenswrapper[4943]: I0307 15:16:17.873057 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/11e09613-e4be-4478-a435-dc66b29c61bd-dispersionconf\") pod \"swift-ring-rebalance-debug-9qnkj\" (UID: \"11e09613-e4be-4478-a435-dc66b29c61bd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-9qnkj" Mar 07 15:16:17 crc kubenswrapper[4943]: I0307 15:16:17.873135 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgdlz\" (UniqueName: \"kubernetes.io/projected/11e09613-e4be-4478-a435-dc66b29c61bd-kube-api-access-tgdlz\") pod \"swift-ring-rebalance-debug-9qnkj\" (UID: \"11e09613-e4be-4478-a435-dc66b29c61bd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-9qnkj" Mar 07 15:16:17 crc kubenswrapper[4943]: I0307 15:16:17.873221 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/11e09613-e4be-4478-a435-dc66b29c61bd-swiftconf\") pod \"swift-ring-rebalance-debug-9qnkj\" (UID: \"11e09613-e4be-4478-a435-dc66b29c61bd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-9qnkj" Mar 07 15:16:17 crc kubenswrapper[4943]: I0307 15:16:17.873267 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/11e09613-e4be-4478-a435-dc66b29c61bd-etc-swift\") pod \"swift-ring-rebalance-debug-9qnkj\" (UID: \"11e09613-e4be-4478-a435-dc66b29c61bd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-9qnkj" Mar 07 15:16:17 crc kubenswrapper[4943]: I0307 15:16:17.873984 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/11e09613-e4be-4478-a435-dc66b29c61bd-ring-data-devices\") pod \"swift-ring-rebalance-debug-9qnkj\" (UID: \"11e09613-e4be-4478-a435-dc66b29c61bd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-9qnkj" Mar 07 15:16:17 crc kubenswrapper[4943]: I0307 15:16:17.874155 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/11e09613-e4be-4478-a435-dc66b29c61bd-etc-swift\") pod \"swift-ring-rebalance-debug-9qnkj\" (UID: \"11e09613-e4be-4478-a435-dc66b29c61bd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-9qnkj" Mar 07 15:16:17 crc kubenswrapper[4943]: I0307 15:16:17.875668 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/11e09613-e4be-4478-a435-dc66b29c61bd-scripts\") pod \"swift-ring-rebalance-debug-9qnkj\" (UID: \"11e09613-e4be-4478-a435-dc66b29c61bd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-9qnkj" Mar 07 15:16:17 crc kubenswrapper[4943]: I0307 15:16:17.878529 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/11e09613-e4be-4478-a435-dc66b29c61bd-dispersionconf\") pod \"swift-ring-rebalance-debug-9qnkj\" (UID: \"11e09613-e4be-4478-a435-dc66b29c61bd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-9qnkj" Mar 07 15:16:17 crc kubenswrapper[4943]: I0307 15:16:17.878876 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/11e09613-e4be-4478-a435-dc66b29c61bd-swiftconf\") pod \"swift-ring-rebalance-debug-9qnkj\" (UID: \"11e09613-e4be-4478-a435-dc66b29c61bd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-9qnkj" Mar 07 15:16:17 crc kubenswrapper[4943]: I0307 15:16:17.900399 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgdlz\" (UniqueName: \"kubernetes.io/projected/11e09613-e4be-4478-a435-dc66b29c61bd-kube-api-access-tgdlz\") pod \"swift-ring-rebalance-debug-9qnkj\" (UID: \"11e09613-e4be-4478-a435-dc66b29c61bd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-9qnkj" Mar 07 15:16:17 crc kubenswrapper[4943]: I0307 15:16:17.981611 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-9qnkj" Mar 07 15:16:18 crc kubenswrapper[4943]: I0307 15:16:18.228563 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-9qnkj"] Mar 07 15:16:18 crc kubenswrapper[4943]: W0307 15:16:18.231152 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod11e09613_e4be_4478_a435_dc66b29c61bd.slice/crio-6b2b9eaabeeb3fc4ff62389824a79b374cbf9ffbeccb8077fcf8f72702d4694e WatchSource:0}: Error finding container 6b2b9eaabeeb3fc4ff62389824a79b374cbf9ffbeccb8077fcf8f72702d4694e: Status 404 returned error can't find the container with id 6b2b9eaabeeb3fc4ff62389824a79b374cbf9ffbeccb8077fcf8f72702d4694e Mar 07 15:16:19 crc kubenswrapper[4943]: I0307 15:16:19.035719 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-9qnkj" event={"ID":"11e09613-e4be-4478-a435-dc66b29c61bd","Type":"ContainerStarted","Data":"d486221e7cf6b2e38945d6f110d67cdb105d1da60f995808ad70faebec537173"} Mar 07 15:16:19 crc kubenswrapper[4943]: I0307 15:16:19.036091 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-9qnkj" event={"ID":"11e09613-e4be-4478-a435-dc66b29c61bd","Type":"ContainerStarted","Data":"6b2b9eaabeeb3fc4ff62389824a79b374cbf9ffbeccb8077fcf8f72702d4694e"} Mar 07 15:16:19 crc kubenswrapper[4943]: I0307 15:16:19.064985 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-9qnkj" podStartSLOduration=2.064905376 podStartE2EDuration="2.064905376s" podCreationTimestamp="2026-03-07 15:16:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:16:19.052121824 +0000 UTC m=+2221.004258332" watchObservedRunningTime="2026-03-07 15:16:19.064905376 +0000 UTC m=+2221.017041914" Mar 07 15:16:20 crc kubenswrapper[4943]: I0307 15:16:20.047739 4943 generic.go:334] "Generic (PLEG): container finished" podID="11e09613-e4be-4478-a435-dc66b29c61bd" containerID="d486221e7cf6b2e38945d6f110d67cdb105d1da60f995808ad70faebec537173" exitCode=0 Mar 07 15:16:20 crc kubenswrapper[4943]: I0307 15:16:20.047798 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-9qnkj" event={"ID":"11e09613-e4be-4478-a435-dc66b29c61bd","Type":"ContainerDied","Data":"d486221e7cf6b2e38945d6f110d67cdb105d1da60f995808ad70faebec537173"} Mar 07 15:16:21 crc kubenswrapper[4943]: I0307 15:16:21.397854 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-9qnkj" Mar 07 15:16:21 crc kubenswrapper[4943]: I0307 15:16:21.450463 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-9qnkj"] Mar 07 15:16:21 crc kubenswrapper[4943]: I0307 15:16:21.459163 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-9qnkj"] Mar 07 15:16:21 crc kubenswrapper[4943]: I0307 15:16:21.551609 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tgdlz\" (UniqueName: \"kubernetes.io/projected/11e09613-e4be-4478-a435-dc66b29c61bd-kube-api-access-tgdlz\") pod \"11e09613-e4be-4478-a435-dc66b29c61bd\" (UID: \"11e09613-e4be-4478-a435-dc66b29c61bd\") " Mar 07 15:16:21 crc kubenswrapper[4943]: I0307 15:16:21.551968 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/11e09613-e4be-4478-a435-dc66b29c61bd-ring-data-devices\") pod \"11e09613-e4be-4478-a435-dc66b29c61bd\" (UID: \"11e09613-e4be-4478-a435-dc66b29c61bd\") " Mar 07 15:16:21 crc kubenswrapper[4943]: I0307 15:16:21.552012 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/11e09613-e4be-4478-a435-dc66b29c61bd-scripts\") pod \"11e09613-e4be-4478-a435-dc66b29c61bd\" (UID: \"11e09613-e4be-4478-a435-dc66b29c61bd\") " Mar 07 15:16:21 crc kubenswrapper[4943]: I0307 15:16:21.552077 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/11e09613-e4be-4478-a435-dc66b29c61bd-swiftconf\") pod \"11e09613-e4be-4478-a435-dc66b29c61bd\" (UID: \"11e09613-e4be-4478-a435-dc66b29c61bd\") " Mar 07 15:16:21 crc kubenswrapper[4943]: I0307 15:16:21.552123 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/11e09613-e4be-4478-a435-dc66b29c61bd-dispersionconf\") pod \"11e09613-e4be-4478-a435-dc66b29c61bd\" (UID: \"11e09613-e4be-4478-a435-dc66b29c61bd\") " Mar 07 15:16:21 crc kubenswrapper[4943]: I0307 15:16:21.552188 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/11e09613-e4be-4478-a435-dc66b29c61bd-etc-swift\") pod \"11e09613-e4be-4478-a435-dc66b29c61bd\" (UID: \"11e09613-e4be-4478-a435-dc66b29c61bd\") " Mar 07 15:16:21 crc kubenswrapper[4943]: I0307 15:16:21.554022 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11e09613-e4be-4478-a435-dc66b29c61bd-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "11e09613-e4be-4478-a435-dc66b29c61bd" (UID: "11e09613-e4be-4478-a435-dc66b29c61bd"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:16:21 crc kubenswrapper[4943]: I0307 15:16:21.554883 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11e09613-e4be-4478-a435-dc66b29c61bd-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "11e09613-e4be-4478-a435-dc66b29c61bd" (UID: "11e09613-e4be-4478-a435-dc66b29c61bd"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:16:21 crc kubenswrapper[4943]: I0307 15:16:21.560011 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11e09613-e4be-4478-a435-dc66b29c61bd-kube-api-access-tgdlz" (OuterVolumeSpecName: "kube-api-access-tgdlz") pod "11e09613-e4be-4478-a435-dc66b29c61bd" (UID: "11e09613-e4be-4478-a435-dc66b29c61bd"). InnerVolumeSpecName "kube-api-access-tgdlz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:16:21 crc kubenswrapper[4943]: I0307 15:16:21.573170 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11e09613-e4be-4478-a435-dc66b29c61bd-scripts" (OuterVolumeSpecName: "scripts") pod "11e09613-e4be-4478-a435-dc66b29c61bd" (UID: "11e09613-e4be-4478-a435-dc66b29c61bd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:16:21 crc kubenswrapper[4943]: I0307 15:16:21.591270 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11e09613-e4be-4478-a435-dc66b29c61bd-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "11e09613-e4be-4478-a435-dc66b29c61bd" (UID: "11e09613-e4be-4478-a435-dc66b29c61bd"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:16:21 crc kubenswrapper[4943]: I0307 15:16:21.593065 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11e09613-e4be-4478-a435-dc66b29c61bd-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "11e09613-e4be-4478-a435-dc66b29c61bd" (UID: "11e09613-e4be-4478-a435-dc66b29c61bd"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:16:21 crc kubenswrapper[4943]: I0307 15:16:21.654270 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tgdlz\" (UniqueName: \"kubernetes.io/projected/11e09613-e4be-4478-a435-dc66b29c61bd-kube-api-access-tgdlz\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:21 crc kubenswrapper[4943]: I0307 15:16:21.654325 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/11e09613-e4be-4478-a435-dc66b29c61bd-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:21 crc kubenswrapper[4943]: I0307 15:16:21.654347 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/11e09613-e4be-4478-a435-dc66b29c61bd-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:21 crc kubenswrapper[4943]: I0307 15:16:21.654364 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/11e09613-e4be-4478-a435-dc66b29c61bd-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:21 crc kubenswrapper[4943]: I0307 15:16:21.654382 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/11e09613-e4be-4478-a435-dc66b29c61bd-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:21 crc kubenswrapper[4943]: I0307 15:16:21.654399 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/11e09613-e4be-4478-a435-dc66b29c61bd-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:21 crc kubenswrapper[4943]: I0307 15:16:21.942904 4943 scope.go:117] "RemoveContainer" containerID="be5eb23beadf94e639477c15837fe80c8e5410f72b6bfc4038ab40de44979f6d" Mar 07 15:16:21 crc kubenswrapper[4943]: I0307 15:16:21.991561 4943 scope.go:117] "RemoveContainer" containerID="4532bdcc9922a6a649a935ca6b7bc782b1aa3e570e68991386637b2ea30daef0" Mar 07 15:16:22 crc kubenswrapper[4943]: I0307 15:16:22.061896 4943 scope.go:117] "RemoveContainer" containerID="0b3d2f580d1445cec8ffea6df297252f539ecf099f6b63cb72727022bdf8093d" Mar 07 15:16:22 crc kubenswrapper[4943]: I0307 15:16:22.074315 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6b2b9eaabeeb3fc4ff62389824a79b374cbf9ffbeccb8077fcf8f72702d4694e" Mar 07 15:16:22 crc kubenswrapper[4943]: I0307 15:16:22.074341 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-9qnkj" Mar 07 15:16:22 crc kubenswrapper[4943]: I0307 15:16:22.571631 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-gklrn"] Mar 07 15:16:22 crc kubenswrapper[4943]: E0307 15:16:22.572141 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11e09613-e4be-4478-a435-dc66b29c61bd" containerName="swift-ring-rebalance" Mar 07 15:16:22 crc kubenswrapper[4943]: I0307 15:16:22.572165 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="11e09613-e4be-4478-a435-dc66b29c61bd" containerName="swift-ring-rebalance" Mar 07 15:16:22 crc kubenswrapper[4943]: I0307 15:16:22.572428 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="11e09613-e4be-4478-a435-dc66b29c61bd" containerName="swift-ring-rebalance" Mar 07 15:16:22 crc kubenswrapper[4943]: I0307 15:16:22.573256 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-gklrn" Mar 07 15:16:22 crc kubenswrapper[4943]: I0307 15:16:22.578347 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:16:22 crc kubenswrapper[4943]: I0307 15:16:22.579658 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:16:22 crc kubenswrapper[4943]: I0307 15:16:22.586427 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-gklrn"] Mar 07 15:16:22 crc kubenswrapper[4943]: I0307 15:16:22.669170 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/41bc02a1-69c1-445f-a49a-d82368dba646-swiftconf\") pod \"swift-ring-rebalance-debug-gklrn\" (UID: \"41bc02a1-69c1-445f-a49a-d82368dba646\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gklrn" Mar 07 15:16:22 crc kubenswrapper[4943]: I0307 15:16:22.669315 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zl25\" (UniqueName: \"kubernetes.io/projected/41bc02a1-69c1-445f-a49a-d82368dba646-kube-api-access-8zl25\") pod \"swift-ring-rebalance-debug-gklrn\" (UID: \"41bc02a1-69c1-445f-a49a-d82368dba646\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gklrn" Mar 07 15:16:22 crc kubenswrapper[4943]: I0307 15:16:22.669506 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/41bc02a1-69c1-445f-a49a-d82368dba646-etc-swift\") pod \"swift-ring-rebalance-debug-gklrn\" (UID: \"41bc02a1-69c1-445f-a49a-d82368dba646\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gklrn" Mar 07 15:16:22 crc kubenswrapper[4943]: I0307 15:16:22.669634 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/41bc02a1-69c1-445f-a49a-d82368dba646-dispersionconf\") pod \"swift-ring-rebalance-debug-gklrn\" (UID: \"41bc02a1-69c1-445f-a49a-d82368dba646\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gklrn" Mar 07 15:16:22 crc kubenswrapper[4943]: I0307 15:16:22.669752 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/41bc02a1-69c1-445f-a49a-d82368dba646-ring-data-devices\") pod \"swift-ring-rebalance-debug-gklrn\" (UID: \"41bc02a1-69c1-445f-a49a-d82368dba646\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gklrn" Mar 07 15:16:22 crc kubenswrapper[4943]: I0307 15:16:22.669813 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/41bc02a1-69c1-445f-a49a-d82368dba646-scripts\") pod \"swift-ring-rebalance-debug-gklrn\" (UID: \"41bc02a1-69c1-445f-a49a-d82368dba646\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gklrn" Mar 07 15:16:22 crc kubenswrapper[4943]: I0307 15:16:22.770889 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/41bc02a1-69c1-445f-a49a-d82368dba646-ring-data-devices\") pod \"swift-ring-rebalance-debug-gklrn\" (UID: \"41bc02a1-69c1-445f-a49a-d82368dba646\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gklrn" Mar 07 15:16:22 crc kubenswrapper[4943]: I0307 15:16:22.770971 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/41bc02a1-69c1-445f-a49a-d82368dba646-scripts\") pod \"swift-ring-rebalance-debug-gklrn\" (UID: \"41bc02a1-69c1-445f-a49a-d82368dba646\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gklrn" Mar 07 15:16:22 crc kubenswrapper[4943]: I0307 15:16:22.771004 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/41bc02a1-69c1-445f-a49a-d82368dba646-swiftconf\") pod \"swift-ring-rebalance-debug-gklrn\" (UID: \"41bc02a1-69c1-445f-a49a-d82368dba646\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gklrn" Mar 07 15:16:22 crc kubenswrapper[4943]: I0307 15:16:22.771048 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zl25\" (UniqueName: \"kubernetes.io/projected/41bc02a1-69c1-445f-a49a-d82368dba646-kube-api-access-8zl25\") pod \"swift-ring-rebalance-debug-gklrn\" (UID: \"41bc02a1-69c1-445f-a49a-d82368dba646\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gklrn" Mar 07 15:16:22 crc kubenswrapper[4943]: I0307 15:16:22.771104 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/41bc02a1-69c1-445f-a49a-d82368dba646-etc-swift\") pod \"swift-ring-rebalance-debug-gklrn\" (UID: \"41bc02a1-69c1-445f-a49a-d82368dba646\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gklrn" Mar 07 15:16:22 crc kubenswrapper[4943]: I0307 15:16:22.771146 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/41bc02a1-69c1-445f-a49a-d82368dba646-dispersionconf\") pod \"swift-ring-rebalance-debug-gklrn\" (UID: \"41bc02a1-69c1-445f-a49a-d82368dba646\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gklrn" Mar 07 15:16:22 crc kubenswrapper[4943]: I0307 15:16:22.771896 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/41bc02a1-69c1-445f-a49a-d82368dba646-scripts\") pod \"swift-ring-rebalance-debug-gklrn\" (UID: \"41bc02a1-69c1-445f-a49a-d82368dba646\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gklrn" Mar 07 15:16:22 crc kubenswrapper[4943]: I0307 15:16:22.772189 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/41bc02a1-69c1-445f-a49a-d82368dba646-ring-data-devices\") pod \"swift-ring-rebalance-debug-gklrn\" (UID: \"41bc02a1-69c1-445f-a49a-d82368dba646\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gklrn" Mar 07 15:16:22 crc kubenswrapper[4943]: I0307 15:16:22.772196 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/41bc02a1-69c1-445f-a49a-d82368dba646-etc-swift\") pod \"swift-ring-rebalance-debug-gklrn\" (UID: \"41bc02a1-69c1-445f-a49a-d82368dba646\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gklrn" Mar 07 15:16:22 crc kubenswrapper[4943]: I0307 15:16:22.776895 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11e09613-e4be-4478-a435-dc66b29c61bd" path="/var/lib/kubelet/pods/11e09613-e4be-4478-a435-dc66b29c61bd/volumes" Mar 07 15:16:22 crc kubenswrapper[4943]: I0307 15:16:22.790584 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/41bc02a1-69c1-445f-a49a-d82368dba646-dispersionconf\") pod \"swift-ring-rebalance-debug-gklrn\" (UID: \"41bc02a1-69c1-445f-a49a-d82368dba646\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gklrn" Mar 07 15:16:22 crc kubenswrapper[4943]: I0307 15:16:22.790589 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/41bc02a1-69c1-445f-a49a-d82368dba646-swiftconf\") pod \"swift-ring-rebalance-debug-gklrn\" (UID: \"41bc02a1-69c1-445f-a49a-d82368dba646\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gklrn" Mar 07 15:16:22 crc kubenswrapper[4943]: I0307 15:16:22.795747 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zl25\" (UniqueName: \"kubernetes.io/projected/41bc02a1-69c1-445f-a49a-d82368dba646-kube-api-access-8zl25\") pod \"swift-ring-rebalance-debug-gklrn\" (UID: \"41bc02a1-69c1-445f-a49a-d82368dba646\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gklrn" Mar 07 15:16:22 crc kubenswrapper[4943]: I0307 15:16:22.901232 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-gklrn" Mar 07 15:16:23 crc kubenswrapper[4943]: I0307 15:16:23.181713 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-gklrn"] Mar 07 15:16:23 crc kubenswrapper[4943]: W0307 15:16:23.189858 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod41bc02a1_69c1_445f_a49a_d82368dba646.slice/crio-7b1ed8996f36f8b12c35afef7b35638a382e2c21fd90da082cb57b5c71199607 WatchSource:0}: Error finding container 7b1ed8996f36f8b12c35afef7b35638a382e2c21fd90da082cb57b5c71199607: Status 404 returned error can't find the container with id 7b1ed8996f36f8b12c35afef7b35638a382e2c21fd90da082cb57b5c71199607 Mar 07 15:16:24 crc kubenswrapper[4943]: I0307 15:16:24.107201 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-gklrn" event={"ID":"41bc02a1-69c1-445f-a49a-d82368dba646","Type":"ContainerStarted","Data":"bf7c5ce789618cb0adfe3fd542245a42d5eb016674e0bcfa39ab3d80b342fa61"} Mar 07 15:16:24 crc kubenswrapper[4943]: I0307 15:16:24.107505 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-gklrn" event={"ID":"41bc02a1-69c1-445f-a49a-d82368dba646","Type":"ContainerStarted","Data":"7b1ed8996f36f8b12c35afef7b35638a382e2c21fd90da082cb57b5c71199607"} Mar 07 15:16:24 crc kubenswrapper[4943]: I0307 15:16:24.142450 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-gklrn" podStartSLOduration=2.142425528 podStartE2EDuration="2.142425528s" podCreationTimestamp="2026-03-07 15:16:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:16:24.130285502 +0000 UTC m=+2226.082422080" watchObservedRunningTime="2026-03-07 15:16:24.142425528 +0000 UTC m=+2226.094562066" Mar 07 15:16:25 crc kubenswrapper[4943]: I0307 15:16:25.136526 4943 generic.go:334] "Generic (PLEG): container finished" podID="41bc02a1-69c1-445f-a49a-d82368dba646" containerID="bf7c5ce789618cb0adfe3fd542245a42d5eb016674e0bcfa39ab3d80b342fa61" exitCode=0 Mar 07 15:16:25 crc kubenswrapper[4943]: I0307 15:16:25.136589 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-gklrn" event={"ID":"41bc02a1-69c1-445f-a49a-d82368dba646","Type":"ContainerDied","Data":"bf7c5ce789618cb0adfe3fd542245a42d5eb016674e0bcfa39ab3d80b342fa61"} Mar 07 15:16:26 crc kubenswrapper[4943]: I0307 15:16:26.508915 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-gklrn" Mar 07 15:16:26 crc kubenswrapper[4943]: I0307 15:16:26.559079 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-gklrn"] Mar 07 15:16:26 crc kubenswrapper[4943]: I0307 15:16:26.569123 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-gklrn"] Mar 07 15:16:26 crc kubenswrapper[4943]: I0307 15:16:26.630417 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/41bc02a1-69c1-445f-a49a-d82368dba646-swiftconf\") pod \"41bc02a1-69c1-445f-a49a-d82368dba646\" (UID: \"41bc02a1-69c1-445f-a49a-d82368dba646\") " Mar 07 15:16:26 crc kubenswrapper[4943]: I0307 15:16:26.630490 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/41bc02a1-69c1-445f-a49a-d82368dba646-scripts\") pod \"41bc02a1-69c1-445f-a49a-d82368dba646\" (UID: \"41bc02a1-69c1-445f-a49a-d82368dba646\") " Mar 07 15:16:26 crc kubenswrapper[4943]: I0307 15:16:26.630555 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8zl25\" (UniqueName: \"kubernetes.io/projected/41bc02a1-69c1-445f-a49a-d82368dba646-kube-api-access-8zl25\") pod \"41bc02a1-69c1-445f-a49a-d82368dba646\" (UID: \"41bc02a1-69c1-445f-a49a-d82368dba646\") " Mar 07 15:16:26 crc kubenswrapper[4943]: I0307 15:16:26.630608 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/41bc02a1-69c1-445f-a49a-d82368dba646-ring-data-devices\") pod \"41bc02a1-69c1-445f-a49a-d82368dba646\" (UID: \"41bc02a1-69c1-445f-a49a-d82368dba646\") " Mar 07 15:16:26 crc kubenswrapper[4943]: I0307 15:16:26.630635 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/41bc02a1-69c1-445f-a49a-d82368dba646-dispersionconf\") pod \"41bc02a1-69c1-445f-a49a-d82368dba646\" (UID: \"41bc02a1-69c1-445f-a49a-d82368dba646\") " Mar 07 15:16:26 crc kubenswrapper[4943]: I0307 15:16:26.630746 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/41bc02a1-69c1-445f-a49a-d82368dba646-etc-swift\") pod \"41bc02a1-69c1-445f-a49a-d82368dba646\" (UID: \"41bc02a1-69c1-445f-a49a-d82368dba646\") " Mar 07 15:16:26 crc kubenswrapper[4943]: I0307 15:16:26.631465 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/41bc02a1-69c1-445f-a49a-d82368dba646-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "41bc02a1-69c1-445f-a49a-d82368dba646" (UID: "41bc02a1-69c1-445f-a49a-d82368dba646"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:16:26 crc kubenswrapper[4943]: I0307 15:16:26.631919 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41bc02a1-69c1-445f-a49a-d82368dba646-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "41bc02a1-69c1-445f-a49a-d82368dba646" (UID: "41bc02a1-69c1-445f-a49a-d82368dba646"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:16:26 crc kubenswrapper[4943]: I0307 15:16:26.635832 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41bc02a1-69c1-445f-a49a-d82368dba646-kube-api-access-8zl25" (OuterVolumeSpecName: "kube-api-access-8zl25") pod "41bc02a1-69c1-445f-a49a-d82368dba646" (UID: "41bc02a1-69c1-445f-a49a-d82368dba646"). InnerVolumeSpecName "kube-api-access-8zl25". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:16:26 crc kubenswrapper[4943]: I0307 15:16:26.654062 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41bc02a1-69c1-445f-a49a-d82368dba646-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "41bc02a1-69c1-445f-a49a-d82368dba646" (UID: "41bc02a1-69c1-445f-a49a-d82368dba646"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:16:26 crc kubenswrapper[4943]: I0307 15:16:26.664293 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/41bc02a1-69c1-445f-a49a-d82368dba646-scripts" (OuterVolumeSpecName: "scripts") pod "41bc02a1-69c1-445f-a49a-d82368dba646" (UID: "41bc02a1-69c1-445f-a49a-d82368dba646"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:16:26 crc kubenswrapper[4943]: I0307 15:16:26.667483 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41bc02a1-69c1-445f-a49a-d82368dba646-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "41bc02a1-69c1-445f-a49a-d82368dba646" (UID: "41bc02a1-69c1-445f-a49a-d82368dba646"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:16:26 crc kubenswrapper[4943]: I0307 15:16:26.732499 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/41bc02a1-69c1-445f-a49a-d82368dba646-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:26 crc kubenswrapper[4943]: I0307 15:16:26.732540 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/41bc02a1-69c1-445f-a49a-d82368dba646-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:26 crc kubenswrapper[4943]: I0307 15:16:26.732557 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/41bc02a1-69c1-445f-a49a-d82368dba646-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:26 crc kubenswrapper[4943]: I0307 15:16:26.732572 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/41bc02a1-69c1-445f-a49a-d82368dba646-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:26 crc kubenswrapper[4943]: I0307 15:16:26.732588 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8zl25\" (UniqueName: \"kubernetes.io/projected/41bc02a1-69c1-445f-a49a-d82368dba646-kube-api-access-8zl25\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:26 crc kubenswrapper[4943]: I0307 15:16:26.732607 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/41bc02a1-69c1-445f-a49a-d82368dba646-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:26 crc kubenswrapper[4943]: I0307 15:16:26.770681 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41bc02a1-69c1-445f-a49a-d82368dba646" path="/var/lib/kubelet/pods/41bc02a1-69c1-445f-a49a-d82368dba646/volumes" Mar 07 15:16:27 crc kubenswrapper[4943]: I0307 15:16:27.161721 4943 scope.go:117] "RemoveContainer" containerID="bf7c5ce789618cb0adfe3fd542245a42d5eb016674e0bcfa39ab3d80b342fa61" Mar 07 15:16:27 crc kubenswrapper[4943]: I0307 15:16:27.162039 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-gklrn" Mar 07 15:16:27 crc kubenswrapper[4943]: I0307 15:16:27.722874 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-5mxqq"] Mar 07 15:16:27 crc kubenswrapper[4943]: E0307 15:16:27.723657 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41bc02a1-69c1-445f-a49a-d82368dba646" containerName="swift-ring-rebalance" Mar 07 15:16:27 crc kubenswrapper[4943]: I0307 15:16:27.723679 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="41bc02a1-69c1-445f-a49a-d82368dba646" containerName="swift-ring-rebalance" Mar 07 15:16:27 crc kubenswrapper[4943]: I0307 15:16:27.723996 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="41bc02a1-69c1-445f-a49a-d82368dba646" containerName="swift-ring-rebalance" Mar 07 15:16:27 crc kubenswrapper[4943]: I0307 15:16:27.724855 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mxqq" Mar 07 15:16:27 crc kubenswrapper[4943]: I0307 15:16:27.733568 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:16:27 crc kubenswrapper[4943]: I0307 15:16:27.733626 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:16:27 crc kubenswrapper[4943]: I0307 15:16:27.737451 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-5mxqq"] Mar 07 15:16:27 crc kubenswrapper[4943]: I0307 15:16:27.855635 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/61d5f204-7421-466f-a993-22e4608be96a-swiftconf\") pod \"swift-ring-rebalance-debug-5mxqq\" (UID: \"61d5f204-7421-466f-a993-22e4608be96a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mxqq" Mar 07 15:16:27 crc kubenswrapper[4943]: I0307 15:16:27.856065 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/61d5f204-7421-466f-a993-22e4608be96a-ring-data-devices\") pod \"swift-ring-rebalance-debug-5mxqq\" (UID: \"61d5f204-7421-466f-a993-22e4608be96a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mxqq" Mar 07 15:16:27 crc kubenswrapper[4943]: I0307 15:16:27.856572 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/61d5f204-7421-466f-a993-22e4608be96a-etc-swift\") pod \"swift-ring-rebalance-debug-5mxqq\" (UID: \"61d5f204-7421-466f-a993-22e4608be96a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mxqq" Mar 07 15:16:27 crc kubenswrapper[4943]: I0307 15:16:27.856698 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4skpm\" (UniqueName: \"kubernetes.io/projected/61d5f204-7421-466f-a993-22e4608be96a-kube-api-access-4skpm\") pod \"swift-ring-rebalance-debug-5mxqq\" (UID: \"61d5f204-7421-466f-a993-22e4608be96a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mxqq" Mar 07 15:16:27 crc kubenswrapper[4943]: I0307 15:16:27.856744 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/61d5f204-7421-466f-a993-22e4608be96a-scripts\") pod \"swift-ring-rebalance-debug-5mxqq\" (UID: \"61d5f204-7421-466f-a993-22e4608be96a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mxqq" Mar 07 15:16:27 crc kubenswrapper[4943]: I0307 15:16:27.856774 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/61d5f204-7421-466f-a993-22e4608be96a-dispersionconf\") pod \"swift-ring-rebalance-debug-5mxqq\" (UID: \"61d5f204-7421-466f-a993-22e4608be96a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mxqq" Mar 07 15:16:27 crc kubenswrapper[4943]: I0307 15:16:27.957856 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4skpm\" (UniqueName: \"kubernetes.io/projected/61d5f204-7421-466f-a993-22e4608be96a-kube-api-access-4skpm\") pod \"swift-ring-rebalance-debug-5mxqq\" (UID: \"61d5f204-7421-466f-a993-22e4608be96a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mxqq" Mar 07 15:16:27 crc kubenswrapper[4943]: I0307 15:16:27.958007 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/61d5f204-7421-466f-a993-22e4608be96a-scripts\") pod \"swift-ring-rebalance-debug-5mxqq\" (UID: \"61d5f204-7421-466f-a993-22e4608be96a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mxqq" Mar 07 15:16:27 crc kubenswrapper[4943]: I0307 15:16:27.958066 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/61d5f204-7421-466f-a993-22e4608be96a-dispersionconf\") pod \"swift-ring-rebalance-debug-5mxqq\" (UID: \"61d5f204-7421-466f-a993-22e4608be96a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mxqq" Mar 07 15:16:27 crc kubenswrapper[4943]: I0307 15:16:27.958135 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/61d5f204-7421-466f-a993-22e4608be96a-swiftconf\") pod \"swift-ring-rebalance-debug-5mxqq\" (UID: \"61d5f204-7421-466f-a993-22e4608be96a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mxqq" Mar 07 15:16:27 crc kubenswrapper[4943]: I0307 15:16:27.958228 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/61d5f204-7421-466f-a993-22e4608be96a-ring-data-devices\") pod \"swift-ring-rebalance-debug-5mxqq\" (UID: \"61d5f204-7421-466f-a993-22e4608be96a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mxqq" Mar 07 15:16:27 crc kubenswrapper[4943]: I0307 15:16:27.958298 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/61d5f204-7421-466f-a993-22e4608be96a-etc-swift\") pod \"swift-ring-rebalance-debug-5mxqq\" (UID: \"61d5f204-7421-466f-a993-22e4608be96a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mxqq" Mar 07 15:16:27 crc kubenswrapper[4943]: I0307 15:16:27.959209 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/61d5f204-7421-466f-a993-22e4608be96a-etc-swift\") pod \"swift-ring-rebalance-debug-5mxqq\" (UID: \"61d5f204-7421-466f-a993-22e4608be96a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mxqq" Mar 07 15:16:27 crc kubenswrapper[4943]: I0307 15:16:27.959568 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/61d5f204-7421-466f-a993-22e4608be96a-scripts\") pod \"swift-ring-rebalance-debug-5mxqq\" (UID: \"61d5f204-7421-466f-a993-22e4608be96a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mxqq" Mar 07 15:16:27 crc kubenswrapper[4943]: I0307 15:16:27.959889 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/61d5f204-7421-466f-a993-22e4608be96a-ring-data-devices\") pod \"swift-ring-rebalance-debug-5mxqq\" (UID: \"61d5f204-7421-466f-a993-22e4608be96a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mxqq" Mar 07 15:16:27 crc kubenswrapper[4943]: I0307 15:16:27.964048 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/61d5f204-7421-466f-a993-22e4608be96a-dispersionconf\") pod \"swift-ring-rebalance-debug-5mxqq\" (UID: \"61d5f204-7421-466f-a993-22e4608be96a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mxqq" Mar 07 15:16:27 crc kubenswrapper[4943]: I0307 15:16:27.964794 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/61d5f204-7421-466f-a993-22e4608be96a-swiftconf\") pod \"swift-ring-rebalance-debug-5mxqq\" (UID: \"61d5f204-7421-466f-a993-22e4608be96a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mxqq" Mar 07 15:16:27 crc kubenswrapper[4943]: I0307 15:16:27.991115 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4skpm\" (UniqueName: \"kubernetes.io/projected/61d5f204-7421-466f-a993-22e4608be96a-kube-api-access-4skpm\") pod \"swift-ring-rebalance-debug-5mxqq\" (UID: \"61d5f204-7421-466f-a993-22e4608be96a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mxqq" Mar 07 15:16:28 crc kubenswrapper[4943]: I0307 15:16:28.061965 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mxqq" Mar 07 15:16:29 crc kubenswrapper[4943]: I0307 15:16:29.158343 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-5mxqq"] Mar 07 15:16:29 crc kubenswrapper[4943]: W0307 15:16:29.168867 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod61d5f204_7421_466f_a993_22e4608be96a.slice/crio-2365ed2809a875639c1c851d51453491ed7990c9969099ae9dfc4df8690babc1 WatchSource:0}: Error finding container 2365ed2809a875639c1c851d51453491ed7990c9969099ae9dfc4df8690babc1: Status 404 returned error can't find the container with id 2365ed2809a875639c1c851d51453491ed7990c9969099ae9dfc4df8690babc1 Mar 07 15:16:29 crc kubenswrapper[4943]: I0307 15:16:29.193887 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mxqq" event={"ID":"61d5f204-7421-466f-a993-22e4608be96a","Type":"ContainerStarted","Data":"2365ed2809a875639c1c851d51453491ed7990c9969099ae9dfc4df8690babc1"} Mar 07 15:16:30 crc kubenswrapper[4943]: I0307 15:16:30.207455 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mxqq" event={"ID":"61d5f204-7421-466f-a993-22e4608be96a","Type":"ContainerStarted","Data":"7d6cf01202fa6a363e293598c058fe9f61876f954e9ade445dc4135733d76d72"} Mar 07 15:16:30 crc kubenswrapper[4943]: I0307 15:16:30.232966 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mxqq" podStartSLOduration=3.232920487 podStartE2EDuration="3.232920487s" podCreationTimestamp="2026-03-07 15:16:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:16:30.224386299 +0000 UTC m=+2232.176522837" watchObservedRunningTime="2026-03-07 15:16:30.232920487 +0000 UTC m=+2232.185056985" Mar 07 15:16:31 crc kubenswrapper[4943]: I0307 15:16:31.221569 4943 generic.go:334] "Generic (PLEG): container finished" podID="61d5f204-7421-466f-a993-22e4608be96a" containerID="7d6cf01202fa6a363e293598c058fe9f61876f954e9ade445dc4135733d76d72" exitCode=0 Mar 07 15:16:31 crc kubenswrapper[4943]: I0307 15:16:31.221635 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mxqq" event={"ID":"61d5f204-7421-466f-a993-22e4608be96a","Type":"ContainerDied","Data":"7d6cf01202fa6a363e293598c058fe9f61876f954e9ade445dc4135733d76d72"} Mar 07 15:16:32 crc kubenswrapper[4943]: I0307 15:16:32.566655 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mxqq" Mar 07 15:16:32 crc kubenswrapper[4943]: I0307 15:16:32.600049 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-5mxqq"] Mar 07 15:16:32 crc kubenswrapper[4943]: I0307 15:16:32.604927 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-5mxqq"] Mar 07 15:16:32 crc kubenswrapper[4943]: I0307 15:16:32.752374 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/61d5f204-7421-466f-a993-22e4608be96a-etc-swift\") pod \"61d5f204-7421-466f-a993-22e4608be96a\" (UID: \"61d5f204-7421-466f-a993-22e4608be96a\") " Mar 07 15:16:32 crc kubenswrapper[4943]: I0307 15:16:32.752472 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/61d5f204-7421-466f-a993-22e4608be96a-dispersionconf\") pod \"61d5f204-7421-466f-a993-22e4608be96a\" (UID: \"61d5f204-7421-466f-a993-22e4608be96a\") " Mar 07 15:16:32 crc kubenswrapper[4943]: I0307 15:16:32.752523 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/61d5f204-7421-466f-a993-22e4608be96a-swiftconf\") pod \"61d5f204-7421-466f-a993-22e4608be96a\" (UID: \"61d5f204-7421-466f-a993-22e4608be96a\") " Mar 07 15:16:32 crc kubenswrapper[4943]: I0307 15:16:32.752584 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4skpm\" (UniqueName: \"kubernetes.io/projected/61d5f204-7421-466f-a993-22e4608be96a-kube-api-access-4skpm\") pod \"61d5f204-7421-466f-a993-22e4608be96a\" (UID: \"61d5f204-7421-466f-a993-22e4608be96a\") " Mar 07 15:16:32 crc kubenswrapper[4943]: I0307 15:16:32.752648 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/61d5f204-7421-466f-a993-22e4608be96a-scripts\") pod \"61d5f204-7421-466f-a993-22e4608be96a\" (UID: \"61d5f204-7421-466f-a993-22e4608be96a\") " Mar 07 15:16:32 crc kubenswrapper[4943]: I0307 15:16:32.752691 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/61d5f204-7421-466f-a993-22e4608be96a-ring-data-devices\") pod \"61d5f204-7421-466f-a993-22e4608be96a\" (UID: \"61d5f204-7421-466f-a993-22e4608be96a\") " Mar 07 15:16:32 crc kubenswrapper[4943]: I0307 15:16:32.752987 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61d5f204-7421-466f-a993-22e4608be96a-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "61d5f204-7421-466f-a993-22e4608be96a" (UID: "61d5f204-7421-466f-a993-22e4608be96a"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:16:32 crc kubenswrapper[4943]: I0307 15:16:32.753305 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/61d5f204-7421-466f-a993-22e4608be96a-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:32 crc kubenswrapper[4943]: I0307 15:16:32.753535 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61d5f204-7421-466f-a993-22e4608be96a-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "61d5f204-7421-466f-a993-22e4608be96a" (UID: "61d5f204-7421-466f-a993-22e4608be96a"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:16:32 crc kubenswrapper[4943]: I0307 15:16:32.762993 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61d5f204-7421-466f-a993-22e4608be96a-kube-api-access-4skpm" (OuterVolumeSpecName: "kube-api-access-4skpm") pod "61d5f204-7421-466f-a993-22e4608be96a" (UID: "61d5f204-7421-466f-a993-22e4608be96a"). InnerVolumeSpecName "kube-api-access-4skpm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:16:32 crc kubenswrapper[4943]: I0307 15:16:32.793519 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61d5f204-7421-466f-a993-22e4608be96a-scripts" (OuterVolumeSpecName: "scripts") pod "61d5f204-7421-466f-a993-22e4608be96a" (UID: "61d5f204-7421-466f-a993-22e4608be96a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:16:32 crc kubenswrapper[4943]: I0307 15:16:32.803852 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61d5f204-7421-466f-a993-22e4608be96a-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "61d5f204-7421-466f-a993-22e4608be96a" (UID: "61d5f204-7421-466f-a993-22e4608be96a"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:16:32 crc kubenswrapper[4943]: I0307 15:16:32.818126 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61d5f204-7421-466f-a993-22e4608be96a-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "61d5f204-7421-466f-a993-22e4608be96a" (UID: "61d5f204-7421-466f-a993-22e4608be96a"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:16:32 crc kubenswrapper[4943]: I0307 15:16:32.854446 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/61d5f204-7421-466f-a993-22e4608be96a-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:32 crc kubenswrapper[4943]: I0307 15:16:32.854474 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4skpm\" (UniqueName: \"kubernetes.io/projected/61d5f204-7421-466f-a993-22e4608be96a-kube-api-access-4skpm\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:32 crc kubenswrapper[4943]: I0307 15:16:32.854486 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/61d5f204-7421-466f-a993-22e4608be96a-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:32 crc kubenswrapper[4943]: I0307 15:16:32.854495 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/61d5f204-7421-466f-a993-22e4608be96a-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:32 crc kubenswrapper[4943]: I0307 15:16:32.854503 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/61d5f204-7421-466f-a993-22e4608be96a-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:33 crc kubenswrapper[4943]: I0307 15:16:33.249876 4943 scope.go:117] "RemoveContainer" containerID="7d6cf01202fa6a363e293598c058fe9f61876f954e9ade445dc4135733d76d72" Mar 07 15:16:33 crc kubenswrapper[4943]: I0307 15:16:33.249948 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mxqq" Mar 07 15:16:33 crc kubenswrapper[4943]: E0307 15:16:33.412757 4943 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod61d5f204_7421_466f_a993_22e4608be96a.slice\": RecentStats: unable to find data in memory cache]" Mar 07 15:16:33 crc kubenswrapper[4943]: I0307 15:16:33.765868 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-wppg5"] Mar 07 15:16:33 crc kubenswrapper[4943]: E0307 15:16:33.766220 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61d5f204-7421-466f-a993-22e4608be96a" containerName="swift-ring-rebalance" Mar 07 15:16:33 crc kubenswrapper[4943]: I0307 15:16:33.766236 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="61d5f204-7421-466f-a993-22e4608be96a" containerName="swift-ring-rebalance" Mar 07 15:16:33 crc kubenswrapper[4943]: I0307 15:16:33.766427 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="61d5f204-7421-466f-a993-22e4608be96a" containerName="swift-ring-rebalance" Mar 07 15:16:33 crc kubenswrapper[4943]: I0307 15:16:33.767266 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wppg5" Mar 07 15:16:33 crc kubenswrapper[4943]: I0307 15:16:33.769185 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:16:33 crc kubenswrapper[4943]: I0307 15:16:33.769480 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:16:33 crc kubenswrapper[4943]: I0307 15:16:33.776724 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-wppg5"] Mar 07 15:16:33 crc kubenswrapper[4943]: I0307 15:16:33.784281 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b5b3f507-50b2-40c5-be8a-dd4a1a8fc414-dispersionconf\") pod \"swift-ring-rebalance-debug-wppg5\" (UID: \"b5b3f507-50b2-40c5-be8a-dd4a1a8fc414\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wppg5" Mar 07 15:16:33 crc kubenswrapper[4943]: I0307 15:16:33.784375 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b5b3f507-50b2-40c5-be8a-dd4a1a8fc414-ring-data-devices\") pod \"swift-ring-rebalance-debug-wppg5\" (UID: \"b5b3f507-50b2-40c5-be8a-dd4a1a8fc414\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wppg5" Mar 07 15:16:33 crc kubenswrapper[4943]: I0307 15:16:33.784419 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b5b3f507-50b2-40c5-be8a-dd4a1a8fc414-swiftconf\") pod \"swift-ring-rebalance-debug-wppg5\" (UID: \"b5b3f507-50b2-40c5-be8a-dd4a1a8fc414\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wppg5" Mar 07 15:16:33 crc kubenswrapper[4943]: I0307 15:16:33.784444 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwp8h\" (UniqueName: \"kubernetes.io/projected/b5b3f507-50b2-40c5-be8a-dd4a1a8fc414-kube-api-access-fwp8h\") pod \"swift-ring-rebalance-debug-wppg5\" (UID: \"b5b3f507-50b2-40c5-be8a-dd4a1a8fc414\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wppg5" Mar 07 15:16:33 crc kubenswrapper[4943]: I0307 15:16:33.784516 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b5b3f507-50b2-40c5-be8a-dd4a1a8fc414-scripts\") pod \"swift-ring-rebalance-debug-wppg5\" (UID: \"b5b3f507-50b2-40c5-be8a-dd4a1a8fc414\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wppg5" Mar 07 15:16:33 crc kubenswrapper[4943]: I0307 15:16:33.784543 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b5b3f507-50b2-40c5-be8a-dd4a1a8fc414-etc-swift\") pod \"swift-ring-rebalance-debug-wppg5\" (UID: \"b5b3f507-50b2-40c5-be8a-dd4a1a8fc414\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wppg5" Mar 07 15:16:33 crc kubenswrapper[4943]: I0307 15:16:33.885807 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b5b3f507-50b2-40c5-be8a-dd4a1a8fc414-scripts\") pod \"swift-ring-rebalance-debug-wppg5\" (UID: \"b5b3f507-50b2-40c5-be8a-dd4a1a8fc414\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wppg5" Mar 07 15:16:33 crc kubenswrapper[4943]: I0307 15:16:33.885888 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b5b3f507-50b2-40c5-be8a-dd4a1a8fc414-etc-swift\") pod \"swift-ring-rebalance-debug-wppg5\" (UID: \"b5b3f507-50b2-40c5-be8a-dd4a1a8fc414\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wppg5" Mar 07 15:16:33 crc kubenswrapper[4943]: I0307 15:16:33.885979 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b5b3f507-50b2-40c5-be8a-dd4a1a8fc414-dispersionconf\") pod \"swift-ring-rebalance-debug-wppg5\" (UID: \"b5b3f507-50b2-40c5-be8a-dd4a1a8fc414\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wppg5" Mar 07 15:16:33 crc kubenswrapper[4943]: I0307 15:16:33.886098 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b5b3f507-50b2-40c5-be8a-dd4a1a8fc414-ring-data-devices\") pod \"swift-ring-rebalance-debug-wppg5\" (UID: \"b5b3f507-50b2-40c5-be8a-dd4a1a8fc414\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wppg5" Mar 07 15:16:33 crc kubenswrapper[4943]: I0307 15:16:33.886198 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b5b3f507-50b2-40c5-be8a-dd4a1a8fc414-swiftconf\") pod \"swift-ring-rebalance-debug-wppg5\" (UID: \"b5b3f507-50b2-40c5-be8a-dd4a1a8fc414\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wppg5" Mar 07 15:16:33 crc kubenswrapper[4943]: I0307 15:16:33.886251 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwp8h\" (UniqueName: \"kubernetes.io/projected/b5b3f507-50b2-40c5-be8a-dd4a1a8fc414-kube-api-access-fwp8h\") pod \"swift-ring-rebalance-debug-wppg5\" (UID: \"b5b3f507-50b2-40c5-be8a-dd4a1a8fc414\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wppg5" Mar 07 15:16:33 crc kubenswrapper[4943]: I0307 15:16:33.886474 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b5b3f507-50b2-40c5-be8a-dd4a1a8fc414-etc-swift\") pod \"swift-ring-rebalance-debug-wppg5\" (UID: \"b5b3f507-50b2-40c5-be8a-dd4a1a8fc414\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wppg5" Mar 07 15:16:33 crc kubenswrapper[4943]: I0307 15:16:33.886897 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b5b3f507-50b2-40c5-be8a-dd4a1a8fc414-scripts\") pod \"swift-ring-rebalance-debug-wppg5\" (UID: \"b5b3f507-50b2-40c5-be8a-dd4a1a8fc414\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wppg5" Mar 07 15:16:33 crc kubenswrapper[4943]: I0307 15:16:33.887141 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b5b3f507-50b2-40c5-be8a-dd4a1a8fc414-ring-data-devices\") pod \"swift-ring-rebalance-debug-wppg5\" (UID: \"b5b3f507-50b2-40c5-be8a-dd4a1a8fc414\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wppg5" Mar 07 15:16:33 crc kubenswrapper[4943]: I0307 15:16:33.890974 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b5b3f507-50b2-40c5-be8a-dd4a1a8fc414-dispersionconf\") pod \"swift-ring-rebalance-debug-wppg5\" (UID: \"b5b3f507-50b2-40c5-be8a-dd4a1a8fc414\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wppg5" Mar 07 15:16:33 crc kubenswrapper[4943]: I0307 15:16:33.891456 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b5b3f507-50b2-40c5-be8a-dd4a1a8fc414-swiftconf\") pod \"swift-ring-rebalance-debug-wppg5\" (UID: \"b5b3f507-50b2-40c5-be8a-dd4a1a8fc414\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wppg5" Mar 07 15:16:33 crc kubenswrapper[4943]: I0307 15:16:33.914572 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwp8h\" (UniqueName: \"kubernetes.io/projected/b5b3f507-50b2-40c5-be8a-dd4a1a8fc414-kube-api-access-fwp8h\") pod \"swift-ring-rebalance-debug-wppg5\" (UID: \"b5b3f507-50b2-40c5-be8a-dd4a1a8fc414\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wppg5" Mar 07 15:16:34 crc kubenswrapper[4943]: I0307 15:16:34.090295 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wppg5" Mar 07 15:16:34 crc kubenswrapper[4943]: I0307 15:16:34.584392 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-wppg5"] Mar 07 15:16:34 crc kubenswrapper[4943]: I0307 15:16:34.764669 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61d5f204-7421-466f-a993-22e4608be96a" path="/var/lib/kubelet/pods/61d5f204-7421-466f-a993-22e4608be96a/volumes" Mar 07 15:16:35 crc kubenswrapper[4943]: I0307 15:16:35.275428 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wppg5" event={"ID":"b5b3f507-50b2-40c5-be8a-dd4a1a8fc414","Type":"ContainerStarted","Data":"af2f7beea5f0bb15968e7e190112bbc6e1218ddef638e26cd31bfc3092e19963"} Mar 07 15:16:35 crc kubenswrapper[4943]: I0307 15:16:35.275480 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wppg5" event={"ID":"b5b3f507-50b2-40c5-be8a-dd4a1a8fc414","Type":"ContainerStarted","Data":"13756ef3f16cb810f29793acdb74646a89e6d9fa0c0b6ef322c3243a2255677c"} Mar 07 15:16:35 crc kubenswrapper[4943]: I0307 15:16:35.300103 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wppg5" podStartSLOduration=2.300073686 podStartE2EDuration="2.300073686s" podCreationTimestamp="2026-03-07 15:16:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:16:35.292236765 +0000 UTC m=+2237.244373293" watchObservedRunningTime="2026-03-07 15:16:35.300073686 +0000 UTC m=+2237.252210214" Mar 07 15:16:36 crc kubenswrapper[4943]: I0307 15:16:36.074095 4943 patch_prober.go:28] interesting pod/machine-config-daemon-cpdmm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 07 15:16:36 crc kubenswrapper[4943]: I0307 15:16:36.074170 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 07 15:16:36 crc kubenswrapper[4943]: I0307 15:16:36.074224 4943 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" Mar 07 15:16:36 crc kubenswrapper[4943]: I0307 15:16:36.075186 4943 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"448308ad6e34ac922d66b3c90e66d0927407a94dd28633ec350502d69e4ca8aa"} pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 07 15:16:36 crc kubenswrapper[4943]: I0307 15:16:36.075281 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerName="machine-config-daemon" containerID="cri-o://448308ad6e34ac922d66b3c90e66d0927407a94dd28633ec350502d69e4ca8aa" gracePeriod=600 Mar 07 15:16:36 crc kubenswrapper[4943]: E0307 15:16:36.198519 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cpdmm_openshift-machine-config-operator(4c1f18ea-e557-41fc-95dd-ce4e1722a61e)\"" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" Mar 07 15:16:36 crc kubenswrapper[4943]: I0307 15:16:36.292327 4943 generic.go:334] "Generic (PLEG): container finished" podID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerID="448308ad6e34ac922d66b3c90e66d0927407a94dd28633ec350502d69e4ca8aa" exitCode=0 Mar 07 15:16:36 crc kubenswrapper[4943]: I0307 15:16:36.292404 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" event={"ID":"4c1f18ea-e557-41fc-95dd-ce4e1722a61e","Type":"ContainerDied","Data":"448308ad6e34ac922d66b3c90e66d0927407a94dd28633ec350502d69e4ca8aa"} Mar 07 15:16:36 crc kubenswrapper[4943]: I0307 15:16:36.292443 4943 scope.go:117] "RemoveContainer" containerID="796971a8f0687fda2fdc2038d9ae78273781f30a32f7c084798a95ee29d28df7" Mar 07 15:16:36 crc kubenswrapper[4943]: I0307 15:16:36.293072 4943 scope.go:117] "RemoveContainer" containerID="448308ad6e34ac922d66b3c90e66d0927407a94dd28633ec350502d69e4ca8aa" Mar 07 15:16:36 crc kubenswrapper[4943]: E0307 15:16:36.293385 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cpdmm_openshift-machine-config-operator(4c1f18ea-e557-41fc-95dd-ce4e1722a61e)\"" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" Mar 07 15:16:36 crc kubenswrapper[4943]: I0307 15:16:36.296373 4943 generic.go:334] "Generic (PLEG): container finished" podID="b5b3f507-50b2-40c5-be8a-dd4a1a8fc414" containerID="af2f7beea5f0bb15968e7e190112bbc6e1218ddef638e26cd31bfc3092e19963" exitCode=0 Mar 07 15:16:36 crc kubenswrapper[4943]: I0307 15:16:36.296425 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wppg5" event={"ID":"b5b3f507-50b2-40c5-be8a-dd4a1a8fc414","Type":"ContainerDied","Data":"af2f7beea5f0bb15968e7e190112bbc6e1218ddef638e26cd31bfc3092e19963"} Mar 07 15:16:37 crc kubenswrapper[4943]: I0307 15:16:37.635167 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wppg5" Mar 07 15:16:37 crc kubenswrapper[4943]: I0307 15:16:37.677777 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-wppg5"] Mar 07 15:16:37 crc kubenswrapper[4943]: I0307 15:16:37.684202 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-wppg5"] Mar 07 15:16:37 crc kubenswrapper[4943]: I0307 15:16:37.754639 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b5b3f507-50b2-40c5-be8a-dd4a1a8fc414-scripts\") pod \"b5b3f507-50b2-40c5-be8a-dd4a1a8fc414\" (UID: \"b5b3f507-50b2-40c5-be8a-dd4a1a8fc414\") " Mar 07 15:16:37 crc kubenswrapper[4943]: I0307 15:16:37.754726 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b5b3f507-50b2-40c5-be8a-dd4a1a8fc414-ring-data-devices\") pod \"b5b3f507-50b2-40c5-be8a-dd4a1a8fc414\" (UID: \"b5b3f507-50b2-40c5-be8a-dd4a1a8fc414\") " Mar 07 15:16:37 crc kubenswrapper[4943]: I0307 15:16:37.754784 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b5b3f507-50b2-40c5-be8a-dd4a1a8fc414-etc-swift\") pod \"b5b3f507-50b2-40c5-be8a-dd4a1a8fc414\" (UID: \"b5b3f507-50b2-40c5-be8a-dd4a1a8fc414\") " Mar 07 15:16:37 crc kubenswrapper[4943]: I0307 15:16:37.754870 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b5b3f507-50b2-40c5-be8a-dd4a1a8fc414-swiftconf\") pod \"b5b3f507-50b2-40c5-be8a-dd4a1a8fc414\" (UID: \"b5b3f507-50b2-40c5-be8a-dd4a1a8fc414\") " Mar 07 15:16:37 crc kubenswrapper[4943]: I0307 15:16:37.754973 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b5b3f507-50b2-40c5-be8a-dd4a1a8fc414-dispersionconf\") pod \"b5b3f507-50b2-40c5-be8a-dd4a1a8fc414\" (UID: \"b5b3f507-50b2-40c5-be8a-dd4a1a8fc414\") " Mar 07 15:16:37 crc kubenswrapper[4943]: I0307 15:16:37.755030 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwp8h\" (UniqueName: \"kubernetes.io/projected/b5b3f507-50b2-40c5-be8a-dd4a1a8fc414-kube-api-access-fwp8h\") pod \"b5b3f507-50b2-40c5-be8a-dd4a1a8fc414\" (UID: \"b5b3f507-50b2-40c5-be8a-dd4a1a8fc414\") " Mar 07 15:16:37 crc kubenswrapper[4943]: I0307 15:16:37.756158 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5b3f507-50b2-40c5-be8a-dd4a1a8fc414-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "b5b3f507-50b2-40c5-be8a-dd4a1a8fc414" (UID: "b5b3f507-50b2-40c5-be8a-dd4a1a8fc414"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:16:37 crc kubenswrapper[4943]: I0307 15:16:37.756706 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5b3f507-50b2-40c5-be8a-dd4a1a8fc414-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "b5b3f507-50b2-40c5-be8a-dd4a1a8fc414" (UID: "b5b3f507-50b2-40c5-be8a-dd4a1a8fc414"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:16:37 crc kubenswrapper[4943]: I0307 15:16:37.765648 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5b3f507-50b2-40c5-be8a-dd4a1a8fc414-kube-api-access-fwp8h" (OuterVolumeSpecName: "kube-api-access-fwp8h") pod "b5b3f507-50b2-40c5-be8a-dd4a1a8fc414" (UID: "b5b3f507-50b2-40c5-be8a-dd4a1a8fc414"). InnerVolumeSpecName "kube-api-access-fwp8h". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:16:37 crc kubenswrapper[4943]: I0307 15:16:37.782478 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5b3f507-50b2-40c5-be8a-dd4a1a8fc414-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "b5b3f507-50b2-40c5-be8a-dd4a1a8fc414" (UID: "b5b3f507-50b2-40c5-be8a-dd4a1a8fc414"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:16:37 crc kubenswrapper[4943]: I0307 15:16:37.788302 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5b3f507-50b2-40c5-be8a-dd4a1a8fc414-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "b5b3f507-50b2-40c5-be8a-dd4a1a8fc414" (UID: "b5b3f507-50b2-40c5-be8a-dd4a1a8fc414"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:16:37 crc kubenswrapper[4943]: I0307 15:16:37.788811 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5b3f507-50b2-40c5-be8a-dd4a1a8fc414-scripts" (OuterVolumeSpecName: "scripts") pod "b5b3f507-50b2-40c5-be8a-dd4a1a8fc414" (UID: "b5b3f507-50b2-40c5-be8a-dd4a1a8fc414"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:16:37 crc kubenswrapper[4943]: I0307 15:16:37.857209 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b5b3f507-50b2-40c5-be8a-dd4a1a8fc414-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:37 crc kubenswrapper[4943]: I0307 15:16:37.857247 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b5b3f507-50b2-40c5-be8a-dd4a1a8fc414-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:37 crc kubenswrapper[4943]: I0307 15:16:37.857259 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwp8h\" (UniqueName: \"kubernetes.io/projected/b5b3f507-50b2-40c5-be8a-dd4a1a8fc414-kube-api-access-fwp8h\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:37 crc kubenswrapper[4943]: I0307 15:16:37.857272 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b5b3f507-50b2-40c5-be8a-dd4a1a8fc414-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:37 crc kubenswrapper[4943]: I0307 15:16:37.857285 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b5b3f507-50b2-40c5-be8a-dd4a1a8fc414-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:37 crc kubenswrapper[4943]: I0307 15:16:37.857295 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b5b3f507-50b2-40c5-be8a-dd4a1a8fc414-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:38 crc kubenswrapper[4943]: I0307 15:16:38.326724 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="13756ef3f16cb810f29793acdb74646a89e6d9fa0c0b6ef322c3243a2255677c" Mar 07 15:16:38 crc kubenswrapper[4943]: I0307 15:16:38.326837 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wppg5" Mar 07 15:16:38 crc kubenswrapper[4943]: I0307 15:16:38.770288 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5b3f507-50b2-40c5-be8a-dd4a1a8fc414" path="/var/lib/kubelet/pods/b5b3f507-50b2-40c5-be8a-dd4a1a8fc414/volumes" Mar 07 15:16:38 crc kubenswrapper[4943]: I0307 15:16:38.854868 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-pqk4g"] Mar 07 15:16:38 crc kubenswrapper[4943]: E0307 15:16:38.855314 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5b3f507-50b2-40c5-be8a-dd4a1a8fc414" containerName="swift-ring-rebalance" Mar 07 15:16:38 crc kubenswrapper[4943]: I0307 15:16:38.855344 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5b3f507-50b2-40c5-be8a-dd4a1a8fc414" containerName="swift-ring-rebalance" Mar 07 15:16:38 crc kubenswrapper[4943]: I0307 15:16:38.855627 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5b3f507-50b2-40c5-be8a-dd4a1a8fc414" containerName="swift-ring-rebalance" Mar 07 15:16:38 crc kubenswrapper[4943]: I0307 15:16:38.856427 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pqk4g" Mar 07 15:16:38 crc kubenswrapper[4943]: I0307 15:16:38.859213 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:16:38 crc kubenswrapper[4943]: I0307 15:16:38.860223 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:16:38 crc kubenswrapper[4943]: I0307 15:16:38.875447 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-pqk4g"] Mar 07 15:16:38 crc kubenswrapper[4943]: I0307 15:16:38.977439 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/807d9e51-601e-4178-b773-26c1357d2b85-swiftconf\") pod \"swift-ring-rebalance-debug-pqk4g\" (UID: \"807d9e51-601e-4178-b773-26c1357d2b85\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pqk4g" Mar 07 15:16:38 crc kubenswrapper[4943]: I0307 15:16:38.977618 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/807d9e51-601e-4178-b773-26c1357d2b85-dispersionconf\") pod \"swift-ring-rebalance-debug-pqk4g\" (UID: \"807d9e51-601e-4178-b773-26c1357d2b85\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pqk4g" Mar 07 15:16:38 crc kubenswrapper[4943]: I0307 15:16:38.977708 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6kzh9\" (UniqueName: \"kubernetes.io/projected/807d9e51-601e-4178-b773-26c1357d2b85-kube-api-access-6kzh9\") pod \"swift-ring-rebalance-debug-pqk4g\" (UID: \"807d9e51-601e-4178-b773-26c1357d2b85\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pqk4g" Mar 07 15:16:38 crc kubenswrapper[4943]: I0307 15:16:38.977834 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/807d9e51-601e-4178-b773-26c1357d2b85-scripts\") pod \"swift-ring-rebalance-debug-pqk4g\" (UID: \"807d9e51-601e-4178-b773-26c1357d2b85\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pqk4g" Mar 07 15:16:38 crc kubenswrapper[4943]: I0307 15:16:38.977895 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/807d9e51-601e-4178-b773-26c1357d2b85-ring-data-devices\") pod \"swift-ring-rebalance-debug-pqk4g\" (UID: \"807d9e51-601e-4178-b773-26c1357d2b85\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pqk4g" Mar 07 15:16:38 crc kubenswrapper[4943]: I0307 15:16:38.978054 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/807d9e51-601e-4178-b773-26c1357d2b85-etc-swift\") pod \"swift-ring-rebalance-debug-pqk4g\" (UID: \"807d9e51-601e-4178-b773-26c1357d2b85\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pqk4g" Mar 07 15:16:39 crc kubenswrapper[4943]: I0307 15:16:39.079984 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/807d9e51-601e-4178-b773-26c1357d2b85-swiftconf\") pod \"swift-ring-rebalance-debug-pqk4g\" (UID: \"807d9e51-601e-4178-b773-26c1357d2b85\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pqk4g" Mar 07 15:16:39 crc kubenswrapper[4943]: I0307 15:16:39.080110 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/807d9e51-601e-4178-b773-26c1357d2b85-dispersionconf\") pod \"swift-ring-rebalance-debug-pqk4g\" (UID: \"807d9e51-601e-4178-b773-26c1357d2b85\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pqk4g" Mar 07 15:16:39 crc kubenswrapper[4943]: I0307 15:16:39.080153 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6kzh9\" (UniqueName: \"kubernetes.io/projected/807d9e51-601e-4178-b773-26c1357d2b85-kube-api-access-6kzh9\") pod \"swift-ring-rebalance-debug-pqk4g\" (UID: \"807d9e51-601e-4178-b773-26c1357d2b85\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pqk4g" Mar 07 15:16:39 crc kubenswrapper[4943]: I0307 15:16:39.080216 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/807d9e51-601e-4178-b773-26c1357d2b85-scripts\") pod \"swift-ring-rebalance-debug-pqk4g\" (UID: \"807d9e51-601e-4178-b773-26c1357d2b85\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pqk4g" Mar 07 15:16:39 crc kubenswrapper[4943]: I0307 15:16:39.080256 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/807d9e51-601e-4178-b773-26c1357d2b85-ring-data-devices\") pod \"swift-ring-rebalance-debug-pqk4g\" (UID: \"807d9e51-601e-4178-b773-26c1357d2b85\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pqk4g" Mar 07 15:16:39 crc kubenswrapper[4943]: I0307 15:16:39.080321 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/807d9e51-601e-4178-b773-26c1357d2b85-etc-swift\") pod \"swift-ring-rebalance-debug-pqk4g\" (UID: \"807d9e51-601e-4178-b773-26c1357d2b85\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pqk4g" Mar 07 15:16:39 crc kubenswrapper[4943]: I0307 15:16:39.082099 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/807d9e51-601e-4178-b773-26c1357d2b85-etc-swift\") pod \"swift-ring-rebalance-debug-pqk4g\" (UID: \"807d9e51-601e-4178-b773-26c1357d2b85\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pqk4g" Mar 07 15:16:39 crc kubenswrapper[4943]: I0307 15:16:39.083143 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/807d9e51-601e-4178-b773-26c1357d2b85-scripts\") pod \"swift-ring-rebalance-debug-pqk4g\" (UID: \"807d9e51-601e-4178-b773-26c1357d2b85\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pqk4g" Mar 07 15:16:39 crc kubenswrapper[4943]: I0307 15:16:39.083175 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/807d9e51-601e-4178-b773-26c1357d2b85-ring-data-devices\") pod \"swift-ring-rebalance-debug-pqk4g\" (UID: \"807d9e51-601e-4178-b773-26c1357d2b85\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pqk4g" Mar 07 15:16:39 crc kubenswrapper[4943]: I0307 15:16:39.085694 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/807d9e51-601e-4178-b773-26c1357d2b85-dispersionconf\") pod \"swift-ring-rebalance-debug-pqk4g\" (UID: \"807d9e51-601e-4178-b773-26c1357d2b85\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pqk4g" Mar 07 15:16:39 crc kubenswrapper[4943]: I0307 15:16:39.086491 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/807d9e51-601e-4178-b773-26c1357d2b85-swiftconf\") pod \"swift-ring-rebalance-debug-pqk4g\" (UID: \"807d9e51-601e-4178-b773-26c1357d2b85\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pqk4g" Mar 07 15:16:39 crc kubenswrapper[4943]: I0307 15:16:39.108018 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6kzh9\" (UniqueName: \"kubernetes.io/projected/807d9e51-601e-4178-b773-26c1357d2b85-kube-api-access-6kzh9\") pod \"swift-ring-rebalance-debug-pqk4g\" (UID: \"807d9e51-601e-4178-b773-26c1357d2b85\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pqk4g" Mar 07 15:16:39 crc kubenswrapper[4943]: I0307 15:16:39.189270 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pqk4g" Mar 07 15:16:39 crc kubenswrapper[4943]: I0307 15:16:39.484915 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-pqk4g"] Mar 07 15:16:40 crc kubenswrapper[4943]: I0307 15:16:40.358148 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pqk4g" event={"ID":"807d9e51-601e-4178-b773-26c1357d2b85","Type":"ContainerStarted","Data":"3bd33192c0d7c38312fb29368fbc516d43db1bc3bf3816323cbacf54b8f2a1d2"} Mar 07 15:16:40 crc kubenswrapper[4943]: I0307 15:16:40.358511 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pqk4g" event={"ID":"807d9e51-601e-4178-b773-26c1357d2b85","Type":"ContainerStarted","Data":"0e25162587f29223eafd2ba9090fb2e25a0d02ce0d5549e7621fb3f3211b7aaf"} Mar 07 15:16:40 crc kubenswrapper[4943]: I0307 15:16:40.392666 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pqk4g" podStartSLOduration=2.392639577 podStartE2EDuration="2.392639577s" podCreationTimestamp="2026-03-07 15:16:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:16:40.380975622 +0000 UTC m=+2242.333112180" watchObservedRunningTime="2026-03-07 15:16:40.392639577 +0000 UTC m=+2242.344776115" Mar 07 15:16:41 crc kubenswrapper[4943]: I0307 15:16:41.368190 4943 generic.go:334] "Generic (PLEG): container finished" podID="807d9e51-601e-4178-b773-26c1357d2b85" containerID="3bd33192c0d7c38312fb29368fbc516d43db1bc3bf3816323cbacf54b8f2a1d2" exitCode=0 Mar 07 15:16:41 crc kubenswrapper[4943]: I0307 15:16:41.368251 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pqk4g" event={"ID":"807d9e51-601e-4178-b773-26c1357d2b85","Type":"ContainerDied","Data":"3bd33192c0d7c38312fb29368fbc516d43db1bc3bf3816323cbacf54b8f2a1d2"} Mar 07 15:16:42 crc kubenswrapper[4943]: I0307 15:16:42.801557 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pqk4g" Mar 07 15:16:42 crc kubenswrapper[4943]: I0307 15:16:42.857271 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-pqk4g"] Mar 07 15:16:42 crc kubenswrapper[4943]: I0307 15:16:42.884005 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-pqk4g"] Mar 07 15:16:42 crc kubenswrapper[4943]: I0307 15:16:42.940702 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/807d9e51-601e-4178-b773-26c1357d2b85-dispersionconf\") pod \"807d9e51-601e-4178-b773-26c1357d2b85\" (UID: \"807d9e51-601e-4178-b773-26c1357d2b85\") " Mar 07 15:16:42 crc kubenswrapper[4943]: I0307 15:16:42.940760 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/807d9e51-601e-4178-b773-26c1357d2b85-etc-swift\") pod \"807d9e51-601e-4178-b773-26c1357d2b85\" (UID: \"807d9e51-601e-4178-b773-26c1357d2b85\") " Mar 07 15:16:42 crc kubenswrapper[4943]: I0307 15:16:42.940829 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/807d9e51-601e-4178-b773-26c1357d2b85-swiftconf\") pod \"807d9e51-601e-4178-b773-26c1357d2b85\" (UID: \"807d9e51-601e-4178-b773-26c1357d2b85\") " Mar 07 15:16:42 crc kubenswrapper[4943]: I0307 15:16:42.940906 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/807d9e51-601e-4178-b773-26c1357d2b85-ring-data-devices\") pod \"807d9e51-601e-4178-b773-26c1357d2b85\" (UID: \"807d9e51-601e-4178-b773-26c1357d2b85\") " Mar 07 15:16:42 crc kubenswrapper[4943]: I0307 15:16:42.941011 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/807d9e51-601e-4178-b773-26c1357d2b85-scripts\") pod \"807d9e51-601e-4178-b773-26c1357d2b85\" (UID: \"807d9e51-601e-4178-b773-26c1357d2b85\") " Mar 07 15:16:42 crc kubenswrapper[4943]: I0307 15:16:42.941036 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6kzh9\" (UniqueName: \"kubernetes.io/projected/807d9e51-601e-4178-b773-26c1357d2b85-kube-api-access-6kzh9\") pod \"807d9e51-601e-4178-b773-26c1357d2b85\" (UID: \"807d9e51-601e-4178-b773-26c1357d2b85\") " Mar 07 15:16:42 crc kubenswrapper[4943]: I0307 15:16:42.941887 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/807d9e51-601e-4178-b773-26c1357d2b85-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "807d9e51-601e-4178-b773-26c1357d2b85" (UID: "807d9e51-601e-4178-b773-26c1357d2b85"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:16:42 crc kubenswrapper[4943]: I0307 15:16:42.942430 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/807d9e51-601e-4178-b773-26c1357d2b85-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "807d9e51-601e-4178-b773-26c1357d2b85" (UID: "807d9e51-601e-4178-b773-26c1357d2b85"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:16:42 crc kubenswrapper[4943]: I0307 15:16:42.946714 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/807d9e51-601e-4178-b773-26c1357d2b85-kube-api-access-6kzh9" (OuterVolumeSpecName: "kube-api-access-6kzh9") pod "807d9e51-601e-4178-b773-26c1357d2b85" (UID: "807d9e51-601e-4178-b773-26c1357d2b85"). InnerVolumeSpecName "kube-api-access-6kzh9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:16:42 crc kubenswrapper[4943]: I0307 15:16:42.968739 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/807d9e51-601e-4178-b773-26c1357d2b85-scripts" (OuterVolumeSpecName: "scripts") pod "807d9e51-601e-4178-b773-26c1357d2b85" (UID: "807d9e51-601e-4178-b773-26c1357d2b85"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:16:42 crc kubenswrapper[4943]: I0307 15:16:42.969546 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/807d9e51-601e-4178-b773-26c1357d2b85-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "807d9e51-601e-4178-b773-26c1357d2b85" (UID: "807d9e51-601e-4178-b773-26c1357d2b85"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:16:42 crc kubenswrapper[4943]: I0307 15:16:42.973330 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/807d9e51-601e-4178-b773-26c1357d2b85-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "807d9e51-601e-4178-b773-26c1357d2b85" (UID: "807d9e51-601e-4178-b773-26c1357d2b85"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:16:43 crc kubenswrapper[4943]: I0307 15:16:43.043164 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/807d9e51-601e-4178-b773-26c1357d2b85-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:43 crc kubenswrapper[4943]: I0307 15:16:43.043207 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/807d9e51-601e-4178-b773-26c1357d2b85-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:43 crc kubenswrapper[4943]: I0307 15:16:43.043218 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/807d9e51-601e-4178-b773-26c1357d2b85-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:43 crc kubenswrapper[4943]: I0307 15:16:43.043230 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/807d9e51-601e-4178-b773-26c1357d2b85-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:43 crc kubenswrapper[4943]: I0307 15:16:43.043242 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/807d9e51-601e-4178-b773-26c1357d2b85-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:43 crc kubenswrapper[4943]: I0307 15:16:43.043253 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6kzh9\" (UniqueName: \"kubernetes.io/projected/807d9e51-601e-4178-b773-26c1357d2b85-kube-api-access-6kzh9\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:43 crc kubenswrapper[4943]: I0307 15:16:43.390669 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0e25162587f29223eafd2ba9090fb2e25a0d02ce0d5549e7621fb3f3211b7aaf" Mar 07 15:16:43 crc kubenswrapper[4943]: I0307 15:16:43.390727 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pqk4g" Mar 07 15:16:44 crc kubenswrapper[4943]: I0307 15:16:44.011315 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-5ccgt"] Mar 07 15:16:44 crc kubenswrapper[4943]: E0307 15:16:44.012098 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="807d9e51-601e-4178-b773-26c1357d2b85" containerName="swift-ring-rebalance" Mar 07 15:16:44 crc kubenswrapper[4943]: I0307 15:16:44.012120 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="807d9e51-601e-4178-b773-26c1357d2b85" containerName="swift-ring-rebalance" Mar 07 15:16:44 crc kubenswrapper[4943]: I0307 15:16:44.012391 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="807d9e51-601e-4178-b773-26c1357d2b85" containerName="swift-ring-rebalance" Mar 07 15:16:44 crc kubenswrapper[4943]: I0307 15:16:44.013170 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5ccgt" Mar 07 15:16:44 crc kubenswrapper[4943]: I0307 15:16:44.017676 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:16:44 crc kubenswrapper[4943]: I0307 15:16:44.017823 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:16:44 crc kubenswrapper[4943]: I0307 15:16:44.036606 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-5ccgt"] Mar 07 15:16:44 crc kubenswrapper[4943]: I0307 15:16:44.160307 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff-dispersionconf\") pod \"swift-ring-rebalance-debug-5ccgt\" (UID: \"1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5ccgt" Mar 07 15:16:44 crc kubenswrapper[4943]: I0307 15:16:44.160520 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff-etc-swift\") pod \"swift-ring-rebalance-debug-5ccgt\" (UID: \"1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5ccgt" Mar 07 15:16:44 crc kubenswrapper[4943]: I0307 15:16:44.160581 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwsz6\" (UniqueName: \"kubernetes.io/projected/1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff-kube-api-access-dwsz6\") pod \"swift-ring-rebalance-debug-5ccgt\" (UID: \"1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5ccgt" Mar 07 15:16:44 crc kubenswrapper[4943]: I0307 15:16:44.160738 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff-swiftconf\") pod \"swift-ring-rebalance-debug-5ccgt\" (UID: \"1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5ccgt" Mar 07 15:16:44 crc kubenswrapper[4943]: I0307 15:16:44.160853 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff-ring-data-devices\") pod \"swift-ring-rebalance-debug-5ccgt\" (UID: \"1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5ccgt" Mar 07 15:16:44 crc kubenswrapper[4943]: I0307 15:16:44.160948 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff-scripts\") pod \"swift-ring-rebalance-debug-5ccgt\" (UID: \"1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5ccgt" Mar 07 15:16:44 crc kubenswrapper[4943]: I0307 15:16:44.262741 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff-etc-swift\") pod \"swift-ring-rebalance-debug-5ccgt\" (UID: \"1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5ccgt" Mar 07 15:16:44 crc kubenswrapper[4943]: I0307 15:16:44.262809 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwsz6\" (UniqueName: \"kubernetes.io/projected/1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff-kube-api-access-dwsz6\") pod \"swift-ring-rebalance-debug-5ccgt\" (UID: \"1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5ccgt" Mar 07 15:16:44 crc kubenswrapper[4943]: I0307 15:16:44.262856 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff-swiftconf\") pod \"swift-ring-rebalance-debug-5ccgt\" (UID: \"1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5ccgt" Mar 07 15:16:44 crc kubenswrapper[4943]: I0307 15:16:44.262907 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff-ring-data-devices\") pod \"swift-ring-rebalance-debug-5ccgt\" (UID: \"1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5ccgt" Mar 07 15:16:44 crc kubenswrapper[4943]: I0307 15:16:44.262973 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff-scripts\") pod \"swift-ring-rebalance-debug-5ccgt\" (UID: \"1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5ccgt" Mar 07 15:16:44 crc kubenswrapper[4943]: I0307 15:16:44.263085 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff-dispersionconf\") pod \"swift-ring-rebalance-debug-5ccgt\" (UID: \"1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5ccgt" Mar 07 15:16:44 crc kubenswrapper[4943]: I0307 15:16:44.263535 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff-etc-swift\") pod \"swift-ring-rebalance-debug-5ccgt\" (UID: \"1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5ccgt" Mar 07 15:16:44 crc kubenswrapper[4943]: I0307 15:16:44.264237 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff-ring-data-devices\") pod \"swift-ring-rebalance-debug-5ccgt\" (UID: \"1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5ccgt" Mar 07 15:16:44 crc kubenswrapper[4943]: I0307 15:16:44.264503 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff-scripts\") pod \"swift-ring-rebalance-debug-5ccgt\" (UID: \"1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5ccgt" Mar 07 15:16:44 crc kubenswrapper[4943]: I0307 15:16:44.287919 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff-swiftconf\") pod \"swift-ring-rebalance-debug-5ccgt\" (UID: \"1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5ccgt" Mar 07 15:16:44 crc kubenswrapper[4943]: I0307 15:16:44.288019 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff-dispersionconf\") pod \"swift-ring-rebalance-debug-5ccgt\" (UID: \"1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5ccgt" Mar 07 15:16:44 crc kubenswrapper[4943]: I0307 15:16:44.292653 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwsz6\" (UniqueName: \"kubernetes.io/projected/1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff-kube-api-access-dwsz6\") pod \"swift-ring-rebalance-debug-5ccgt\" (UID: \"1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5ccgt" Mar 07 15:16:44 crc kubenswrapper[4943]: I0307 15:16:44.356962 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5ccgt" Mar 07 15:16:44 crc kubenswrapper[4943]: I0307 15:16:44.663797 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-5ccgt"] Mar 07 15:16:44 crc kubenswrapper[4943]: I0307 15:16:44.765698 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="807d9e51-601e-4178-b773-26c1357d2b85" path="/var/lib/kubelet/pods/807d9e51-601e-4178-b773-26c1357d2b85/volumes" Mar 07 15:16:45 crc kubenswrapper[4943]: I0307 15:16:45.411607 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5ccgt" event={"ID":"1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff","Type":"ContainerStarted","Data":"38cb674cbb0f8f9c160c67ea4a51102703e4b2c2261fffab9b7cca100ed589b1"} Mar 07 15:16:45 crc kubenswrapper[4943]: I0307 15:16:45.412025 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5ccgt" event={"ID":"1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff","Type":"ContainerStarted","Data":"8893f92a399f3cd1a4a76063c4506b663ca52e300c70df2d3b2df1739e176b3a"} Mar 07 15:16:45 crc kubenswrapper[4943]: I0307 15:16:45.436422 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5ccgt" podStartSLOduration=2.4364023550000002 podStartE2EDuration="2.436402355s" podCreationTimestamp="2026-03-07 15:16:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:16:45.432621063 +0000 UTC m=+2247.384757571" watchObservedRunningTime="2026-03-07 15:16:45.436402355 +0000 UTC m=+2247.388538863" Mar 07 15:16:46 crc kubenswrapper[4943]: I0307 15:16:46.423993 4943 generic.go:334] "Generic (PLEG): container finished" podID="1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff" containerID="38cb674cbb0f8f9c160c67ea4a51102703e4b2c2261fffab9b7cca100ed589b1" exitCode=0 Mar 07 15:16:46 crc kubenswrapper[4943]: I0307 15:16:46.424148 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5ccgt" event={"ID":"1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff","Type":"ContainerDied","Data":"38cb674cbb0f8f9c160c67ea4a51102703e4b2c2261fffab9b7cca100ed589b1"} Mar 07 15:16:47 crc kubenswrapper[4943]: I0307 15:16:47.787480 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5ccgt" Mar 07 15:16:47 crc kubenswrapper[4943]: I0307 15:16:47.816839 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-5ccgt"] Mar 07 15:16:47 crc kubenswrapper[4943]: I0307 15:16:47.823042 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-5ccgt"] Mar 07 15:16:47 crc kubenswrapper[4943]: I0307 15:16:47.927971 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff-scripts\") pod \"1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff\" (UID: \"1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff\") " Mar 07 15:16:47 crc kubenswrapper[4943]: I0307 15:16:47.928043 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff-dispersionconf\") pod \"1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff\" (UID: \"1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff\") " Mar 07 15:16:47 crc kubenswrapper[4943]: I0307 15:16:47.928094 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff-ring-data-devices\") pod \"1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff\" (UID: \"1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff\") " Mar 07 15:16:47 crc kubenswrapper[4943]: I0307 15:16:47.928222 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff-etc-swift\") pod \"1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff\" (UID: \"1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff\") " Mar 07 15:16:47 crc kubenswrapper[4943]: I0307 15:16:47.928247 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwsz6\" (UniqueName: \"kubernetes.io/projected/1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff-kube-api-access-dwsz6\") pod \"1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff\" (UID: \"1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff\") " Mar 07 15:16:47 crc kubenswrapper[4943]: I0307 15:16:47.928292 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff-swiftconf\") pod \"1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff\" (UID: \"1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff\") " Mar 07 15:16:47 crc kubenswrapper[4943]: I0307 15:16:47.933083 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff" (UID: "1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:16:47 crc kubenswrapper[4943]: I0307 15:16:47.933700 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff" (UID: "1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:16:47 crc kubenswrapper[4943]: I0307 15:16:47.944065 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff-kube-api-access-dwsz6" (OuterVolumeSpecName: "kube-api-access-dwsz6") pod "1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff" (UID: "1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff"). InnerVolumeSpecName "kube-api-access-dwsz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:16:47 crc kubenswrapper[4943]: I0307 15:16:47.976085 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff" (UID: "1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:16:47 crc kubenswrapper[4943]: I0307 15:16:47.993098 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff" (UID: "1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:16:47 crc kubenswrapper[4943]: I0307 15:16:47.993547 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff-scripts" (OuterVolumeSpecName: "scripts") pod "1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff" (UID: "1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:16:48 crc kubenswrapper[4943]: I0307 15:16:48.029748 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:48 crc kubenswrapper[4943]: I0307 15:16:48.029784 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:48 crc kubenswrapper[4943]: I0307 15:16:48.029793 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwsz6\" (UniqueName: \"kubernetes.io/projected/1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff-kube-api-access-dwsz6\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:48 crc kubenswrapper[4943]: I0307 15:16:48.029803 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:48 crc kubenswrapper[4943]: I0307 15:16:48.029811 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:48 crc kubenswrapper[4943]: I0307 15:16:48.029819 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:48 crc kubenswrapper[4943]: I0307 15:16:48.461300 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8893f92a399f3cd1a4a76063c4506b663ca52e300c70df2d3b2df1739e176b3a" Mar 07 15:16:48 crc kubenswrapper[4943]: I0307 15:16:48.461412 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5ccgt" Mar 07 15:16:48 crc kubenswrapper[4943]: I0307 15:16:48.770917 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff" path="/var/lib/kubelet/pods/1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff/volumes" Mar 07 15:16:49 crc kubenswrapper[4943]: I0307 15:16:49.045892 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-c7gj4"] Mar 07 15:16:49 crc kubenswrapper[4943]: E0307 15:16:49.046425 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff" containerName="swift-ring-rebalance" Mar 07 15:16:49 crc kubenswrapper[4943]: I0307 15:16:49.046450 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff" containerName="swift-ring-rebalance" Mar 07 15:16:49 crc kubenswrapper[4943]: I0307 15:16:49.046728 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="1835edd3-c2d4-4b0f-989b-4a14dc5fe5ff" containerName="swift-ring-rebalance" Mar 07 15:16:49 crc kubenswrapper[4943]: I0307 15:16:49.047661 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-c7gj4" Mar 07 15:16:49 crc kubenswrapper[4943]: I0307 15:16:49.055638 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:16:49 crc kubenswrapper[4943]: I0307 15:16:49.062195 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-c7gj4"] Mar 07 15:16:49 crc kubenswrapper[4943]: I0307 15:16:49.062989 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:16:49 crc kubenswrapper[4943]: I0307 15:16:49.145447 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1490d08b-b319-42bc-9259-ff749a56d136-swiftconf\") pod \"swift-ring-rebalance-debug-c7gj4\" (UID: \"1490d08b-b319-42bc-9259-ff749a56d136\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-c7gj4" Mar 07 15:16:49 crc kubenswrapper[4943]: I0307 15:16:49.145514 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1490d08b-b319-42bc-9259-ff749a56d136-dispersionconf\") pod \"swift-ring-rebalance-debug-c7gj4\" (UID: \"1490d08b-b319-42bc-9259-ff749a56d136\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-c7gj4" Mar 07 15:16:49 crc kubenswrapper[4943]: I0307 15:16:49.145583 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gzf7\" (UniqueName: \"kubernetes.io/projected/1490d08b-b319-42bc-9259-ff749a56d136-kube-api-access-6gzf7\") pod \"swift-ring-rebalance-debug-c7gj4\" (UID: \"1490d08b-b319-42bc-9259-ff749a56d136\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-c7gj4" Mar 07 15:16:49 crc kubenswrapper[4943]: I0307 15:16:49.145793 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1490d08b-b319-42bc-9259-ff749a56d136-ring-data-devices\") pod \"swift-ring-rebalance-debug-c7gj4\" (UID: \"1490d08b-b319-42bc-9259-ff749a56d136\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-c7gj4" Mar 07 15:16:49 crc kubenswrapper[4943]: I0307 15:16:49.145867 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1490d08b-b319-42bc-9259-ff749a56d136-scripts\") pod \"swift-ring-rebalance-debug-c7gj4\" (UID: \"1490d08b-b319-42bc-9259-ff749a56d136\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-c7gj4" Mar 07 15:16:49 crc kubenswrapper[4943]: I0307 15:16:49.145895 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1490d08b-b319-42bc-9259-ff749a56d136-etc-swift\") pod \"swift-ring-rebalance-debug-c7gj4\" (UID: \"1490d08b-b319-42bc-9259-ff749a56d136\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-c7gj4" Mar 07 15:16:49 crc kubenswrapper[4943]: I0307 15:16:49.248042 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gzf7\" (UniqueName: \"kubernetes.io/projected/1490d08b-b319-42bc-9259-ff749a56d136-kube-api-access-6gzf7\") pod \"swift-ring-rebalance-debug-c7gj4\" (UID: \"1490d08b-b319-42bc-9259-ff749a56d136\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-c7gj4" Mar 07 15:16:49 crc kubenswrapper[4943]: I0307 15:16:49.248183 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1490d08b-b319-42bc-9259-ff749a56d136-ring-data-devices\") pod \"swift-ring-rebalance-debug-c7gj4\" (UID: \"1490d08b-b319-42bc-9259-ff749a56d136\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-c7gj4" Mar 07 15:16:49 crc kubenswrapper[4943]: I0307 15:16:49.248232 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1490d08b-b319-42bc-9259-ff749a56d136-scripts\") pod \"swift-ring-rebalance-debug-c7gj4\" (UID: \"1490d08b-b319-42bc-9259-ff749a56d136\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-c7gj4" Mar 07 15:16:49 crc kubenswrapper[4943]: I0307 15:16:49.248265 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1490d08b-b319-42bc-9259-ff749a56d136-etc-swift\") pod \"swift-ring-rebalance-debug-c7gj4\" (UID: \"1490d08b-b319-42bc-9259-ff749a56d136\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-c7gj4" Mar 07 15:16:49 crc kubenswrapper[4943]: I0307 15:16:49.248360 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1490d08b-b319-42bc-9259-ff749a56d136-swiftconf\") pod \"swift-ring-rebalance-debug-c7gj4\" (UID: \"1490d08b-b319-42bc-9259-ff749a56d136\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-c7gj4" Mar 07 15:16:49 crc kubenswrapper[4943]: I0307 15:16:49.248399 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1490d08b-b319-42bc-9259-ff749a56d136-dispersionconf\") pod \"swift-ring-rebalance-debug-c7gj4\" (UID: \"1490d08b-b319-42bc-9259-ff749a56d136\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-c7gj4" Mar 07 15:16:49 crc kubenswrapper[4943]: I0307 15:16:49.248917 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1490d08b-b319-42bc-9259-ff749a56d136-etc-swift\") pod \"swift-ring-rebalance-debug-c7gj4\" (UID: \"1490d08b-b319-42bc-9259-ff749a56d136\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-c7gj4" Mar 07 15:16:49 crc kubenswrapper[4943]: I0307 15:16:49.249264 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1490d08b-b319-42bc-9259-ff749a56d136-scripts\") pod \"swift-ring-rebalance-debug-c7gj4\" (UID: \"1490d08b-b319-42bc-9259-ff749a56d136\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-c7gj4" Mar 07 15:16:49 crc kubenswrapper[4943]: I0307 15:16:49.249794 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1490d08b-b319-42bc-9259-ff749a56d136-ring-data-devices\") pod \"swift-ring-rebalance-debug-c7gj4\" (UID: \"1490d08b-b319-42bc-9259-ff749a56d136\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-c7gj4" Mar 07 15:16:49 crc kubenswrapper[4943]: I0307 15:16:49.255496 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1490d08b-b319-42bc-9259-ff749a56d136-dispersionconf\") pod \"swift-ring-rebalance-debug-c7gj4\" (UID: \"1490d08b-b319-42bc-9259-ff749a56d136\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-c7gj4" Mar 07 15:16:49 crc kubenswrapper[4943]: I0307 15:16:49.256025 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1490d08b-b319-42bc-9259-ff749a56d136-swiftconf\") pod \"swift-ring-rebalance-debug-c7gj4\" (UID: \"1490d08b-b319-42bc-9259-ff749a56d136\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-c7gj4" Mar 07 15:16:49 crc kubenswrapper[4943]: I0307 15:16:49.280148 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gzf7\" (UniqueName: \"kubernetes.io/projected/1490d08b-b319-42bc-9259-ff749a56d136-kube-api-access-6gzf7\") pod \"swift-ring-rebalance-debug-c7gj4\" (UID: \"1490d08b-b319-42bc-9259-ff749a56d136\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-c7gj4" Mar 07 15:16:49 crc kubenswrapper[4943]: I0307 15:16:49.377798 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-c7gj4" Mar 07 15:16:49 crc kubenswrapper[4943]: I0307 15:16:49.691224 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-c7gj4"] Mar 07 15:16:49 crc kubenswrapper[4943]: W0307 15:16:49.700352 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1490d08b_b319_42bc_9259_ff749a56d136.slice/crio-891ff0d0a5a322741661597af57264716bd7f25090eb9e90ca6d764c54ec4ead WatchSource:0}: Error finding container 891ff0d0a5a322741661597af57264716bd7f25090eb9e90ca6d764c54ec4ead: Status 404 returned error can't find the container with id 891ff0d0a5a322741661597af57264716bd7f25090eb9e90ca6d764c54ec4ead Mar 07 15:16:49 crc kubenswrapper[4943]: I0307 15:16:49.755493 4943 scope.go:117] "RemoveContainer" containerID="448308ad6e34ac922d66b3c90e66d0927407a94dd28633ec350502d69e4ca8aa" Mar 07 15:16:49 crc kubenswrapper[4943]: E0307 15:16:49.755695 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cpdmm_openshift-machine-config-operator(4c1f18ea-e557-41fc-95dd-ce4e1722a61e)\"" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" Mar 07 15:16:50 crc kubenswrapper[4943]: I0307 15:16:50.486083 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-c7gj4" event={"ID":"1490d08b-b319-42bc-9259-ff749a56d136","Type":"ContainerStarted","Data":"17364057933ada0f1377538d14281f33b01873033659fcf72f6baf62dda3b21e"} Mar 07 15:16:50 crc kubenswrapper[4943]: I0307 15:16:50.486533 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-c7gj4" event={"ID":"1490d08b-b319-42bc-9259-ff749a56d136","Type":"ContainerStarted","Data":"891ff0d0a5a322741661597af57264716bd7f25090eb9e90ca6d764c54ec4ead"} Mar 07 15:16:50 crc kubenswrapper[4943]: I0307 15:16:50.514550 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-c7gj4" podStartSLOduration=1.514521123 podStartE2EDuration="1.514521123s" podCreationTimestamp="2026-03-07 15:16:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:16:50.504307103 +0000 UTC m=+2252.456443631" watchObservedRunningTime="2026-03-07 15:16:50.514521123 +0000 UTC m=+2252.466657661" Mar 07 15:16:51 crc kubenswrapper[4943]: I0307 15:16:51.512283 4943 generic.go:334] "Generic (PLEG): container finished" podID="1490d08b-b319-42bc-9259-ff749a56d136" containerID="17364057933ada0f1377538d14281f33b01873033659fcf72f6baf62dda3b21e" exitCode=0 Mar 07 15:16:51 crc kubenswrapper[4943]: I0307 15:16:51.512373 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-c7gj4" event={"ID":"1490d08b-b319-42bc-9259-ff749a56d136","Type":"ContainerDied","Data":"17364057933ada0f1377538d14281f33b01873033659fcf72f6baf62dda3b21e"} Mar 07 15:16:52 crc kubenswrapper[4943]: I0307 15:16:52.892478 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-c7gj4" Mar 07 15:16:52 crc kubenswrapper[4943]: I0307 15:16:52.933770 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-c7gj4"] Mar 07 15:16:52 crc kubenswrapper[4943]: I0307 15:16:52.939078 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-c7gj4"] Mar 07 15:16:53 crc kubenswrapper[4943]: I0307 15:16:53.012353 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6gzf7\" (UniqueName: \"kubernetes.io/projected/1490d08b-b319-42bc-9259-ff749a56d136-kube-api-access-6gzf7\") pod \"1490d08b-b319-42bc-9259-ff749a56d136\" (UID: \"1490d08b-b319-42bc-9259-ff749a56d136\") " Mar 07 15:16:53 crc kubenswrapper[4943]: I0307 15:16:53.012427 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1490d08b-b319-42bc-9259-ff749a56d136-ring-data-devices\") pod \"1490d08b-b319-42bc-9259-ff749a56d136\" (UID: \"1490d08b-b319-42bc-9259-ff749a56d136\") " Mar 07 15:16:53 crc kubenswrapper[4943]: I0307 15:16:53.012484 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1490d08b-b319-42bc-9259-ff749a56d136-swiftconf\") pod \"1490d08b-b319-42bc-9259-ff749a56d136\" (UID: \"1490d08b-b319-42bc-9259-ff749a56d136\") " Mar 07 15:16:53 crc kubenswrapper[4943]: I0307 15:16:53.012511 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1490d08b-b319-42bc-9259-ff749a56d136-scripts\") pod \"1490d08b-b319-42bc-9259-ff749a56d136\" (UID: \"1490d08b-b319-42bc-9259-ff749a56d136\") " Mar 07 15:16:53 crc kubenswrapper[4943]: I0307 15:16:53.012544 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1490d08b-b319-42bc-9259-ff749a56d136-dispersionconf\") pod \"1490d08b-b319-42bc-9259-ff749a56d136\" (UID: \"1490d08b-b319-42bc-9259-ff749a56d136\") " Mar 07 15:16:53 crc kubenswrapper[4943]: I0307 15:16:53.012570 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1490d08b-b319-42bc-9259-ff749a56d136-etc-swift\") pod \"1490d08b-b319-42bc-9259-ff749a56d136\" (UID: \"1490d08b-b319-42bc-9259-ff749a56d136\") " Mar 07 15:16:53 crc kubenswrapper[4943]: I0307 15:16:53.013117 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1490d08b-b319-42bc-9259-ff749a56d136-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "1490d08b-b319-42bc-9259-ff749a56d136" (UID: "1490d08b-b319-42bc-9259-ff749a56d136"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:16:53 crc kubenswrapper[4943]: I0307 15:16:53.013407 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1490d08b-b319-42bc-9259-ff749a56d136-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "1490d08b-b319-42bc-9259-ff749a56d136" (UID: "1490d08b-b319-42bc-9259-ff749a56d136"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:16:53 crc kubenswrapper[4943]: I0307 15:16:53.025336 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1490d08b-b319-42bc-9259-ff749a56d136-kube-api-access-6gzf7" (OuterVolumeSpecName: "kube-api-access-6gzf7") pod "1490d08b-b319-42bc-9259-ff749a56d136" (UID: "1490d08b-b319-42bc-9259-ff749a56d136"). InnerVolumeSpecName "kube-api-access-6gzf7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:16:53 crc kubenswrapper[4943]: I0307 15:16:53.045347 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1490d08b-b319-42bc-9259-ff749a56d136-scripts" (OuterVolumeSpecName: "scripts") pod "1490d08b-b319-42bc-9259-ff749a56d136" (UID: "1490d08b-b319-42bc-9259-ff749a56d136"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:16:53 crc kubenswrapper[4943]: I0307 15:16:53.049604 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1490d08b-b319-42bc-9259-ff749a56d136-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "1490d08b-b319-42bc-9259-ff749a56d136" (UID: "1490d08b-b319-42bc-9259-ff749a56d136"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:16:53 crc kubenswrapper[4943]: I0307 15:16:53.056023 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1490d08b-b319-42bc-9259-ff749a56d136-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "1490d08b-b319-42bc-9259-ff749a56d136" (UID: "1490d08b-b319-42bc-9259-ff749a56d136"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:16:53 crc kubenswrapper[4943]: I0307 15:16:53.113955 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1490d08b-b319-42bc-9259-ff749a56d136-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:53 crc kubenswrapper[4943]: I0307 15:16:53.113985 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1490d08b-b319-42bc-9259-ff749a56d136-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:53 crc kubenswrapper[4943]: I0307 15:16:53.113996 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1490d08b-b319-42bc-9259-ff749a56d136-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:53 crc kubenswrapper[4943]: I0307 15:16:53.114005 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6gzf7\" (UniqueName: \"kubernetes.io/projected/1490d08b-b319-42bc-9259-ff749a56d136-kube-api-access-6gzf7\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:53 crc kubenswrapper[4943]: I0307 15:16:53.114013 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1490d08b-b319-42bc-9259-ff749a56d136-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:53 crc kubenswrapper[4943]: I0307 15:16:53.114022 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1490d08b-b319-42bc-9259-ff749a56d136-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:53 crc kubenswrapper[4943]: I0307 15:16:53.537356 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="891ff0d0a5a322741661597af57264716bd7f25090eb9e90ca6d764c54ec4ead" Mar 07 15:16:53 crc kubenswrapper[4943]: I0307 15:16:53.537413 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-c7gj4" Mar 07 15:16:54 crc kubenswrapper[4943]: I0307 15:16:54.073809 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-cwz52"] Mar 07 15:16:54 crc kubenswrapper[4943]: E0307 15:16:54.074612 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1490d08b-b319-42bc-9259-ff749a56d136" containerName="swift-ring-rebalance" Mar 07 15:16:54 crc kubenswrapper[4943]: I0307 15:16:54.074634 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="1490d08b-b319-42bc-9259-ff749a56d136" containerName="swift-ring-rebalance" Mar 07 15:16:54 crc kubenswrapper[4943]: I0307 15:16:54.074883 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="1490d08b-b319-42bc-9259-ff749a56d136" containerName="swift-ring-rebalance" Mar 07 15:16:54 crc kubenswrapper[4943]: I0307 15:16:54.075726 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-cwz52" Mar 07 15:16:54 crc kubenswrapper[4943]: I0307 15:16:54.078734 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:16:54 crc kubenswrapper[4943]: I0307 15:16:54.078806 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:16:54 crc kubenswrapper[4943]: I0307 15:16:54.086536 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-cwz52"] Mar 07 15:16:54 crc kubenswrapper[4943]: I0307 15:16:54.133016 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mccf\" (UniqueName: \"kubernetes.io/projected/62e4f176-732c-4d0e-a8d9-54895e9e72af-kube-api-access-6mccf\") pod \"swift-ring-rebalance-debug-cwz52\" (UID: \"62e4f176-732c-4d0e-a8d9-54895e9e72af\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-cwz52" Mar 07 15:16:54 crc kubenswrapper[4943]: I0307 15:16:54.133111 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/62e4f176-732c-4d0e-a8d9-54895e9e72af-dispersionconf\") pod \"swift-ring-rebalance-debug-cwz52\" (UID: \"62e4f176-732c-4d0e-a8d9-54895e9e72af\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-cwz52" Mar 07 15:16:54 crc kubenswrapper[4943]: I0307 15:16:54.133151 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/62e4f176-732c-4d0e-a8d9-54895e9e72af-swiftconf\") pod \"swift-ring-rebalance-debug-cwz52\" (UID: \"62e4f176-732c-4d0e-a8d9-54895e9e72af\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-cwz52" Mar 07 15:16:54 crc kubenswrapper[4943]: I0307 15:16:54.133191 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/62e4f176-732c-4d0e-a8d9-54895e9e72af-etc-swift\") pod \"swift-ring-rebalance-debug-cwz52\" (UID: \"62e4f176-732c-4d0e-a8d9-54895e9e72af\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-cwz52" Mar 07 15:16:54 crc kubenswrapper[4943]: I0307 15:16:54.133323 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/62e4f176-732c-4d0e-a8d9-54895e9e72af-ring-data-devices\") pod \"swift-ring-rebalance-debug-cwz52\" (UID: \"62e4f176-732c-4d0e-a8d9-54895e9e72af\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-cwz52" Mar 07 15:16:54 crc kubenswrapper[4943]: I0307 15:16:54.133408 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/62e4f176-732c-4d0e-a8d9-54895e9e72af-scripts\") pod \"swift-ring-rebalance-debug-cwz52\" (UID: \"62e4f176-732c-4d0e-a8d9-54895e9e72af\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-cwz52" Mar 07 15:16:54 crc kubenswrapper[4943]: I0307 15:16:54.234788 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/62e4f176-732c-4d0e-a8d9-54895e9e72af-ring-data-devices\") pod \"swift-ring-rebalance-debug-cwz52\" (UID: \"62e4f176-732c-4d0e-a8d9-54895e9e72af\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-cwz52" Mar 07 15:16:54 crc kubenswrapper[4943]: I0307 15:16:54.234884 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/62e4f176-732c-4d0e-a8d9-54895e9e72af-scripts\") pod \"swift-ring-rebalance-debug-cwz52\" (UID: \"62e4f176-732c-4d0e-a8d9-54895e9e72af\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-cwz52" Mar 07 15:16:54 crc kubenswrapper[4943]: I0307 15:16:54.235009 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mccf\" (UniqueName: \"kubernetes.io/projected/62e4f176-732c-4d0e-a8d9-54895e9e72af-kube-api-access-6mccf\") pod \"swift-ring-rebalance-debug-cwz52\" (UID: \"62e4f176-732c-4d0e-a8d9-54895e9e72af\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-cwz52" Mar 07 15:16:54 crc kubenswrapper[4943]: I0307 15:16:54.235073 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/62e4f176-732c-4d0e-a8d9-54895e9e72af-dispersionconf\") pod \"swift-ring-rebalance-debug-cwz52\" (UID: \"62e4f176-732c-4d0e-a8d9-54895e9e72af\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-cwz52" Mar 07 15:16:54 crc kubenswrapper[4943]: I0307 15:16:54.235113 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/62e4f176-732c-4d0e-a8d9-54895e9e72af-swiftconf\") pod \"swift-ring-rebalance-debug-cwz52\" (UID: \"62e4f176-732c-4d0e-a8d9-54895e9e72af\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-cwz52" Mar 07 15:16:54 crc kubenswrapper[4943]: I0307 15:16:54.235192 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/62e4f176-732c-4d0e-a8d9-54895e9e72af-etc-swift\") pod \"swift-ring-rebalance-debug-cwz52\" (UID: \"62e4f176-732c-4d0e-a8d9-54895e9e72af\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-cwz52" Mar 07 15:16:54 crc kubenswrapper[4943]: I0307 15:16:54.235859 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/62e4f176-732c-4d0e-a8d9-54895e9e72af-etc-swift\") pod \"swift-ring-rebalance-debug-cwz52\" (UID: \"62e4f176-732c-4d0e-a8d9-54895e9e72af\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-cwz52" Mar 07 15:16:54 crc kubenswrapper[4943]: I0307 15:16:54.236188 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/62e4f176-732c-4d0e-a8d9-54895e9e72af-scripts\") pod \"swift-ring-rebalance-debug-cwz52\" (UID: \"62e4f176-732c-4d0e-a8d9-54895e9e72af\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-cwz52" Mar 07 15:16:54 crc kubenswrapper[4943]: I0307 15:16:54.237178 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/62e4f176-732c-4d0e-a8d9-54895e9e72af-ring-data-devices\") pod \"swift-ring-rebalance-debug-cwz52\" (UID: \"62e4f176-732c-4d0e-a8d9-54895e9e72af\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-cwz52" Mar 07 15:16:54 crc kubenswrapper[4943]: I0307 15:16:54.241221 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/62e4f176-732c-4d0e-a8d9-54895e9e72af-dispersionconf\") pod \"swift-ring-rebalance-debug-cwz52\" (UID: \"62e4f176-732c-4d0e-a8d9-54895e9e72af\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-cwz52" Mar 07 15:16:54 crc kubenswrapper[4943]: I0307 15:16:54.241811 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/62e4f176-732c-4d0e-a8d9-54895e9e72af-swiftconf\") pod \"swift-ring-rebalance-debug-cwz52\" (UID: \"62e4f176-732c-4d0e-a8d9-54895e9e72af\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-cwz52" Mar 07 15:16:54 crc kubenswrapper[4943]: I0307 15:16:54.261993 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mccf\" (UniqueName: \"kubernetes.io/projected/62e4f176-732c-4d0e-a8d9-54895e9e72af-kube-api-access-6mccf\") pod \"swift-ring-rebalance-debug-cwz52\" (UID: \"62e4f176-732c-4d0e-a8d9-54895e9e72af\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-cwz52" Mar 07 15:16:54 crc kubenswrapper[4943]: I0307 15:16:54.445080 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-cwz52" Mar 07 15:16:54 crc kubenswrapper[4943]: I0307 15:16:54.770488 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1490d08b-b319-42bc-9259-ff749a56d136" path="/var/lib/kubelet/pods/1490d08b-b319-42bc-9259-ff749a56d136/volumes" Mar 07 15:16:54 crc kubenswrapper[4943]: I0307 15:16:54.955648 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-cwz52"] Mar 07 15:16:54 crc kubenswrapper[4943]: W0307 15:16:54.964646 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod62e4f176_732c_4d0e_a8d9_54895e9e72af.slice/crio-1e553535fe48a6a614b9e4c4108fa3ee2ad551092e0dbaecfd8206430a81d8ea WatchSource:0}: Error finding container 1e553535fe48a6a614b9e4c4108fa3ee2ad551092e0dbaecfd8206430a81d8ea: Status 404 returned error can't find the container with id 1e553535fe48a6a614b9e4c4108fa3ee2ad551092e0dbaecfd8206430a81d8ea Mar 07 15:16:55 crc kubenswrapper[4943]: I0307 15:16:55.566883 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-cwz52" event={"ID":"62e4f176-732c-4d0e-a8d9-54895e9e72af","Type":"ContainerStarted","Data":"61cd6cf3937822a98ccfe8d816ceac8893280a0ec691733b1fb03f468d36b867"} Mar 07 15:16:55 crc kubenswrapper[4943]: I0307 15:16:55.566946 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-cwz52" event={"ID":"62e4f176-732c-4d0e-a8d9-54895e9e72af","Type":"ContainerStarted","Data":"1e553535fe48a6a614b9e4c4108fa3ee2ad551092e0dbaecfd8206430a81d8ea"} Mar 07 15:16:56 crc kubenswrapper[4943]: I0307 15:16:56.579574 4943 generic.go:334] "Generic (PLEG): container finished" podID="62e4f176-732c-4d0e-a8d9-54895e9e72af" containerID="61cd6cf3937822a98ccfe8d816ceac8893280a0ec691733b1fb03f468d36b867" exitCode=0 Mar 07 15:16:56 crc kubenswrapper[4943]: I0307 15:16:56.579635 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-cwz52" event={"ID":"62e4f176-732c-4d0e-a8d9-54895e9e72af","Type":"ContainerDied","Data":"61cd6cf3937822a98ccfe8d816ceac8893280a0ec691733b1fb03f468d36b867"} Mar 07 15:16:58 crc kubenswrapper[4943]: I0307 15:16:58.047837 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-cwz52" Mar 07 15:16:58 crc kubenswrapper[4943]: I0307 15:16:58.087594 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-cwz52"] Mar 07 15:16:58 crc kubenswrapper[4943]: I0307 15:16:58.095647 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-cwz52"] Mar 07 15:16:58 crc kubenswrapper[4943]: I0307 15:16:58.109679 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/62e4f176-732c-4d0e-a8d9-54895e9e72af-dispersionconf\") pod \"62e4f176-732c-4d0e-a8d9-54895e9e72af\" (UID: \"62e4f176-732c-4d0e-a8d9-54895e9e72af\") " Mar 07 15:16:58 crc kubenswrapper[4943]: I0307 15:16:58.109782 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/62e4f176-732c-4d0e-a8d9-54895e9e72af-etc-swift\") pod \"62e4f176-732c-4d0e-a8d9-54895e9e72af\" (UID: \"62e4f176-732c-4d0e-a8d9-54895e9e72af\") " Mar 07 15:16:58 crc kubenswrapper[4943]: I0307 15:16:58.109843 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/62e4f176-732c-4d0e-a8d9-54895e9e72af-swiftconf\") pod \"62e4f176-732c-4d0e-a8d9-54895e9e72af\" (UID: \"62e4f176-732c-4d0e-a8d9-54895e9e72af\") " Mar 07 15:16:58 crc kubenswrapper[4943]: I0307 15:16:58.109913 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6mccf\" (UniqueName: \"kubernetes.io/projected/62e4f176-732c-4d0e-a8d9-54895e9e72af-kube-api-access-6mccf\") pod \"62e4f176-732c-4d0e-a8d9-54895e9e72af\" (UID: \"62e4f176-732c-4d0e-a8d9-54895e9e72af\") " Mar 07 15:16:58 crc kubenswrapper[4943]: I0307 15:16:58.109989 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/62e4f176-732c-4d0e-a8d9-54895e9e72af-ring-data-devices\") pod \"62e4f176-732c-4d0e-a8d9-54895e9e72af\" (UID: \"62e4f176-732c-4d0e-a8d9-54895e9e72af\") " Mar 07 15:16:58 crc kubenswrapper[4943]: I0307 15:16:58.110051 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/62e4f176-732c-4d0e-a8d9-54895e9e72af-scripts\") pod \"62e4f176-732c-4d0e-a8d9-54895e9e72af\" (UID: \"62e4f176-732c-4d0e-a8d9-54895e9e72af\") " Mar 07 15:16:58 crc kubenswrapper[4943]: I0307 15:16:58.111676 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62e4f176-732c-4d0e-a8d9-54895e9e72af-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "62e4f176-732c-4d0e-a8d9-54895e9e72af" (UID: "62e4f176-732c-4d0e-a8d9-54895e9e72af"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:16:58 crc kubenswrapper[4943]: I0307 15:16:58.111739 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62e4f176-732c-4d0e-a8d9-54895e9e72af-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "62e4f176-732c-4d0e-a8d9-54895e9e72af" (UID: "62e4f176-732c-4d0e-a8d9-54895e9e72af"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:16:58 crc kubenswrapper[4943]: I0307 15:16:58.139141 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62e4f176-732c-4d0e-a8d9-54895e9e72af-kube-api-access-6mccf" (OuterVolumeSpecName: "kube-api-access-6mccf") pod "62e4f176-732c-4d0e-a8d9-54895e9e72af" (UID: "62e4f176-732c-4d0e-a8d9-54895e9e72af"). InnerVolumeSpecName "kube-api-access-6mccf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:16:58 crc kubenswrapper[4943]: I0307 15:16:58.143170 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62e4f176-732c-4d0e-a8d9-54895e9e72af-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "62e4f176-732c-4d0e-a8d9-54895e9e72af" (UID: "62e4f176-732c-4d0e-a8d9-54895e9e72af"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:16:58 crc kubenswrapper[4943]: I0307 15:16:58.146384 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62e4f176-732c-4d0e-a8d9-54895e9e72af-scripts" (OuterVolumeSpecName: "scripts") pod "62e4f176-732c-4d0e-a8d9-54895e9e72af" (UID: "62e4f176-732c-4d0e-a8d9-54895e9e72af"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:16:58 crc kubenswrapper[4943]: I0307 15:16:58.150854 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62e4f176-732c-4d0e-a8d9-54895e9e72af-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "62e4f176-732c-4d0e-a8d9-54895e9e72af" (UID: "62e4f176-732c-4d0e-a8d9-54895e9e72af"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:16:58 crc kubenswrapper[4943]: I0307 15:16:58.212454 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/62e4f176-732c-4d0e-a8d9-54895e9e72af-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:58 crc kubenswrapper[4943]: I0307 15:16:58.212509 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/62e4f176-732c-4d0e-a8d9-54895e9e72af-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:58 crc kubenswrapper[4943]: I0307 15:16:58.212522 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/62e4f176-732c-4d0e-a8d9-54895e9e72af-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:58 crc kubenswrapper[4943]: I0307 15:16:58.212535 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6mccf\" (UniqueName: \"kubernetes.io/projected/62e4f176-732c-4d0e-a8d9-54895e9e72af-kube-api-access-6mccf\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:58 crc kubenswrapper[4943]: I0307 15:16:58.212550 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/62e4f176-732c-4d0e-a8d9-54895e9e72af-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:58 crc kubenswrapper[4943]: I0307 15:16:58.212562 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/62e4f176-732c-4d0e-a8d9-54895e9e72af-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:16:58 crc kubenswrapper[4943]: I0307 15:16:58.620229 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1e553535fe48a6a614b9e4c4108fa3ee2ad551092e0dbaecfd8206430a81d8ea" Mar 07 15:16:58 crc kubenswrapper[4943]: I0307 15:16:58.620301 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-cwz52" Mar 07 15:16:58 crc kubenswrapper[4943]: I0307 15:16:58.769879 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62e4f176-732c-4d0e-a8d9-54895e9e72af" path="/var/lib/kubelet/pods/62e4f176-732c-4d0e-a8d9-54895e9e72af/volumes" Mar 07 15:16:59 crc kubenswrapper[4943]: I0307 15:16:59.298664 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-mbkd8"] Mar 07 15:16:59 crc kubenswrapper[4943]: E0307 15:16:59.299158 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62e4f176-732c-4d0e-a8d9-54895e9e72af" containerName="swift-ring-rebalance" Mar 07 15:16:59 crc kubenswrapper[4943]: I0307 15:16:59.299180 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="62e4f176-732c-4d0e-a8d9-54895e9e72af" containerName="swift-ring-rebalance" Mar 07 15:16:59 crc kubenswrapper[4943]: I0307 15:16:59.299461 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="62e4f176-732c-4d0e-a8d9-54895e9e72af" containerName="swift-ring-rebalance" Mar 07 15:16:59 crc kubenswrapper[4943]: I0307 15:16:59.300214 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-mbkd8" Mar 07 15:16:59 crc kubenswrapper[4943]: I0307 15:16:59.303524 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:16:59 crc kubenswrapper[4943]: I0307 15:16:59.303857 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:16:59 crc kubenswrapper[4943]: I0307 15:16:59.324244 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-mbkd8"] Mar 07 15:16:59 crc kubenswrapper[4943]: I0307 15:16:59.436236 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/eaa02776-103b-4290-ac42-dff319f86dd5-ring-data-devices\") pod \"swift-ring-rebalance-debug-mbkd8\" (UID: \"eaa02776-103b-4290-ac42-dff319f86dd5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-mbkd8" Mar 07 15:16:59 crc kubenswrapper[4943]: I0307 15:16:59.436338 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/eaa02776-103b-4290-ac42-dff319f86dd5-swiftconf\") pod \"swift-ring-rebalance-debug-mbkd8\" (UID: \"eaa02776-103b-4290-ac42-dff319f86dd5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-mbkd8" Mar 07 15:16:59 crc kubenswrapper[4943]: I0307 15:16:59.436404 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gw486\" (UniqueName: \"kubernetes.io/projected/eaa02776-103b-4290-ac42-dff319f86dd5-kube-api-access-gw486\") pod \"swift-ring-rebalance-debug-mbkd8\" (UID: \"eaa02776-103b-4290-ac42-dff319f86dd5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-mbkd8" Mar 07 15:16:59 crc kubenswrapper[4943]: I0307 15:16:59.436516 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eaa02776-103b-4290-ac42-dff319f86dd5-scripts\") pod \"swift-ring-rebalance-debug-mbkd8\" (UID: \"eaa02776-103b-4290-ac42-dff319f86dd5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-mbkd8" Mar 07 15:16:59 crc kubenswrapper[4943]: I0307 15:16:59.436546 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/eaa02776-103b-4290-ac42-dff319f86dd5-dispersionconf\") pod \"swift-ring-rebalance-debug-mbkd8\" (UID: \"eaa02776-103b-4290-ac42-dff319f86dd5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-mbkd8" Mar 07 15:16:59 crc kubenswrapper[4943]: I0307 15:16:59.436582 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/eaa02776-103b-4290-ac42-dff319f86dd5-etc-swift\") pod \"swift-ring-rebalance-debug-mbkd8\" (UID: \"eaa02776-103b-4290-ac42-dff319f86dd5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-mbkd8" Mar 07 15:16:59 crc kubenswrapper[4943]: I0307 15:16:59.538366 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/eaa02776-103b-4290-ac42-dff319f86dd5-ring-data-devices\") pod \"swift-ring-rebalance-debug-mbkd8\" (UID: \"eaa02776-103b-4290-ac42-dff319f86dd5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-mbkd8" Mar 07 15:16:59 crc kubenswrapper[4943]: I0307 15:16:59.538466 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/eaa02776-103b-4290-ac42-dff319f86dd5-swiftconf\") pod \"swift-ring-rebalance-debug-mbkd8\" (UID: \"eaa02776-103b-4290-ac42-dff319f86dd5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-mbkd8" Mar 07 15:16:59 crc kubenswrapper[4943]: I0307 15:16:59.538534 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gw486\" (UniqueName: \"kubernetes.io/projected/eaa02776-103b-4290-ac42-dff319f86dd5-kube-api-access-gw486\") pod \"swift-ring-rebalance-debug-mbkd8\" (UID: \"eaa02776-103b-4290-ac42-dff319f86dd5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-mbkd8" Mar 07 15:16:59 crc kubenswrapper[4943]: I0307 15:16:59.538643 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eaa02776-103b-4290-ac42-dff319f86dd5-scripts\") pod \"swift-ring-rebalance-debug-mbkd8\" (UID: \"eaa02776-103b-4290-ac42-dff319f86dd5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-mbkd8" Mar 07 15:16:59 crc kubenswrapper[4943]: I0307 15:16:59.538677 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/eaa02776-103b-4290-ac42-dff319f86dd5-dispersionconf\") pod \"swift-ring-rebalance-debug-mbkd8\" (UID: \"eaa02776-103b-4290-ac42-dff319f86dd5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-mbkd8" Mar 07 15:16:59 crc kubenswrapper[4943]: I0307 15:16:59.538717 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/eaa02776-103b-4290-ac42-dff319f86dd5-etc-swift\") pod \"swift-ring-rebalance-debug-mbkd8\" (UID: \"eaa02776-103b-4290-ac42-dff319f86dd5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-mbkd8" Mar 07 15:16:59 crc kubenswrapper[4943]: I0307 15:16:59.539530 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/eaa02776-103b-4290-ac42-dff319f86dd5-etc-swift\") pod \"swift-ring-rebalance-debug-mbkd8\" (UID: \"eaa02776-103b-4290-ac42-dff319f86dd5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-mbkd8" Mar 07 15:16:59 crc kubenswrapper[4943]: I0307 15:16:59.540008 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/eaa02776-103b-4290-ac42-dff319f86dd5-ring-data-devices\") pod \"swift-ring-rebalance-debug-mbkd8\" (UID: \"eaa02776-103b-4290-ac42-dff319f86dd5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-mbkd8" Mar 07 15:16:59 crc kubenswrapper[4943]: I0307 15:16:59.540168 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eaa02776-103b-4290-ac42-dff319f86dd5-scripts\") pod \"swift-ring-rebalance-debug-mbkd8\" (UID: \"eaa02776-103b-4290-ac42-dff319f86dd5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-mbkd8" Mar 07 15:16:59 crc kubenswrapper[4943]: I0307 15:16:59.544696 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/eaa02776-103b-4290-ac42-dff319f86dd5-dispersionconf\") pod \"swift-ring-rebalance-debug-mbkd8\" (UID: \"eaa02776-103b-4290-ac42-dff319f86dd5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-mbkd8" Mar 07 15:16:59 crc kubenswrapper[4943]: I0307 15:16:59.544715 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/eaa02776-103b-4290-ac42-dff319f86dd5-swiftconf\") pod \"swift-ring-rebalance-debug-mbkd8\" (UID: \"eaa02776-103b-4290-ac42-dff319f86dd5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-mbkd8" Mar 07 15:16:59 crc kubenswrapper[4943]: I0307 15:16:59.560676 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gw486\" (UniqueName: \"kubernetes.io/projected/eaa02776-103b-4290-ac42-dff319f86dd5-kube-api-access-gw486\") pod \"swift-ring-rebalance-debug-mbkd8\" (UID: \"eaa02776-103b-4290-ac42-dff319f86dd5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-mbkd8" Mar 07 15:16:59 crc kubenswrapper[4943]: I0307 15:16:59.624770 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-mbkd8" Mar 07 15:17:00 crc kubenswrapper[4943]: I0307 15:17:00.220870 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-mbkd8"] Mar 07 15:17:00 crc kubenswrapper[4943]: I0307 15:17:00.648798 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-mbkd8" event={"ID":"eaa02776-103b-4290-ac42-dff319f86dd5","Type":"ContainerStarted","Data":"d460a3d5d4eb2979f3dedb56c7984013529157d7cfdc0f799c927072a626b3bf"} Mar 07 15:17:00 crc kubenswrapper[4943]: I0307 15:17:00.648845 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-mbkd8" event={"ID":"eaa02776-103b-4290-ac42-dff319f86dd5","Type":"ContainerStarted","Data":"8b9b81cace84605b6d3fe82308206e41feaa0f18ee1a8e056348fb13ab097561"} Mar 07 15:17:00 crc kubenswrapper[4943]: I0307 15:17:00.675990 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-mbkd8" podStartSLOduration=1.6759711560000001 podStartE2EDuration="1.675971156s" podCreationTimestamp="2026-03-07 15:16:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:17:00.670339398 +0000 UTC m=+2262.622475956" watchObservedRunningTime="2026-03-07 15:17:00.675971156 +0000 UTC m=+2262.628107644" Mar 07 15:17:01 crc kubenswrapper[4943]: I0307 15:17:01.830440 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rm5zj"] Mar 07 15:17:01 crc kubenswrapper[4943]: I0307 15:17:01.834292 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rm5zj" Mar 07 15:17:01 crc kubenswrapper[4943]: I0307 15:17:01.851246 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rm5zj"] Mar 07 15:17:02 crc kubenswrapper[4943]: I0307 15:17:02.011508 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c107b9f-c2c1-4261-8c88-bf93b6752bca-catalog-content\") pod \"redhat-operators-rm5zj\" (UID: \"6c107b9f-c2c1-4261-8c88-bf93b6752bca\") " pod="openshift-marketplace/redhat-operators-rm5zj" Mar 07 15:17:02 crc kubenswrapper[4943]: I0307 15:17:02.011574 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c107b9f-c2c1-4261-8c88-bf93b6752bca-utilities\") pod \"redhat-operators-rm5zj\" (UID: \"6c107b9f-c2c1-4261-8c88-bf93b6752bca\") " pod="openshift-marketplace/redhat-operators-rm5zj" Mar 07 15:17:02 crc kubenswrapper[4943]: I0307 15:17:02.011636 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vtbp\" (UniqueName: \"kubernetes.io/projected/6c107b9f-c2c1-4261-8c88-bf93b6752bca-kube-api-access-7vtbp\") pod \"redhat-operators-rm5zj\" (UID: \"6c107b9f-c2c1-4261-8c88-bf93b6752bca\") " pod="openshift-marketplace/redhat-operators-rm5zj" Mar 07 15:17:02 crc kubenswrapper[4943]: I0307 15:17:02.113163 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c107b9f-c2c1-4261-8c88-bf93b6752bca-catalog-content\") pod \"redhat-operators-rm5zj\" (UID: \"6c107b9f-c2c1-4261-8c88-bf93b6752bca\") " pod="openshift-marketplace/redhat-operators-rm5zj" Mar 07 15:17:02 crc kubenswrapper[4943]: I0307 15:17:02.113557 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c107b9f-c2c1-4261-8c88-bf93b6752bca-utilities\") pod \"redhat-operators-rm5zj\" (UID: \"6c107b9f-c2c1-4261-8c88-bf93b6752bca\") " pod="openshift-marketplace/redhat-operators-rm5zj" Mar 07 15:17:02 crc kubenswrapper[4943]: I0307 15:17:02.113642 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vtbp\" (UniqueName: \"kubernetes.io/projected/6c107b9f-c2c1-4261-8c88-bf93b6752bca-kube-api-access-7vtbp\") pod \"redhat-operators-rm5zj\" (UID: \"6c107b9f-c2c1-4261-8c88-bf93b6752bca\") " pod="openshift-marketplace/redhat-operators-rm5zj" Mar 07 15:17:02 crc kubenswrapper[4943]: I0307 15:17:02.113672 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c107b9f-c2c1-4261-8c88-bf93b6752bca-catalog-content\") pod \"redhat-operators-rm5zj\" (UID: \"6c107b9f-c2c1-4261-8c88-bf93b6752bca\") " pod="openshift-marketplace/redhat-operators-rm5zj" Mar 07 15:17:02 crc kubenswrapper[4943]: I0307 15:17:02.114136 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c107b9f-c2c1-4261-8c88-bf93b6752bca-utilities\") pod \"redhat-operators-rm5zj\" (UID: \"6c107b9f-c2c1-4261-8c88-bf93b6752bca\") " pod="openshift-marketplace/redhat-operators-rm5zj" Mar 07 15:17:02 crc kubenswrapper[4943]: I0307 15:17:02.136862 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vtbp\" (UniqueName: \"kubernetes.io/projected/6c107b9f-c2c1-4261-8c88-bf93b6752bca-kube-api-access-7vtbp\") pod \"redhat-operators-rm5zj\" (UID: \"6c107b9f-c2c1-4261-8c88-bf93b6752bca\") " pod="openshift-marketplace/redhat-operators-rm5zj" Mar 07 15:17:02 crc kubenswrapper[4943]: I0307 15:17:02.213299 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rm5zj" Mar 07 15:17:02 crc kubenswrapper[4943]: I0307 15:17:02.633999 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rm5zj"] Mar 07 15:17:02 crc kubenswrapper[4943]: W0307 15:17:02.638826 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6c107b9f_c2c1_4261_8c88_bf93b6752bca.slice/crio-118be207ee89b2d2d5996edd79d245faa29d050a7388709f6febbcdfd419c84c WatchSource:0}: Error finding container 118be207ee89b2d2d5996edd79d245faa29d050a7388709f6febbcdfd419c84c: Status 404 returned error can't find the container with id 118be207ee89b2d2d5996edd79d245faa29d050a7388709f6febbcdfd419c84c Mar 07 15:17:02 crc kubenswrapper[4943]: I0307 15:17:02.667230 4943 generic.go:334] "Generic (PLEG): container finished" podID="eaa02776-103b-4290-ac42-dff319f86dd5" containerID="d460a3d5d4eb2979f3dedb56c7984013529157d7cfdc0f799c927072a626b3bf" exitCode=0 Mar 07 15:17:02 crc kubenswrapper[4943]: I0307 15:17:02.667294 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-mbkd8" event={"ID":"eaa02776-103b-4290-ac42-dff319f86dd5","Type":"ContainerDied","Data":"d460a3d5d4eb2979f3dedb56c7984013529157d7cfdc0f799c927072a626b3bf"} Mar 07 15:17:02 crc kubenswrapper[4943]: I0307 15:17:02.668838 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rm5zj" event={"ID":"6c107b9f-c2c1-4261-8c88-bf93b6752bca","Type":"ContainerStarted","Data":"118be207ee89b2d2d5996edd79d245faa29d050a7388709f6febbcdfd419c84c"} Mar 07 15:17:03 crc kubenswrapper[4943]: I0307 15:17:03.686264 4943 generic.go:334] "Generic (PLEG): container finished" podID="6c107b9f-c2c1-4261-8c88-bf93b6752bca" containerID="f5e05905a886de492549a325357cf76509a37d26a6e1592fa227bfe023e6bc01" exitCode=0 Mar 07 15:17:03 crc kubenswrapper[4943]: I0307 15:17:03.686366 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rm5zj" event={"ID":"6c107b9f-c2c1-4261-8c88-bf93b6752bca","Type":"ContainerDied","Data":"f5e05905a886de492549a325357cf76509a37d26a6e1592fa227bfe023e6bc01"} Mar 07 15:17:03 crc kubenswrapper[4943]: I0307 15:17:03.690780 4943 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 07 15:17:03 crc kubenswrapper[4943]: I0307 15:17:03.756363 4943 scope.go:117] "RemoveContainer" containerID="448308ad6e34ac922d66b3c90e66d0927407a94dd28633ec350502d69e4ca8aa" Mar 07 15:17:03 crc kubenswrapper[4943]: E0307 15:17:03.756564 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cpdmm_openshift-machine-config-operator(4c1f18ea-e557-41fc-95dd-ce4e1722a61e)\"" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" Mar 07 15:17:04 crc kubenswrapper[4943]: I0307 15:17:04.029000 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-mbkd8" Mar 07 15:17:04 crc kubenswrapper[4943]: I0307 15:17:04.063760 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-mbkd8"] Mar 07 15:17:04 crc kubenswrapper[4943]: I0307 15:17:04.070491 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-mbkd8"] Mar 07 15:17:04 crc kubenswrapper[4943]: I0307 15:17:04.158012 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eaa02776-103b-4290-ac42-dff319f86dd5-scripts\") pod \"eaa02776-103b-4290-ac42-dff319f86dd5\" (UID: \"eaa02776-103b-4290-ac42-dff319f86dd5\") " Mar 07 15:17:04 crc kubenswrapper[4943]: I0307 15:17:04.158171 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/eaa02776-103b-4290-ac42-dff319f86dd5-ring-data-devices\") pod \"eaa02776-103b-4290-ac42-dff319f86dd5\" (UID: \"eaa02776-103b-4290-ac42-dff319f86dd5\") " Mar 07 15:17:04 crc kubenswrapper[4943]: I0307 15:17:04.158239 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gw486\" (UniqueName: \"kubernetes.io/projected/eaa02776-103b-4290-ac42-dff319f86dd5-kube-api-access-gw486\") pod \"eaa02776-103b-4290-ac42-dff319f86dd5\" (UID: \"eaa02776-103b-4290-ac42-dff319f86dd5\") " Mar 07 15:17:04 crc kubenswrapper[4943]: I0307 15:17:04.158272 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/eaa02776-103b-4290-ac42-dff319f86dd5-dispersionconf\") pod \"eaa02776-103b-4290-ac42-dff319f86dd5\" (UID: \"eaa02776-103b-4290-ac42-dff319f86dd5\") " Mar 07 15:17:04 crc kubenswrapper[4943]: I0307 15:17:04.158328 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/eaa02776-103b-4290-ac42-dff319f86dd5-etc-swift\") pod \"eaa02776-103b-4290-ac42-dff319f86dd5\" (UID: \"eaa02776-103b-4290-ac42-dff319f86dd5\") " Mar 07 15:17:04 crc kubenswrapper[4943]: I0307 15:17:04.158371 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/eaa02776-103b-4290-ac42-dff319f86dd5-swiftconf\") pod \"eaa02776-103b-4290-ac42-dff319f86dd5\" (UID: \"eaa02776-103b-4290-ac42-dff319f86dd5\") " Mar 07 15:17:04 crc kubenswrapper[4943]: I0307 15:17:04.158918 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eaa02776-103b-4290-ac42-dff319f86dd5-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "eaa02776-103b-4290-ac42-dff319f86dd5" (UID: "eaa02776-103b-4290-ac42-dff319f86dd5"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:17:04 crc kubenswrapper[4943]: I0307 15:17:04.159701 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eaa02776-103b-4290-ac42-dff319f86dd5-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "eaa02776-103b-4290-ac42-dff319f86dd5" (UID: "eaa02776-103b-4290-ac42-dff319f86dd5"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:17:04 crc kubenswrapper[4943]: I0307 15:17:04.166997 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eaa02776-103b-4290-ac42-dff319f86dd5-kube-api-access-gw486" (OuterVolumeSpecName: "kube-api-access-gw486") pod "eaa02776-103b-4290-ac42-dff319f86dd5" (UID: "eaa02776-103b-4290-ac42-dff319f86dd5"). InnerVolumeSpecName "kube-api-access-gw486". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:17:04 crc kubenswrapper[4943]: I0307 15:17:04.183158 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eaa02776-103b-4290-ac42-dff319f86dd5-scripts" (OuterVolumeSpecName: "scripts") pod "eaa02776-103b-4290-ac42-dff319f86dd5" (UID: "eaa02776-103b-4290-ac42-dff319f86dd5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:17:04 crc kubenswrapper[4943]: I0307 15:17:04.187098 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eaa02776-103b-4290-ac42-dff319f86dd5-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "eaa02776-103b-4290-ac42-dff319f86dd5" (UID: "eaa02776-103b-4290-ac42-dff319f86dd5"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:17:04 crc kubenswrapper[4943]: I0307 15:17:04.189324 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eaa02776-103b-4290-ac42-dff319f86dd5-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "eaa02776-103b-4290-ac42-dff319f86dd5" (UID: "eaa02776-103b-4290-ac42-dff319f86dd5"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:17:04 crc kubenswrapper[4943]: I0307 15:17:04.260213 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/eaa02776-103b-4290-ac42-dff319f86dd5-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:04 crc kubenswrapper[4943]: I0307 15:17:04.260241 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gw486\" (UniqueName: \"kubernetes.io/projected/eaa02776-103b-4290-ac42-dff319f86dd5-kube-api-access-gw486\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:04 crc kubenswrapper[4943]: I0307 15:17:04.260254 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/eaa02776-103b-4290-ac42-dff319f86dd5-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:04 crc kubenswrapper[4943]: I0307 15:17:04.260262 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/eaa02776-103b-4290-ac42-dff319f86dd5-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:04 crc kubenswrapper[4943]: I0307 15:17:04.260271 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/eaa02776-103b-4290-ac42-dff319f86dd5-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:04 crc kubenswrapper[4943]: I0307 15:17:04.260279 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eaa02776-103b-4290-ac42-dff319f86dd5-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:04 crc kubenswrapper[4943]: I0307 15:17:04.705069 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8b9b81cace84605b6d3fe82308206e41feaa0f18ee1a8e056348fb13ab097561" Mar 07 15:17:04 crc kubenswrapper[4943]: I0307 15:17:04.706273 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-mbkd8" Mar 07 15:17:04 crc kubenswrapper[4943]: I0307 15:17:04.770447 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eaa02776-103b-4290-ac42-dff319f86dd5" path="/var/lib/kubelet/pods/eaa02776-103b-4290-ac42-dff319f86dd5/volumes" Mar 07 15:17:05 crc kubenswrapper[4943]: I0307 15:17:05.264424 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-sbxxd"] Mar 07 15:17:05 crc kubenswrapper[4943]: E0307 15:17:05.265149 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaa02776-103b-4290-ac42-dff319f86dd5" containerName="swift-ring-rebalance" Mar 07 15:17:05 crc kubenswrapper[4943]: I0307 15:17:05.265168 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaa02776-103b-4290-ac42-dff319f86dd5" containerName="swift-ring-rebalance" Mar 07 15:17:05 crc kubenswrapper[4943]: I0307 15:17:05.265342 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaa02776-103b-4290-ac42-dff319f86dd5" containerName="swift-ring-rebalance" Mar 07 15:17:05 crc kubenswrapper[4943]: I0307 15:17:05.265956 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-sbxxd" Mar 07 15:17:05 crc kubenswrapper[4943]: I0307 15:17:05.268691 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:17:05 crc kubenswrapper[4943]: I0307 15:17:05.268779 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:17:05 crc kubenswrapper[4943]: I0307 15:17:05.284845 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-sbxxd"] Mar 07 15:17:05 crc kubenswrapper[4943]: I0307 15:17:05.375316 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2d91c98f-b635-43c9-a873-21fe0021631c-scripts\") pod \"swift-ring-rebalance-debug-sbxxd\" (UID: \"2d91c98f-b635-43c9-a873-21fe0021631c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sbxxd" Mar 07 15:17:05 crc kubenswrapper[4943]: I0307 15:17:05.375375 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgpzr\" (UniqueName: \"kubernetes.io/projected/2d91c98f-b635-43c9-a873-21fe0021631c-kube-api-access-wgpzr\") pod \"swift-ring-rebalance-debug-sbxxd\" (UID: \"2d91c98f-b635-43c9-a873-21fe0021631c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sbxxd" Mar 07 15:17:05 crc kubenswrapper[4943]: I0307 15:17:05.375415 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2d91c98f-b635-43c9-a873-21fe0021631c-swiftconf\") pod \"swift-ring-rebalance-debug-sbxxd\" (UID: \"2d91c98f-b635-43c9-a873-21fe0021631c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sbxxd" Mar 07 15:17:05 crc kubenswrapper[4943]: I0307 15:17:05.375484 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2d91c98f-b635-43c9-a873-21fe0021631c-ring-data-devices\") pod \"swift-ring-rebalance-debug-sbxxd\" (UID: \"2d91c98f-b635-43c9-a873-21fe0021631c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sbxxd" Mar 07 15:17:05 crc kubenswrapper[4943]: I0307 15:17:05.375572 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2d91c98f-b635-43c9-a873-21fe0021631c-dispersionconf\") pod \"swift-ring-rebalance-debug-sbxxd\" (UID: \"2d91c98f-b635-43c9-a873-21fe0021631c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sbxxd" Mar 07 15:17:05 crc kubenswrapper[4943]: I0307 15:17:05.375667 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2d91c98f-b635-43c9-a873-21fe0021631c-etc-swift\") pod \"swift-ring-rebalance-debug-sbxxd\" (UID: \"2d91c98f-b635-43c9-a873-21fe0021631c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sbxxd" Mar 07 15:17:05 crc kubenswrapper[4943]: I0307 15:17:05.477201 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2d91c98f-b635-43c9-a873-21fe0021631c-scripts\") pod \"swift-ring-rebalance-debug-sbxxd\" (UID: \"2d91c98f-b635-43c9-a873-21fe0021631c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sbxxd" Mar 07 15:17:05 crc kubenswrapper[4943]: I0307 15:17:05.477590 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgpzr\" (UniqueName: \"kubernetes.io/projected/2d91c98f-b635-43c9-a873-21fe0021631c-kube-api-access-wgpzr\") pod \"swift-ring-rebalance-debug-sbxxd\" (UID: \"2d91c98f-b635-43c9-a873-21fe0021631c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sbxxd" Mar 07 15:17:05 crc kubenswrapper[4943]: I0307 15:17:05.477678 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2d91c98f-b635-43c9-a873-21fe0021631c-swiftconf\") pod \"swift-ring-rebalance-debug-sbxxd\" (UID: \"2d91c98f-b635-43c9-a873-21fe0021631c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sbxxd" Mar 07 15:17:05 crc kubenswrapper[4943]: I0307 15:17:05.477765 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2d91c98f-b635-43c9-a873-21fe0021631c-ring-data-devices\") pod \"swift-ring-rebalance-debug-sbxxd\" (UID: \"2d91c98f-b635-43c9-a873-21fe0021631c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sbxxd" Mar 07 15:17:05 crc kubenswrapper[4943]: I0307 15:17:05.477850 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2d91c98f-b635-43c9-a873-21fe0021631c-dispersionconf\") pod \"swift-ring-rebalance-debug-sbxxd\" (UID: \"2d91c98f-b635-43c9-a873-21fe0021631c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sbxxd" Mar 07 15:17:05 crc kubenswrapper[4943]: I0307 15:17:05.477927 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2d91c98f-b635-43c9-a873-21fe0021631c-etc-swift\") pod \"swift-ring-rebalance-debug-sbxxd\" (UID: \"2d91c98f-b635-43c9-a873-21fe0021631c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sbxxd" Mar 07 15:17:05 crc kubenswrapper[4943]: I0307 15:17:05.478299 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2d91c98f-b635-43c9-a873-21fe0021631c-scripts\") pod \"swift-ring-rebalance-debug-sbxxd\" (UID: \"2d91c98f-b635-43c9-a873-21fe0021631c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sbxxd" Mar 07 15:17:05 crc kubenswrapper[4943]: I0307 15:17:05.478422 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2d91c98f-b635-43c9-a873-21fe0021631c-etc-swift\") pod \"swift-ring-rebalance-debug-sbxxd\" (UID: \"2d91c98f-b635-43c9-a873-21fe0021631c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sbxxd" Mar 07 15:17:05 crc kubenswrapper[4943]: I0307 15:17:05.478961 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2d91c98f-b635-43c9-a873-21fe0021631c-ring-data-devices\") pod \"swift-ring-rebalance-debug-sbxxd\" (UID: \"2d91c98f-b635-43c9-a873-21fe0021631c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sbxxd" Mar 07 15:17:05 crc kubenswrapper[4943]: I0307 15:17:05.498815 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2d91c98f-b635-43c9-a873-21fe0021631c-dispersionconf\") pod \"swift-ring-rebalance-debug-sbxxd\" (UID: \"2d91c98f-b635-43c9-a873-21fe0021631c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sbxxd" Mar 07 15:17:05 crc kubenswrapper[4943]: I0307 15:17:05.498963 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2d91c98f-b635-43c9-a873-21fe0021631c-swiftconf\") pod \"swift-ring-rebalance-debug-sbxxd\" (UID: \"2d91c98f-b635-43c9-a873-21fe0021631c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sbxxd" Mar 07 15:17:05 crc kubenswrapper[4943]: I0307 15:17:05.504229 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgpzr\" (UniqueName: \"kubernetes.io/projected/2d91c98f-b635-43c9-a873-21fe0021631c-kube-api-access-wgpzr\") pod \"swift-ring-rebalance-debug-sbxxd\" (UID: \"2d91c98f-b635-43c9-a873-21fe0021631c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sbxxd" Mar 07 15:17:05 crc kubenswrapper[4943]: I0307 15:17:05.592694 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-sbxxd" Mar 07 15:17:05 crc kubenswrapper[4943]: I0307 15:17:05.719091 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rm5zj" event={"ID":"6c107b9f-c2c1-4261-8c88-bf93b6752bca","Type":"ContainerStarted","Data":"d1a3ddf6a3d3f6a42c77eeb440e7d8ff56b4d6b3a602e0c5d2ce7bd69cc6fcc9"} Mar 07 15:17:06 crc kubenswrapper[4943]: I0307 15:17:06.080787 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-sbxxd"] Mar 07 15:17:06 crc kubenswrapper[4943]: I0307 15:17:06.734101 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-sbxxd" event={"ID":"2d91c98f-b635-43c9-a873-21fe0021631c","Type":"ContainerStarted","Data":"bb9a7e7e4155450139c98b46ec263f780a53bc2a3533009b58ed766fddd1c6c4"} Mar 07 15:17:06 crc kubenswrapper[4943]: I0307 15:17:06.734637 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-sbxxd" event={"ID":"2d91c98f-b635-43c9-a873-21fe0021631c","Type":"ContainerStarted","Data":"d9d6be96deeb5475c199be9f5d9f4b4b83735f46f04e6f8d0032665d99a71c96"} Mar 07 15:17:06 crc kubenswrapper[4943]: I0307 15:17:06.739318 4943 generic.go:334] "Generic (PLEG): container finished" podID="6c107b9f-c2c1-4261-8c88-bf93b6752bca" containerID="d1a3ddf6a3d3f6a42c77eeb440e7d8ff56b4d6b3a602e0c5d2ce7bd69cc6fcc9" exitCode=0 Mar 07 15:17:06 crc kubenswrapper[4943]: I0307 15:17:06.739851 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rm5zj" event={"ID":"6c107b9f-c2c1-4261-8c88-bf93b6752bca","Type":"ContainerDied","Data":"d1a3ddf6a3d3f6a42c77eeb440e7d8ff56b4d6b3a602e0c5d2ce7bd69cc6fcc9"} Mar 07 15:17:06 crc kubenswrapper[4943]: I0307 15:17:06.764775 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-sbxxd" podStartSLOduration=1.7647511420000002 podStartE2EDuration="1.764751142s" podCreationTimestamp="2026-03-07 15:17:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:17:06.759649258 +0000 UTC m=+2268.711785796" watchObservedRunningTime="2026-03-07 15:17:06.764751142 +0000 UTC m=+2268.716887680" Mar 07 15:17:08 crc kubenswrapper[4943]: I0307 15:17:08.764052 4943 generic.go:334] "Generic (PLEG): container finished" podID="2d91c98f-b635-43c9-a873-21fe0021631c" containerID="bb9a7e7e4155450139c98b46ec263f780a53bc2a3533009b58ed766fddd1c6c4" exitCode=0 Mar 07 15:17:08 crc kubenswrapper[4943]: I0307 15:17:08.773662 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-sbxxd" event={"ID":"2d91c98f-b635-43c9-a873-21fe0021631c","Type":"ContainerDied","Data":"bb9a7e7e4155450139c98b46ec263f780a53bc2a3533009b58ed766fddd1c6c4"} Mar 07 15:17:08 crc kubenswrapper[4943]: I0307 15:17:08.773725 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rm5zj" event={"ID":"6c107b9f-c2c1-4261-8c88-bf93b6752bca","Type":"ContainerStarted","Data":"901f052294e2dd79323ce8e1b852421f9bbf6ec2bc87b028a78e62a35f3a5196"} Mar 07 15:17:08 crc kubenswrapper[4943]: I0307 15:17:08.808474 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rm5zj" podStartSLOduration=4.097507978 podStartE2EDuration="7.808447795s" podCreationTimestamp="2026-03-07 15:17:01 +0000 UTC" firstStartedPulling="2026-03-07 15:17:03.690309719 +0000 UTC m=+2265.642446257" lastFinishedPulling="2026-03-07 15:17:07.401249566 +0000 UTC m=+2269.353386074" observedRunningTime="2026-03-07 15:17:08.799147238 +0000 UTC m=+2270.751283746" watchObservedRunningTime="2026-03-07 15:17:08.808447795 +0000 UTC m=+2270.760584333" Mar 07 15:17:10 crc kubenswrapper[4943]: I0307 15:17:10.174793 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-sbxxd" Mar 07 15:17:10 crc kubenswrapper[4943]: I0307 15:17:10.221014 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-sbxxd"] Mar 07 15:17:10 crc kubenswrapper[4943]: I0307 15:17:10.227996 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-sbxxd"] Mar 07 15:17:10 crc kubenswrapper[4943]: I0307 15:17:10.358876 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2d91c98f-b635-43c9-a873-21fe0021631c-scripts\") pod \"2d91c98f-b635-43c9-a873-21fe0021631c\" (UID: \"2d91c98f-b635-43c9-a873-21fe0021631c\") " Mar 07 15:17:10 crc kubenswrapper[4943]: I0307 15:17:10.358992 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wgpzr\" (UniqueName: \"kubernetes.io/projected/2d91c98f-b635-43c9-a873-21fe0021631c-kube-api-access-wgpzr\") pod \"2d91c98f-b635-43c9-a873-21fe0021631c\" (UID: \"2d91c98f-b635-43c9-a873-21fe0021631c\") " Mar 07 15:17:10 crc kubenswrapper[4943]: I0307 15:17:10.359018 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2d91c98f-b635-43c9-a873-21fe0021631c-swiftconf\") pod \"2d91c98f-b635-43c9-a873-21fe0021631c\" (UID: \"2d91c98f-b635-43c9-a873-21fe0021631c\") " Mar 07 15:17:10 crc kubenswrapper[4943]: I0307 15:17:10.359034 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2d91c98f-b635-43c9-a873-21fe0021631c-ring-data-devices\") pod \"2d91c98f-b635-43c9-a873-21fe0021631c\" (UID: \"2d91c98f-b635-43c9-a873-21fe0021631c\") " Mar 07 15:17:10 crc kubenswrapper[4943]: I0307 15:17:10.359060 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2d91c98f-b635-43c9-a873-21fe0021631c-dispersionconf\") pod \"2d91c98f-b635-43c9-a873-21fe0021631c\" (UID: \"2d91c98f-b635-43c9-a873-21fe0021631c\") " Mar 07 15:17:10 crc kubenswrapper[4943]: I0307 15:17:10.359076 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2d91c98f-b635-43c9-a873-21fe0021631c-etc-swift\") pod \"2d91c98f-b635-43c9-a873-21fe0021631c\" (UID: \"2d91c98f-b635-43c9-a873-21fe0021631c\") " Mar 07 15:17:10 crc kubenswrapper[4943]: I0307 15:17:10.359613 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d91c98f-b635-43c9-a873-21fe0021631c-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "2d91c98f-b635-43c9-a873-21fe0021631c" (UID: "2d91c98f-b635-43c9-a873-21fe0021631c"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:17:10 crc kubenswrapper[4943]: I0307 15:17:10.359839 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2d91c98f-b635-43c9-a873-21fe0021631c-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:10 crc kubenswrapper[4943]: I0307 15:17:10.359912 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d91c98f-b635-43c9-a873-21fe0021631c-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "2d91c98f-b635-43c9-a873-21fe0021631c" (UID: "2d91c98f-b635-43c9-a873-21fe0021631c"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:17:10 crc kubenswrapper[4943]: I0307 15:17:10.364108 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d91c98f-b635-43c9-a873-21fe0021631c-kube-api-access-wgpzr" (OuterVolumeSpecName: "kube-api-access-wgpzr") pod "2d91c98f-b635-43c9-a873-21fe0021631c" (UID: "2d91c98f-b635-43c9-a873-21fe0021631c"). InnerVolumeSpecName "kube-api-access-wgpzr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:17:10 crc kubenswrapper[4943]: I0307 15:17:10.386118 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d91c98f-b635-43c9-a873-21fe0021631c-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "2d91c98f-b635-43c9-a873-21fe0021631c" (UID: "2d91c98f-b635-43c9-a873-21fe0021631c"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:17:10 crc kubenswrapper[4943]: I0307 15:17:10.389743 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d91c98f-b635-43c9-a873-21fe0021631c-scripts" (OuterVolumeSpecName: "scripts") pod "2d91c98f-b635-43c9-a873-21fe0021631c" (UID: "2d91c98f-b635-43c9-a873-21fe0021631c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:17:10 crc kubenswrapper[4943]: I0307 15:17:10.405308 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d91c98f-b635-43c9-a873-21fe0021631c-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "2d91c98f-b635-43c9-a873-21fe0021631c" (UID: "2d91c98f-b635-43c9-a873-21fe0021631c"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:17:10 crc kubenswrapper[4943]: I0307 15:17:10.461216 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2d91c98f-b635-43c9-a873-21fe0021631c-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:10 crc kubenswrapper[4943]: I0307 15:17:10.461264 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2d91c98f-b635-43c9-a873-21fe0021631c-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:10 crc kubenswrapper[4943]: I0307 15:17:10.461284 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2d91c98f-b635-43c9-a873-21fe0021631c-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:10 crc kubenswrapper[4943]: I0307 15:17:10.461302 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wgpzr\" (UniqueName: \"kubernetes.io/projected/2d91c98f-b635-43c9-a873-21fe0021631c-kube-api-access-wgpzr\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:10 crc kubenswrapper[4943]: I0307 15:17:10.461322 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2d91c98f-b635-43c9-a873-21fe0021631c-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:10 crc kubenswrapper[4943]: I0307 15:17:10.770771 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d91c98f-b635-43c9-a873-21fe0021631c" path="/var/lib/kubelet/pods/2d91c98f-b635-43c9-a873-21fe0021631c/volumes" Mar 07 15:17:10 crc kubenswrapper[4943]: I0307 15:17:10.791188 4943 scope.go:117] "RemoveContainer" containerID="bb9a7e7e4155450139c98b46ec263f780a53bc2a3533009b58ed766fddd1c6c4" Mar 07 15:17:10 crc kubenswrapper[4943]: I0307 15:17:10.791319 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-sbxxd" Mar 07 15:17:11 crc kubenswrapper[4943]: I0307 15:17:11.444139 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-s966m"] Mar 07 15:17:11 crc kubenswrapper[4943]: E0307 15:17:11.444900 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d91c98f-b635-43c9-a873-21fe0021631c" containerName="swift-ring-rebalance" Mar 07 15:17:11 crc kubenswrapper[4943]: I0307 15:17:11.444922 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d91c98f-b635-43c9-a873-21fe0021631c" containerName="swift-ring-rebalance" Mar 07 15:17:11 crc kubenswrapper[4943]: I0307 15:17:11.450846 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d91c98f-b635-43c9-a873-21fe0021631c" containerName="swift-ring-rebalance" Mar 07 15:17:11 crc kubenswrapper[4943]: I0307 15:17:11.451629 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-s966m" Mar 07 15:17:11 crc kubenswrapper[4943]: I0307 15:17:11.454118 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:17:11 crc kubenswrapper[4943]: I0307 15:17:11.454648 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:17:11 crc kubenswrapper[4943]: I0307 15:17:11.455712 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-s966m"] Mar 07 15:17:11 crc kubenswrapper[4943]: I0307 15:17:11.576640 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e-ring-data-devices\") pod \"swift-ring-rebalance-debug-s966m\" (UID: \"f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-s966m" Mar 07 15:17:11 crc kubenswrapper[4943]: I0307 15:17:11.577145 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e-swiftconf\") pod \"swift-ring-rebalance-debug-s966m\" (UID: \"f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-s966m" Mar 07 15:17:11 crc kubenswrapper[4943]: I0307 15:17:11.577173 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e-scripts\") pod \"swift-ring-rebalance-debug-s966m\" (UID: \"f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-s966m" Mar 07 15:17:11 crc kubenswrapper[4943]: I0307 15:17:11.577192 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e-etc-swift\") pod \"swift-ring-rebalance-debug-s966m\" (UID: \"f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-s966m" Mar 07 15:17:11 crc kubenswrapper[4943]: I0307 15:17:11.577217 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4cgx\" (UniqueName: \"kubernetes.io/projected/f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e-kube-api-access-b4cgx\") pod \"swift-ring-rebalance-debug-s966m\" (UID: \"f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-s966m" Mar 07 15:17:11 crc kubenswrapper[4943]: I0307 15:17:11.577235 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e-dispersionconf\") pod \"swift-ring-rebalance-debug-s966m\" (UID: \"f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-s966m" Mar 07 15:17:11 crc kubenswrapper[4943]: I0307 15:17:11.678446 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e-ring-data-devices\") pod \"swift-ring-rebalance-debug-s966m\" (UID: \"f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-s966m" Mar 07 15:17:11 crc kubenswrapper[4943]: I0307 15:17:11.678717 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e-swiftconf\") pod \"swift-ring-rebalance-debug-s966m\" (UID: \"f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-s966m" Mar 07 15:17:11 crc kubenswrapper[4943]: I0307 15:17:11.678801 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e-scripts\") pod \"swift-ring-rebalance-debug-s966m\" (UID: \"f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-s966m" Mar 07 15:17:11 crc kubenswrapper[4943]: I0307 15:17:11.678897 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e-etc-swift\") pod \"swift-ring-rebalance-debug-s966m\" (UID: \"f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-s966m" Mar 07 15:17:11 crc kubenswrapper[4943]: I0307 15:17:11.678997 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4cgx\" (UniqueName: \"kubernetes.io/projected/f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e-kube-api-access-b4cgx\") pod \"swift-ring-rebalance-debug-s966m\" (UID: \"f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-s966m" Mar 07 15:17:11 crc kubenswrapper[4943]: I0307 15:17:11.679064 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e-dispersionconf\") pod \"swift-ring-rebalance-debug-s966m\" (UID: \"f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-s966m" Mar 07 15:17:11 crc kubenswrapper[4943]: I0307 15:17:11.679302 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e-etc-swift\") pod \"swift-ring-rebalance-debug-s966m\" (UID: \"f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-s966m" Mar 07 15:17:11 crc kubenswrapper[4943]: I0307 15:17:11.679560 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e-scripts\") pod \"swift-ring-rebalance-debug-s966m\" (UID: \"f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-s966m" Mar 07 15:17:11 crc kubenswrapper[4943]: I0307 15:17:11.680054 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e-ring-data-devices\") pod \"swift-ring-rebalance-debug-s966m\" (UID: \"f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-s966m" Mar 07 15:17:11 crc kubenswrapper[4943]: I0307 15:17:11.687384 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e-swiftconf\") pod \"swift-ring-rebalance-debug-s966m\" (UID: \"f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-s966m" Mar 07 15:17:11 crc kubenswrapper[4943]: I0307 15:17:11.696179 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e-dispersionconf\") pod \"swift-ring-rebalance-debug-s966m\" (UID: \"f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-s966m" Mar 07 15:17:11 crc kubenswrapper[4943]: I0307 15:17:11.697885 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4cgx\" (UniqueName: \"kubernetes.io/projected/f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e-kube-api-access-b4cgx\") pod \"swift-ring-rebalance-debug-s966m\" (UID: \"f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-s966m" Mar 07 15:17:11 crc kubenswrapper[4943]: I0307 15:17:11.771259 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-s966m" Mar 07 15:17:11 crc kubenswrapper[4943]: I0307 15:17:11.998855 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-s966m"] Mar 07 15:17:12 crc kubenswrapper[4943]: W0307 15:17:12.002188 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf3247b97_4cb6_4839_bbcb_13ed8cdbfd8e.slice/crio-96a2fda5679ad42e8805bea507e838efbbe0c107971144ad6fddff3af4589960 WatchSource:0}: Error finding container 96a2fda5679ad42e8805bea507e838efbbe0c107971144ad6fddff3af4589960: Status 404 returned error can't find the container with id 96a2fda5679ad42e8805bea507e838efbbe0c107971144ad6fddff3af4589960 Mar 07 15:17:12 crc kubenswrapper[4943]: I0307 15:17:12.214414 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rm5zj" Mar 07 15:17:12 crc kubenswrapper[4943]: I0307 15:17:12.214495 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rm5zj" Mar 07 15:17:12 crc kubenswrapper[4943]: I0307 15:17:12.817244 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-s966m" event={"ID":"f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e","Type":"ContainerStarted","Data":"5b9e0bcb311794281eb3c56032cd33eacf7b7b4c013bb10cb04e7e7e66100c23"} Mar 07 15:17:12 crc kubenswrapper[4943]: I0307 15:17:12.817306 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-s966m" event={"ID":"f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e","Type":"ContainerStarted","Data":"96a2fda5679ad42e8805bea507e838efbbe0c107971144ad6fddff3af4589960"} Mar 07 15:17:12 crc kubenswrapper[4943]: I0307 15:17:12.847084 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-s966m" podStartSLOduration=1.8470640299999999 podStartE2EDuration="1.84706403s" podCreationTimestamp="2026-03-07 15:17:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:17:12.836397249 +0000 UTC m=+2274.788533757" watchObservedRunningTime="2026-03-07 15:17:12.84706403 +0000 UTC m=+2274.799200548" Mar 07 15:17:13 crc kubenswrapper[4943]: I0307 15:17:13.268000 4943 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rm5zj" podUID="6c107b9f-c2c1-4261-8c88-bf93b6752bca" containerName="registry-server" probeResult="failure" output=< Mar 07 15:17:13 crc kubenswrapper[4943]: timeout: failed to connect service ":50051" within 1s Mar 07 15:17:13 crc kubenswrapper[4943]: > Mar 07 15:17:13 crc kubenswrapper[4943]: I0307 15:17:13.828081 4943 generic.go:334] "Generic (PLEG): container finished" podID="f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e" containerID="5b9e0bcb311794281eb3c56032cd33eacf7b7b4c013bb10cb04e7e7e66100c23" exitCode=0 Mar 07 15:17:13 crc kubenswrapper[4943]: I0307 15:17:13.828147 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-s966m" event={"ID":"f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e","Type":"ContainerDied","Data":"5b9e0bcb311794281eb3c56032cd33eacf7b7b4c013bb10cb04e7e7e66100c23"} Mar 07 15:17:15 crc kubenswrapper[4943]: I0307 15:17:15.212801 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-s966m" Mar 07 15:17:15 crc kubenswrapper[4943]: I0307 15:17:15.246474 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-s966m"] Mar 07 15:17:15 crc kubenswrapper[4943]: I0307 15:17:15.253545 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-s966m"] Mar 07 15:17:15 crc kubenswrapper[4943]: I0307 15:17:15.334349 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e-dispersionconf\") pod \"f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e\" (UID: \"f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e\") " Mar 07 15:17:15 crc kubenswrapper[4943]: I0307 15:17:15.334429 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e-etc-swift\") pod \"f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e\" (UID: \"f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e\") " Mar 07 15:17:15 crc kubenswrapper[4943]: I0307 15:17:15.334477 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b4cgx\" (UniqueName: \"kubernetes.io/projected/f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e-kube-api-access-b4cgx\") pod \"f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e\" (UID: \"f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e\") " Mar 07 15:17:15 crc kubenswrapper[4943]: I0307 15:17:15.334659 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e-scripts\") pod \"f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e\" (UID: \"f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e\") " Mar 07 15:17:15 crc kubenswrapper[4943]: I0307 15:17:15.334696 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e-swiftconf\") pod \"f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e\" (UID: \"f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e\") " Mar 07 15:17:15 crc kubenswrapper[4943]: I0307 15:17:15.334759 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e-ring-data-devices\") pod \"f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e\" (UID: \"f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e\") " Mar 07 15:17:15 crc kubenswrapper[4943]: I0307 15:17:15.337434 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e" (UID: "f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:17:15 crc kubenswrapper[4943]: I0307 15:17:15.339133 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e" (UID: "f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:17:15 crc kubenswrapper[4943]: I0307 15:17:15.345081 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e-kube-api-access-b4cgx" (OuterVolumeSpecName: "kube-api-access-b4cgx") pod "f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e" (UID: "f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e"). InnerVolumeSpecName "kube-api-access-b4cgx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:17:15 crc kubenswrapper[4943]: I0307 15:17:15.377483 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e" (UID: "f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:17:15 crc kubenswrapper[4943]: I0307 15:17:15.379390 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e-scripts" (OuterVolumeSpecName: "scripts") pod "f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e" (UID: "f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:17:15 crc kubenswrapper[4943]: I0307 15:17:15.390690 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e" (UID: "f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:17:15 crc kubenswrapper[4943]: I0307 15:17:15.436857 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b4cgx\" (UniqueName: \"kubernetes.io/projected/f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e-kube-api-access-b4cgx\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:15 crc kubenswrapper[4943]: I0307 15:17:15.436916 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:15 crc kubenswrapper[4943]: I0307 15:17:15.436958 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:15 crc kubenswrapper[4943]: I0307 15:17:15.436977 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:15 crc kubenswrapper[4943]: I0307 15:17:15.436995 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:15 crc kubenswrapper[4943]: I0307 15:17:15.437014 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:15 crc kubenswrapper[4943]: I0307 15:17:15.755398 4943 scope.go:117] "RemoveContainer" containerID="448308ad6e34ac922d66b3c90e66d0927407a94dd28633ec350502d69e4ca8aa" Mar 07 15:17:15 crc kubenswrapper[4943]: E0307 15:17:15.755733 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cpdmm_openshift-machine-config-operator(4c1f18ea-e557-41fc-95dd-ce4e1722a61e)\"" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" Mar 07 15:17:15 crc kubenswrapper[4943]: I0307 15:17:15.857460 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="96a2fda5679ad42e8805bea507e838efbbe0c107971144ad6fddff3af4589960" Mar 07 15:17:15 crc kubenswrapper[4943]: I0307 15:17:15.857570 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-s966m" Mar 07 15:17:16 crc kubenswrapper[4943]: I0307 15:17:16.439906 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-xl6sl"] Mar 07 15:17:16 crc kubenswrapper[4943]: E0307 15:17:16.440253 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e" containerName="swift-ring-rebalance" Mar 07 15:17:16 crc kubenswrapper[4943]: I0307 15:17:16.440267 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e" containerName="swift-ring-rebalance" Mar 07 15:17:16 crc kubenswrapper[4943]: I0307 15:17:16.440432 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e" containerName="swift-ring-rebalance" Mar 07 15:17:16 crc kubenswrapper[4943]: I0307 15:17:16.441273 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-xl6sl" Mar 07 15:17:16 crc kubenswrapper[4943]: I0307 15:17:16.444749 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:17:16 crc kubenswrapper[4943]: I0307 15:17:16.445183 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:17:16 crc kubenswrapper[4943]: I0307 15:17:16.454454 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-xl6sl"] Mar 07 15:17:16 crc kubenswrapper[4943]: I0307 15:17:16.487790 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/7a1f99d8-a307-4d33-b85c-fb453d65a975-ring-data-devices\") pod \"swift-ring-rebalance-debug-xl6sl\" (UID: \"7a1f99d8-a307-4d33-b85c-fb453d65a975\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xl6sl" Mar 07 15:17:16 crc kubenswrapper[4943]: I0307 15:17:16.487951 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/7a1f99d8-a307-4d33-b85c-fb453d65a975-etc-swift\") pod \"swift-ring-rebalance-debug-xl6sl\" (UID: \"7a1f99d8-a307-4d33-b85c-fb453d65a975\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xl6sl" Mar 07 15:17:16 crc kubenswrapper[4943]: I0307 15:17:16.488056 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7a1f99d8-a307-4d33-b85c-fb453d65a975-scripts\") pod \"swift-ring-rebalance-debug-xl6sl\" (UID: \"7a1f99d8-a307-4d33-b85c-fb453d65a975\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xl6sl" Mar 07 15:17:16 crc kubenswrapper[4943]: I0307 15:17:16.488189 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/7a1f99d8-a307-4d33-b85c-fb453d65a975-swiftconf\") pod \"swift-ring-rebalance-debug-xl6sl\" (UID: \"7a1f99d8-a307-4d33-b85c-fb453d65a975\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xl6sl" Mar 07 15:17:16 crc kubenswrapper[4943]: I0307 15:17:16.488270 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mjkn\" (UniqueName: \"kubernetes.io/projected/7a1f99d8-a307-4d33-b85c-fb453d65a975-kube-api-access-4mjkn\") pod \"swift-ring-rebalance-debug-xl6sl\" (UID: \"7a1f99d8-a307-4d33-b85c-fb453d65a975\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xl6sl" Mar 07 15:17:16 crc kubenswrapper[4943]: I0307 15:17:16.488332 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/7a1f99d8-a307-4d33-b85c-fb453d65a975-dispersionconf\") pod \"swift-ring-rebalance-debug-xl6sl\" (UID: \"7a1f99d8-a307-4d33-b85c-fb453d65a975\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xl6sl" Mar 07 15:17:16 crc kubenswrapper[4943]: I0307 15:17:16.589501 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/7a1f99d8-a307-4d33-b85c-fb453d65a975-dispersionconf\") pod \"swift-ring-rebalance-debug-xl6sl\" (UID: \"7a1f99d8-a307-4d33-b85c-fb453d65a975\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xl6sl" Mar 07 15:17:16 crc kubenswrapper[4943]: I0307 15:17:16.589571 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/7a1f99d8-a307-4d33-b85c-fb453d65a975-ring-data-devices\") pod \"swift-ring-rebalance-debug-xl6sl\" (UID: \"7a1f99d8-a307-4d33-b85c-fb453d65a975\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xl6sl" Mar 07 15:17:16 crc kubenswrapper[4943]: I0307 15:17:16.589605 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/7a1f99d8-a307-4d33-b85c-fb453d65a975-etc-swift\") pod \"swift-ring-rebalance-debug-xl6sl\" (UID: \"7a1f99d8-a307-4d33-b85c-fb453d65a975\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xl6sl" Mar 07 15:17:16 crc kubenswrapper[4943]: I0307 15:17:16.589648 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7a1f99d8-a307-4d33-b85c-fb453d65a975-scripts\") pod \"swift-ring-rebalance-debug-xl6sl\" (UID: \"7a1f99d8-a307-4d33-b85c-fb453d65a975\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xl6sl" Mar 07 15:17:16 crc kubenswrapper[4943]: I0307 15:17:16.589699 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/7a1f99d8-a307-4d33-b85c-fb453d65a975-swiftconf\") pod \"swift-ring-rebalance-debug-xl6sl\" (UID: \"7a1f99d8-a307-4d33-b85c-fb453d65a975\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xl6sl" Mar 07 15:17:16 crc kubenswrapper[4943]: I0307 15:17:16.589737 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mjkn\" (UniqueName: \"kubernetes.io/projected/7a1f99d8-a307-4d33-b85c-fb453d65a975-kube-api-access-4mjkn\") pod \"swift-ring-rebalance-debug-xl6sl\" (UID: \"7a1f99d8-a307-4d33-b85c-fb453d65a975\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xl6sl" Mar 07 15:17:16 crc kubenswrapper[4943]: I0307 15:17:16.590526 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/7a1f99d8-a307-4d33-b85c-fb453d65a975-etc-swift\") pod \"swift-ring-rebalance-debug-xl6sl\" (UID: \"7a1f99d8-a307-4d33-b85c-fb453d65a975\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xl6sl" Mar 07 15:17:16 crc kubenswrapper[4943]: I0307 15:17:16.590864 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7a1f99d8-a307-4d33-b85c-fb453d65a975-scripts\") pod \"swift-ring-rebalance-debug-xl6sl\" (UID: \"7a1f99d8-a307-4d33-b85c-fb453d65a975\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xl6sl" Mar 07 15:17:16 crc kubenswrapper[4943]: I0307 15:17:16.591436 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/7a1f99d8-a307-4d33-b85c-fb453d65a975-ring-data-devices\") pod \"swift-ring-rebalance-debug-xl6sl\" (UID: \"7a1f99d8-a307-4d33-b85c-fb453d65a975\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xl6sl" Mar 07 15:17:16 crc kubenswrapper[4943]: I0307 15:17:16.593696 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/7a1f99d8-a307-4d33-b85c-fb453d65a975-dispersionconf\") pod \"swift-ring-rebalance-debug-xl6sl\" (UID: \"7a1f99d8-a307-4d33-b85c-fb453d65a975\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xl6sl" Mar 07 15:17:16 crc kubenswrapper[4943]: I0307 15:17:16.593702 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/7a1f99d8-a307-4d33-b85c-fb453d65a975-swiftconf\") pod \"swift-ring-rebalance-debug-xl6sl\" (UID: \"7a1f99d8-a307-4d33-b85c-fb453d65a975\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xl6sl" Mar 07 15:17:16 crc kubenswrapper[4943]: I0307 15:17:16.610806 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mjkn\" (UniqueName: \"kubernetes.io/projected/7a1f99d8-a307-4d33-b85c-fb453d65a975-kube-api-access-4mjkn\") pod \"swift-ring-rebalance-debug-xl6sl\" (UID: \"7a1f99d8-a307-4d33-b85c-fb453d65a975\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xl6sl" Mar 07 15:17:16 crc kubenswrapper[4943]: I0307 15:17:16.774150 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e" path="/var/lib/kubelet/pods/f3247b97-4cb6-4839-bbcb-13ed8cdbfd8e/volumes" Mar 07 15:17:16 crc kubenswrapper[4943]: I0307 15:17:16.776883 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-xl6sl" Mar 07 15:17:17 crc kubenswrapper[4943]: I0307 15:17:17.285635 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-xl6sl"] Mar 07 15:17:17 crc kubenswrapper[4943]: I0307 15:17:17.882729 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-xl6sl" event={"ID":"7a1f99d8-a307-4d33-b85c-fb453d65a975","Type":"ContainerStarted","Data":"c88a3b7e42abc90dd6878e360903137e3a79b3509edf5556fcd1f163eb4bafde"} Mar 07 15:17:17 crc kubenswrapper[4943]: I0307 15:17:17.883247 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-xl6sl" event={"ID":"7a1f99d8-a307-4d33-b85c-fb453d65a975","Type":"ContainerStarted","Data":"5b5479b160f6cdf588d6327ee2800ac47b68434ce173699f43230fee98894445"} Mar 07 15:17:17 crc kubenswrapper[4943]: I0307 15:17:17.915554 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-xl6sl" podStartSLOduration=1.915524631 podStartE2EDuration="1.915524631s" podCreationTimestamp="2026-03-07 15:17:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:17:17.909970735 +0000 UTC m=+2279.862107273" watchObservedRunningTime="2026-03-07 15:17:17.915524631 +0000 UTC m=+2279.867661169" Mar 07 15:17:18 crc kubenswrapper[4943]: I0307 15:17:18.903686 4943 generic.go:334] "Generic (PLEG): container finished" podID="7a1f99d8-a307-4d33-b85c-fb453d65a975" containerID="c88a3b7e42abc90dd6878e360903137e3a79b3509edf5556fcd1f163eb4bafde" exitCode=0 Mar 07 15:17:18 crc kubenswrapper[4943]: I0307 15:17:18.903728 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-xl6sl" event={"ID":"7a1f99d8-a307-4d33-b85c-fb453d65a975","Type":"ContainerDied","Data":"c88a3b7e42abc90dd6878e360903137e3a79b3509edf5556fcd1f163eb4bafde"} Mar 07 15:17:20 crc kubenswrapper[4943]: I0307 15:17:20.298145 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-xl6sl" Mar 07 15:17:20 crc kubenswrapper[4943]: I0307 15:17:20.333536 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-xl6sl"] Mar 07 15:17:20 crc kubenswrapper[4943]: I0307 15:17:20.338561 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-xl6sl"] Mar 07 15:17:20 crc kubenswrapper[4943]: I0307 15:17:20.357454 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4mjkn\" (UniqueName: \"kubernetes.io/projected/7a1f99d8-a307-4d33-b85c-fb453d65a975-kube-api-access-4mjkn\") pod \"7a1f99d8-a307-4d33-b85c-fb453d65a975\" (UID: \"7a1f99d8-a307-4d33-b85c-fb453d65a975\") " Mar 07 15:17:20 crc kubenswrapper[4943]: I0307 15:17:20.357508 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/7a1f99d8-a307-4d33-b85c-fb453d65a975-etc-swift\") pod \"7a1f99d8-a307-4d33-b85c-fb453d65a975\" (UID: \"7a1f99d8-a307-4d33-b85c-fb453d65a975\") " Mar 07 15:17:20 crc kubenswrapper[4943]: I0307 15:17:20.357545 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/7a1f99d8-a307-4d33-b85c-fb453d65a975-dispersionconf\") pod \"7a1f99d8-a307-4d33-b85c-fb453d65a975\" (UID: \"7a1f99d8-a307-4d33-b85c-fb453d65a975\") " Mar 07 15:17:20 crc kubenswrapper[4943]: I0307 15:17:20.357576 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/7a1f99d8-a307-4d33-b85c-fb453d65a975-ring-data-devices\") pod \"7a1f99d8-a307-4d33-b85c-fb453d65a975\" (UID: \"7a1f99d8-a307-4d33-b85c-fb453d65a975\") " Mar 07 15:17:20 crc kubenswrapper[4943]: I0307 15:17:20.357607 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7a1f99d8-a307-4d33-b85c-fb453d65a975-scripts\") pod \"7a1f99d8-a307-4d33-b85c-fb453d65a975\" (UID: \"7a1f99d8-a307-4d33-b85c-fb453d65a975\") " Mar 07 15:17:20 crc kubenswrapper[4943]: I0307 15:17:20.357662 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/7a1f99d8-a307-4d33-b85c-fb453d65a975-swiftconf\") pod \"7a1f99d8-a307-4d33-b85c-fb453d65a975\" (UID: \"7a1f99d8-a307-4d33-b85c-fb453d65a975\") " Mar 07 15:17:20 crc kubenswrapper[4943]: I0307 15:17:20.358220 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a1f99d8-a307-4d33-b85c-fb453d65a975-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "7a1f99d8-a307-4d33-b85c-fb453d65a975" (UID: "7a1f99d8-a307-4d33-b85c-fb453d65a975"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:17:20 crc kubenswrapper[4943]: I0307 15:17:20.358362 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a1f99d8-a307-4d33-b85c-fb453d65a975-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "7a1f99d8-a307-4d33-b85c-fb453d65a975" (UID: "7a1f99d8-a307-4d33-b85c-fb453d65a975"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:17:20 crc kubenswrapper[4943]: I0307 15:17:20.362615 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a1f99d8-a307-4d33-b85c-fb453d65a975-kube-api-access-4mjkn" (OuterVolumeSpecName: "kube-api-access-4mjkn") pod "7a1f99d8-a307-4d33-b85c-fb453d65a975" (UID: "7a1f99d8-a307-4d33-b85c-fb453d65a975"). InnerVolumeSpecName "kube-api-access-4mjkn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:17:20 crc kubenswrapper[4943]: I0307 15:17:20.379052 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a1f99d8-a307-4d33-b85c-fb453d65a975-scripts" (OuterVolumeSpecName: "scripts") pod "7a1f99d8-a307-4d33-b85c-fb453d65a975" (UID: "7a1f99d8-a307-4d33-b85c-fb453d65a975"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:17:20 crc kubenswrapper[4943]: I0307 15:17:20.394107 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a1f99d8-a307-4d33-b85c-fb453d65a975-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "7a1f99d8-a307-4d33-b85c-fb453d65a975" (UID: "7a1f99d8-a307-4d33-b85c-fb453d65a975"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:17:20 crc kubenswrapper[4943]: I0307 15:17:20.401128 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a1f99d8-a307-4d33-b85c-fb453d65a975-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "7a1f99d8-a307-4d33-b85c-fb453d65a975" (UID: "7a1f99d8-a307-4d33-b85c-fb453d65a975"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:17:20 crc kubenswrapper[4943]: I0307 15:17:20.458748 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/7a1f99d8-a307-4d33-b85c-fb453d65a975-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:20 crc kubenswrapper[4943]: I0307 15:17:20.458782 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4mjkn\" (UniqueName: \"kubernetes.io/projected/7a1f99d8-a307-4d33-b85c-fb453d65a975-kube-api-access-4mjkn\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:20 crc kubenswrapper[4943]: I0307 15:17:20.458795 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/7a1f99d8-a307-4d33-b85c-fb453d65a975-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:20 crc kubenswrapper[4943]: I0307 15:17:20.458809 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/7a1f99d8-a307-4d33-b85c-fb453d65a975-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:20 crc kubenswrapper[4943]: I0307 15:17:20.458821 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/7a1f99d8-a307-4d33-b85c-fb453d65a975-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:20 crc kubenswrapper[4943]: I0307 15:17:20.458833 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7a1f99d8-a307-4d33-b85c-fb453d65a975-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:20 crc kubenswrapper[4943]: I0307 15:17:20.770566 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a1f99d8-a307-4d33-b85c-fb453d65a975" path="/var/lib/kubelet/pods/7a1f99d8-a307-4d33-b85c-fb453d65a975/volumes" Mar 07 15:17:20 crc kubenswrapper[4943]: I0307 15:17:20.929213 4943 scope.go:117] "RemoveContainer" containerID="c88a3b7e42abc90dd6878e360903137e3a79b3509edf5556fcd1f163eb4bafde" Mar 07 15:17:20 crc kubenswrapper[4943]: I0307 15:17:20.929434 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-xl6sl" Mar 07 15:17:21 crc kubenswrapper[4943]: I0307 15:17:21.482640 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-scd6j"] Mar 07 15:17:21 crc kubenswrapper[4943]: E0307 15:17:21.483554 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a1f99d8-a307-4d33-b85c-fb453d65a975" containerName="swift-ring-rebalance" Mar 07 15:17:21 crc kubenswrapper[4943]: I0307 15:17:21.483583 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a1f99d8-a307-4d33-b85c-fb453d65a975" containerName="swift-ring-rebalance" Mar 07 15:17:21 crc kubenswrapper[4943]: I0307 15:17:21.483866 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a1f99d8-a307-4d33-b85c-fb453d65a975" containerName="swift-ring-rebalance" Mar 07 15:17:21 crc kubenswrapper[4943]: I0307 15:17:21.484706 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-scd6j" Mar 07 15:17:21 crc kubenswrapper[4943]: I0307 15:17:21.487533 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:17:21 crc kubenswrapper[4943]: I0307 15:17:21.488008 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:17:21 crc kubenswrapper[4943]: I0307 15:17:21.503383 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-scd6j"] Mar 07 15:17:21 crc kubenswrapper[4943]: I0307 15:17:21.577229 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8q4m\" (UniqueName: \"kubernetes.io/projected/f78aedf3-b272-440c-8adf-8df4a828022a-kube-api-access-t8q4m\") pod \"swift-ring-rebalance-debug-scd6j\" (UID: \"f78aedf3-b272-440c-8adf-8df4a828022a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-scd6j" Mar 07 15:17:21 crc kubenswrapper[4943]: I0307 15:17:21.577312 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f78aedf3-b272-440c-8adf-8df4a828022a-scripts\") pod \"swift-ring-rebalance-debug-scd6j\" (UID: \"f78aedf3-b272-440c-8adf-8df4a828022a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-scd6j" Mar 07 15:17:21 crc kubenswrapper[4943]: I0307 15:17:21.577365 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f78aedf3-b272-440c-8adf-8df4a828022a-etc-swift\") pod \"swift-ring-rebalance-debug-scd6j\" (UID: \"f78aedf3-b272-440c-8adf-8df4a828022a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-scd6j" Mar 07 15:17:21 crc kubenswrapper[4943]: I0307 15:17:21.577423 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f78aedf3-b272-440c-8adf-8df4a828022a-ring-data-devices\") pod \"swift-ring-rebalance-debug-scd6j\" (UID: \"f78aedf3-b272-440c-8adf-8df4a828022a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-scd6j" Mar 07 15:17:21 crc kubenswrapper[4943]: I0307 15:17:21.577460 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f78aedf3-b272-440c-8adf-8df4a828022a-dispersionconf\") pod \"swift-ring-rebalance-debug-scd6j\" (UID: \"f78aedf3-b272-440c-8adf-8df4a828022a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-scd6j" Mar 07 15:17:21 crc kubenswrapper[4943]: I0307 15:17:21.577654 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f78aedf3-b272-440c-8adf-8df4a828022a-swiftconf\") pod \"swift-ring-rebalance-debug-scd6j\" (UID: \"f78aedf3-b272-440c-8adf-8df4a828022a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-scd6j" Mar 07 15:17:21 crc kubenswrapper[4943]: I0307 15:17:21.678842 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f78aedf3-b272-440c-8adf-8df4a828022a-ring-data-devices\") pod \"swift-ring-rebalance-debug-scd6j\" (UID: \"f78aedf3-b272-440c-8adf-8df4a828022a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-scd6j" Mar 07 15:17:21 crc kubenswrapper[4943]: I0307 15:17:21.678992 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f78aedf3-b272-440c-8adf-8df4a828022a-dispersionconf\") pod \"swift-ring-rebalance-debug-scd6j\" (UID: \"f78aedf3-b272-440c-8adf-8df4a828022a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-scd6j" Mar 07 15:17:21 crc kubenswrapper[4943]: I0307 15:17:21.679247 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f78aedf3-b272-440c-8adf-8df4a828022a-swiftconf\") pod \"swift-ring-rebalance-debug-scd6j\" (UID: \"f78aedf3-b272-440c-8adf-8df4a828022a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-scd6j" Mar 07 15:17:21 crc kubenswrapper[4943]: I0307 15:17:21.679467 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8q4m\" (UniqueName: \"kubernetes.io/projected/f78aedf3-b272-440c-8adf-8df4a828022a-kube-api-access-t8q4m\") pod \"swift-ring-rebalance-debug-scd6j\" (UID: \"f78aedf3-b272-440c-8adf-8df4a828022a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-scd6j" Mar 07 15:17:21 crc kubenswrapper[4943]: I0307 15:17:21.679576 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f78aedf3-b272-440c-8adf-8df4a828022a-scripts\") pod \"swift-ring-rebalance-debug-scd6j\" (UID: \"f78aedf3-b272-440c-8adf-8df4a828022a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-scd6j" Mar 07 15:17:21 crc kubenswrapper[4943]: I0307 15:17:21.679649 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f78aedf3-b272-440c-8adf-8df4a828022a-etc-swift\") pod \"swift-ring-rebalance-debug-scd6j\" (UID: \"f78aedf3-b272-440c-8adf-8df4a828022a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-scd6j" Mar 07 15:17:21 crc kubenswrapper[4943]: I0307 15:17:21.679803 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f78aedf3-b272-440c-8adf-8df4a828022a-ring-data-devices\") pod \"swift-ring-rebalance-debug-scd6j\" (UID: \"f78aedf3-b272-440c-8adf-8df4a828022a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-scd6j" Mar 07 15:17:21 crc kubenswrapper[4943]: I0307 15:17:21.680323 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f78aedf3-b272-440c-8adf-8df4a828022a-etc-swift\") pod \"swift-ring-rebalance-debug-scd6j\" (UID: \"f78aedf3-b272-440c-8adf-8df4a828022a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-scd6j" Mar 07 15:17:21 crc kubenswrapper[4943]: I0307 15:17:21.680869 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f78aedf3-b272-440c-8adf-8df4a828022a-scripts\") pod \"swift-ring-rebalance-debug-scd6j\" (UID: \"f78aedf3-b272-440c-8adf-8df4a828022a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-scd6j" Mar 07 15:17:21 crc kubenswrapper[4943]: I0307 15:17:21.684431 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f78aedf3-b272-440c-8adf-8df4a828022a-dispersionconf\") pod \"swift-ring-rebalance-debug-scd6j\" (UID: \"f78aedf3-b272-440c-8adf-8df4a828022a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-scd6j" Mar 07 15:17:21 crc kubenswrapper[4943]: I0307 15:17:21.686832 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f78aedf3-b272-440c-8adf-8df4a828022a-swiftconf\") pod \"swift-ring-rebalance-debug-scd6j\" (UID: \"f78aedf3-b272-440c-8adf-8df4a828022a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-scd6j" Mar 07 15:17:21 crc kubenswrapper[4943]: I0307 15:17:21.702800 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8q4m\" (UniqueName: \"kubernetes.io/projected/f78aedf3-b272-440c-8adf-8df4a828022a-kube-api-access-t8q4m\") pod \"swift-ring-rebalance-debug-scd6j\" (UID: \"f78aedf3-b272-440c-8adf-8df4a828022a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-scd6j" Mar 07 15:17:21 crc kubenswrapper[4943]: I0307 15:17:21.814564 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-scd6j" Mar 07 15:17:22 crc kubenswrapper[4943]: I0307 15:17:22.137237 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-scd6j"] Mar 07 15:17:22 crc kubenswrapper[4943]: W0307 15:17:22.139995 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf78aedf3_b272_440c_8adf_8df4a828022a.slice/crio-233221f1b77af7dbbf2c51904fab6b1fd360bea57f8724e5f84c394ed82c1f2e WatchSource:0}: Error finding container 233221f1b77af7dbbf2c51904fab6b1fd360bea57f8724e5f84c394ed82c1f2e: Status 404 returned error can't find the container with id 233221f1b77af7dbbf2c51904fab6b1fd360bea57f8724e5f84c394ed82c1f2e Mar 07 15:17:22 crc kubenswrapper[4943]: I0307 15:17:22.251321 4943 scope.go:117] "RemoveContainer" containerID="7b29381c6714b878aa56e314ad179792ca8c473e39c9853b2c64065b9ce0bbfd" Mar 07 15:17:22 crc kubenswrapper[4943]: I0307 15:17:22.294823 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rm5zj" Mar 07 15:17:22 crc kubenswrapper[4943]: I0307 15:17:22.355791 4943 scope.go:117] "RemoveContainer" containerID="8787decac2665a4084b059ae56065aeeb268fd1a232ba5d141389c8e154e64ff" Mar 07 15:17:22 crc kubenswrapper[4943]: I0307 15:17:22.381397 4943 scope.go:117] "RemoveContainer" containerID="7df5bfaa2e10322b574e07b3024a8090cf867a6ba151ab94ed79f3a1da20d2d4" Mar 07 15:17:22 crc kubenswrapper[4943]: I0307 15:17:22.398665 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rm5zj" Mar 07 15:17:22 crc kubenswrapper[4943]: I0307 15:17:22.429534 4943 scope.go:117] "RemoveContainer" containerID="c2616fec6a6b33d05df21ca67ab7a148ccc3c6558a0b33d4d9e5af6e21ec9ec0" Mar 07 15:17:22 crc kubenswrapper[4943]: I0307 15:17:22.483886 4943 scope.go:117] "RemoveContainer" containerID="7f636729abd0912c1a4f9afeb6d60b85afa90aede4d7e0dc6f518232bc393483" Mar 07 15:17:22 crc kubenswrapper[4943]: I0307 15:17:22.513413 4943 scope.go:117] "RemoveContainer" containerID="5856b9b9b3ebd822cf233d14dab3b104f8c51a4851c2cd4234792e319f5b72ff" Mar 07 15:17:22 crc kubenswrapper[4943]: I0307 15:17:22.534542 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rm5zj"] Mar 07 15:17:22 crc kubenswrapper[4943]: I0307 15:17:22.556591 4943 scope.go:117] "RemoveContainer" containerID="897e9786c93077b536da1b7c8d55d5b6d0d17ae15c9c7d354cf6ca12256d7e5c" Mar 07 15:17:22 crc kubenswrapper[4943]: I0307 15:17:22.585630 4943 scope.go:117] "RemoveContainer" containerID="4f054df4e375fb2c1e152506750925311e103138ee9c7e4f5dfe2a8284ff67ce" Mar 07 15:17:22 crc kubenswrapper[4943]: I0307 15:17:22.611257 4943 scope.go:117] "RemoveContainer" containerID="68e7707f6cdf208c71b188376640fe9988be69744d8938ac436f849086359f57" Mar 07 15:17:22 crc kubenswrapper[4943]: I0307 15:17:22.637042 4943 scope.go:117] "RemoveContainer" containerID="1581ae5bf3af5855b7ab2b0c264ff111dca284eedfa2e49a4cf21a6b80de367a" Mar 07 15:17:22 crc kubenswrapper[4943]: I0307 15:17:22.973555 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-scd6j" event={"ID":"f78aedf3-b272-440c-8adf-8df4a828022a","Type":"ContainerStarted","Data":"17f36d0267ff2a0a309f993bbbb9d4fc4ab822fd9f800baa15fe969aaad3074c"} Mar 07 15:17:22 crc kubenswrapper[4943]: I0307 15:17:22.973609 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-scd6j" event={"ID":"f78aedf3-b272-440c-8adf-8df4a828022a","Type":"ContainerStarted","Data":"233221f1b77af7dbbf2c51904fab6b1fd360bea57f8724e5f84c394ed82c1f2e"} Mar 07 15:17:22 crc kubenswrapper[4943]: I0307 15:17:22.997785 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-scd6j" podStartSLOduration=1.9977627 podStartE2EDuration="1.9977627s" podCreationTimestamp="2026-03-07 15:17:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:17:22.996355025 +0000 UTC m=+2284.948491553" watchObservedRunningTime="2026-03-07 15:17:22.9977627 +0000 UTC m=+2284.949899208" Mar 07 15:17:23 crc kubenswrapper[4943]: I0307 15:17:23.985055 4943 generic.go:334] "Generic (PLEG): container finished" podID="f78aedf3-b272-440c-8adf-8df4a828022a" containerID="17f36d0267ff2a0a309f993bbbb9d4fc4ab822fd9f800baa15fe969aaad3074c" exitCode=0 Mar 07 15:17:23 crc kubenswrapper[4943]: I0307 15:17:23.985493 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-scd6j" event={"ID":"f78aedf3-b272-440c-8adf-8df4a828022a","Type":"ContainerDied","Data":"17f36d0267ff2a0a309f993bbbb9d4fc4ab822fd9f800baa15fe969aaad3074c"} Mar 07 15:17:23 crc kubenswrapper[4943]: I0307 15:17:23.985647 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rm5zj" podUID="6c107b9f-c2c1-4261-8c88-bf93b6752bca" containerName="registry-server" containerID="cri-o://901f052294e2dd79323ce8e1b852421f9bbf6ec2bc87b028a78e62a35f3a5196" gracePeriod=2 Mar 07 15:17:24 crc kubenswrapper[4943]: I0307 15:17:24.450031 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rm5zj" Mar 07 15:17:24 crc kubenswrapper[4943]: I0307 15:17:24.533568 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c107b9f-c2c1-4261-8c88-bf93b6752bca-catalog-content\") pod \"6c107b9f-c2c1-4261-8c88-bf93b6752bca\" (UID: \"6c107b9f-c2c1-4261-8c88-bf93b6752bca\") " Mar 07 15:17:24 crc kubenswrapper[4943]: I0307 15:17:24.533635 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c107b9f-c2c1-4261-8c88-bf93b6752bca-utilities\") pod \"6c107b9f-c2c1-4261-8c88-bf93b6752bca\" (UID: \"6c107b9f-c2c1-4261-8c88-bf93b6752bca\") " Mar 07 15:17:24 crc kubenswrapper[4943]: I0307 15:17:24.533690 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vtbp\" (UniqueName: \"kubernetes.io/projected/6c107b9f-c2c1-4261-8c88-bf93b6752bca-kube-api-access-7vtbp\") pod \"6c107b9f-c2c1-4261-8c88-bf93b6752bca\" (UID: \"6c107b9f-c2c1-4261-8c88-bf93b6752bca\") " Mar 07 15:17:24 crc kubenswrapper[4943]: I0307 15:17:24.538561 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c107b9f-c2c1-4261-8c88-bf93b6752bca-utilities" (OuterVolumeSpecName: "utilities") pod "6c107b9f-c2c1-4261-8c88-bf93b6752bca" (UID: "6c107b9f-c2c1-4261-8c88-bf93b6752bca"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:17:24 crc kubenswrapper[4943]: I0307 15:17:24.538869 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c107b9f-c2c1-4261-8c88-bf93b6752bca-kube-api-access-7vtbp" (OuterVolumeSpecName: "kube-api-access-7vtbp") pod "6c107b9f-c2c1-4261-8c88-bf93b6752bca" (UID: "6c107b9f-c2c1-4261-8c88-bf93b6752bca"). InnerVolumeSpecName "kube-api-access-7vtbp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:17:24 crc kubenswrapper[4943]: I0307 15:17:24.636642 4943 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c107b9f-c2c1-4261-8c88-bf93b6752bca-utilities\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:24 crc kubenswrapper[4943]: I0307 15:17:24.636731 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7vtbp\" (UniqueName: \"kubernetes.io/projected/6c107b9f-c2c1-4261-8c88-bf93b6752bca-kube-api-access-7vtbp\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:24 crc kubenswrapper[4943]: I0307 15:17:24.655116 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c107b9f-c2c1-4261-8c88-bf93b6752bca-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6c107b9f-c2c1-4261-8c88-bf93b6752bca" (UID: "6c107b9f-c2c1-4261-8c88-bf93b6752bca"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:17:24 crc kubenswrapper[4943]: I0307 15:17:24.738167 4943 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c107b9f-c2c1-4261-8c88-bf93b6752bca-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:25 crc kubenswrapper[4943]: I0307 15:17:25.002762 4943 generic.go:334] "Generic (PLEG): container finished" podID="6c107b9f-c2c1-4261-8c88-bf93b6752bca" containerID="901f052294e2dd79323ce8e1b852421f9bbf6ec2bc87b028a78e62a35f3a5196" exitCode=0 Mar 07 15:17:25 crc kubenswrapper[4943]: I0307 15:17:25.002854 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rm5zj" event={"ID":"6c107b9f-c2c1-4261-8c88-bf93b6752bca","Type":"ContainerDied","Data":"901f052294e2dd79323ce8e1b852421f9bbf6ec2bc87b028a78e62a35f3a5196"} Mar 07 15:17:25 crc kubenswrapper[4943]: I0307 15:17:25.002917 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rm5zj" Mar 07 15:17:25 crc kubenswrapper[4943]: I0307 15:17:25.002984 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rm5zj" event={"ID":"6c107b9f-c2c1-4261-8c88-bf93b6752bca","Type":"ContainerDied","Data":"118be207ee89b2d2d5996edd79d245faa29d050a7388709f6febbcdfd419c84c"} Mar 07 15:17:25 crc kubenswrapper[4943]: I0307 15:17:25.003027 4943 scope.go:117] "RemoveContainer" containerID="901f052294e2dd79323ce8e1b852421f9bbf6ec2bc87b028a78e62a35f3a5196" Mar 07 15:17:25 crc kubenswrapper[4943]: I0307 15:17:25.037114 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rm5zj"] Mar 07 15:17:25 crc kubenswrapper[4943]: I0307 15:17:25.047661 4943 scope.go:117] "RemoveContainer" containerID="d1a3ddf6a3d3f6a42c77eeb440e7d8ff56b4d6b3a602e0c5d2ce7bd69cc6fcc9" Mar 07 15:17:25 crc kubenswrapper[4943]: I0307 15:17:25.048794 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rm5zj"] Mar 07 15:17:25 crc kubenswrapper[4943]: I0307 15:17:25.090910 4943 scope.go:117] "RemoveContainer" containerID="f5e05905a886de492549a325357cf76509a37d26a6e1592fa227bfe023e6bc01" Mar 07 15:17:25 crc kubenswrapper[4943]: I0307 15:17:25.130260 4943 scope.go:117] "RemoveContainer" containerID="901f052294e2dd79323ce8e1b852421f9bbf6ec2bc87b028a78e62a35f3a5196" Mar 07 15:17:25 crc kubenswrapper[4943]: E0307 15:17:25.132557 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"901f052294e2dd79323ce8e1b852421f9bbf6ec2bc87b028a78e62a35f3a5196\": container with ID starting with 901f052294e2dd79323ce8e1b852421f9bbf6ec2bc87b028a78e62a35f3a5196 not found: ID does not exist" containerID="901f052294e2dd79323ce8e1b852421f9bbf6ec2bc87b028a78e62a35f3a5196" Mar 07 15:17:25 crc kubenswrapper[4943]: I0307 15:17:25.132601 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"901f052294e2dd79323ce8e1b852421f9bbf6ec2bc87b028a78e62a35f3a5196"} err="failed to get container status \"901f052294e2dd79323ce8e1b852421f9bbf6ec2bc87b028a78e62a35f3a5196\": rpc error: code = NotFound desc = could not find container \"901f052294e2dd79323ce8e1b852421f9bbf6ec2bc87b028a78e62a35f3a5196\": container with ID starting with 901f052294e2dd79323ce8e1b852421f9bbf6ec2bc87b028a78e62a35f3a5196 not found: ID does not exist" Mar 07 15:17:25 crc kubenswrapper[4943]: I0307 15:17:25.132633 4943 scope.go:117] "RemoveContainer" containerID="d1a3ddf6a3d3f6a42c77eeb440e7d8ff56b4d6b3a602e0c5d2ce7bd69cc6fcc9" Mar 07 15:17:25 crc kubenswrapper[4943]: E0307 15:17:25.132989 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1a3ddf6a3d3f6a42c77eeb440e7d8ff56b4d6b3a602e0c5d2ce7bd69cc6fcc9\": container with ID starting with d1a3ddf6a3d3f6a42c77eeb440e7d8ff56b4d6b3a602e0c5d2ce7bd69cc6fcc9 not found: ID does not exist" containerID="d1a3ddf6a3d3f6a42c77eeb440e7d8ff56b4d6b3a602e0c5d2ce7bd69cc6fcc9" Mar 07 15:17:25 crc kubenswrapper[4943]: I0307 15:17:25.133027 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1a3ddf6a3d3f6a42c77eeb440e7d8ff56b4d6b3a602e0c5d2ce7bd69cc6fcc9"} err="failed to get container status \"d1a3ddf6a3d3f6a42c77eeb440e7d8ff56b4d6b3a602e0c5d2ce7bd69cc6fcc9\": rpc error: code = NotFound desc = could not find container \"d1a3ddf6a3d3f6a42c77eeb440e7d8ff56b4d6b3a602e0c5d2ce7bd69cc6fcc9\": container with ID starting with d1a3ddf6a3d3f6a42c77eeb440e7d8ff56b4d6b3a602e0c5d2ce7bd69cc6fcc9 not found: ID does not exist" Mar 07 15:17:25 crc kubenswrapper[4943]: I0307 15:17:25.133053 4943 scope.go:117] "RemoveContainer" containerID="f5e05905a886de492549a325357cf76509a37d26a6e1592fa227bfe023e6bc01" Mar 07 15:17:25 crc kubenswrapper[4943]: E0307 15:17:25.133505 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5e05905a886de492549a325357cf76509a37d26a6e1592fa227bfe023e6bc01\": container with ID starting with f5e05905a886de492549a325357cf76509a37d26a6e1592fa227bfe023e6bc01 not found: ID does not exist" containerID="f5e05905a886de492549a325357cf76509a37d26a6e1592fa227bfe023e6bc01" Mar 07 15:17:25 crc kubenswrapper[4943]: I0307 15:17:25.133543 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5e05905a886de492549a325357cf76509a37d26a6e1592fa227bfe023e6bc01"} err="failed to get container status \"f5e05905a886de492549a325357cf76509a37d26a6e1592fa227bfe023e6bc01\": rpc error: code = NotFound desc = could not find container \"f5e05905a886de492549a325357cf76509a37d26a6e1592fa227bfe023e6bc01\": container with ID starting with f5e05905a886de492549a325357cf76509a37d26a6e1592fa227bfe023e6bc01 not found: ID does not exist" Mar 07 15:17:25 crc kubenswrapper[4943]: I0307 15:17:25.261496 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-scd6j" Mar 07 15:17:25 crc kubenswrapper[4943]: I0307 15:17:25.295982 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-scd6j"] Mar 07 15:17:25 crc kubenswrapper[4943]: I0307 15:17:25.305165 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-scd6j"] Mar 07 15:17:25 crc kubenswrapper[4943]: I0307 15:17:25.452488 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f78aedf3-b272-440c-8adf-8df4a828022a-swiftconf\") pod \"f78aedf3-b272-440c-8adf-8df4a828022a\" (UID: \"f78aedf3-b272-440c-8adf-8df4a828022a\") " Mar 07 15:17:25 crc kubenswrapper[4943]: I0307 15:17:25.452614 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f78aedf3-b272-440c-8adf-8df4a828022a-etc-swift\") pod \"f78aedf3-b272-440c-8adf-8df4a828022a\" (UID: \"f78aedf3-b272-440c-8adf-8df4a828022a\") " Mar 07 15:17:25 crc kubenswrapper[4943]: I0307 15:17:25.452694 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f78aedf3-b272-440c-8adf-8df4a828022a-scripts\") pod \"f78aedf3-b272-440c-8adf-8df4a828022a\" (UID: \"f78aedf3-b272-440c-8adf-8df4a828022a\") " Mar 07 15:17:25 crc kubenswrapper[4943]: I0307 15:17:25.452825 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f78aedf3-b272-440c-8adf-8df4a828022a-ring-data-devices\") pod \"f78aedf3-b272-440c-8adf-8df4a828022a\" (UID: \"f78aedf3-b272-440c-8adf-8df4a828022a\") " Mar 07 15:17:25 crc kubenswrapper[4943]: I0307 15:17:25.452863 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f78aedf3-b272-440c-8adf-8df4a828022a-dispersionconf\") pod \"f78aedf3-b272-440c-8adf-8df4a828022a\" (UID: \"f78aedf3-b272-440c-8adf-8df4a828022a\") " Mar 07 15:17:25 crc kubenswrapper[4943]: I0307 15:17:25.452974 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t8q4m\" (UniqueName: \"kubernetes.io/projected/f78aedf3-b272-440c-8adf-8df4a828022a-kube-api-access-t8q4m\") pod \"f78aedf3-b272-440c-8adf-8df4a828022a\" (UID: \"f78aedf3-b272-440c-8adf-8df4a828022a\") " Mar 07 15:17:25 crc kubenswrapper[4943]: I0307 15:17:25.453484 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f78aedf3-b272-440c-8adf-8df4a828022a-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "f78aedf3-b272-440c-8adf-8df4a828022a" (UID: "f78aedf3-b272-440c-8adf-8df4a828022a"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:17:25 crc kubenswrapper[4943]: I0307 15:17:25.453555 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f78aedf3-b272-440c-8adf-8df4a828022a-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "f78aedf3-b272-440c-8adf-8df4a828022a" (UID: "f78aedf3-b272-440c-8adf-8df4a828022a"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:17:25 crc kubenswrapper[4943]: I0307 15:17:25.454353 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f78aedf3-b272-440c-8adf-8df4a828022a-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:25 crc kubenswrapper[4943]: I0307 15:17:25.454392 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f78aedf3-b272-440c-8adf-8df4a828022a-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:25 crc kubenswrapper[4943]: I0307 15:17:25.459580 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f78aedf3-b272-440c-8adf-8df4a828022a-kube-api-access-t8q4m" (OuterVolumeSpecName: "kube-api-access-t8q4m") pod "f78aedf3-b272-440c-8adf-8df4a828022a" (UID: "f78aedf3-b272-440c-8adf-8df4a828022a"). InnerVolumeSpecName "kube-api-access-t8q4m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:17:25 crc kubenswrapper[4943]: I0307 15:17:25.480803 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f78aedf3-b272-440c-8adf-8df4a828022a-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "f78aedf3-b272-440c-8adf-8df4a828022a" (UID: "f78aedf3-b272-440c-8adf-8df4a828022a"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:17:25 crc kubenswrapper[4943]: I0307 15:17:25.493168 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f78aedf3-b272-440c-8adf-8df4a828022a-scripts" (OuterVolumeSpecName: "scripts") pod "f78aedf3-b272-440c-8adf-8df4a828022a" (UID: "f78aedf3-b272-440c-8adf-8df4a828022a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:17:25 crc kubenswrapper[4943]: I0307 15:17:25.494591 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f78aedf3-b272-440c-8adf-8df4a828022a-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "f78aedf3-b272-440c-8adf-8df4a828022a" (UID: "f78aedf3-b272-440c-8adf-8df4a828022a"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:17:25 crc kubenswrapper[4943]: I0307 15:17:25.555998 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t8q4m\" (UniqueName: \"kubernetes.io/projected/f78aedf3-b272-440c-8adf-8df4a828022a-kube-api-access-t8q4m\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:25 crc kubenswrapper[4943]: I0307 15:17:25.556150 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f78aedf3-b272-440c-8adf-8df4a828022a-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:25 crc kubenswrapper[4943]: I0307 15:17:25.556230 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f78aedf3-b272-440c-8adf-8df4a828022a-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:25 crc kubenswrapper[4943]: I0307 15:17:25.556304 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f78aedf3-b272-440c-8adf-8df4a828022a-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:26 crc kubenswrapper[4943]: I0307 15:17:26.029801 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="233221f1b77af7dbbf2c51904fab6b1fd360bea57f8724e5f84c394ed82c1f2e" Mar 07 15:17:26 crc kubenswrapper[4943]: I0307 15:17:26.029894 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-scd6j" Mar 07 15:17:26 crc kubenswrapper[4943]: I0307 15:17:26.415728 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-j84nx"] Mar 07 15:17:26 crc kubenswrapper[4943]: E0307 15:17:26.416484 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c107b9f-c2c1-4261-8c88-bf93b6752bca" containerName="extract-content" Mar 07 15:17:26 crc kubenswrapper[4943]: I0307 15:17:26.416553 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c107b9f-c2c1-4261-8c88-bf93b6752bca" containerName="extract-content" Mar 07 15:17:26 crc kubenswrapper[4943]: E0307 15:17:26.416621 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c107b9f-c2c1-4261-8c88-bf93b6752bca" containerName="extract-utilities" Mar 07 15:17:26 crc kubenswrapper[4943]: I0307 15:17:26.416688 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c107b9f-c2c1-4261-8c88-bf93b6752bca" containerName="extract-utilities" Mar 07 15:17:26 crc kubenswrapper[4943]: E0307 15:17:26.416763 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c107b9f-c2c1-4261-8c88-bf93b6752bca" containerName="registry-server" Mar 07 15:17:26 crc kubenswrapper[4943]: I0307 15:17:26.416812 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c107b9f-c2c1-4261-8c88-bf93b6752bca" containerName="registry-server" Mar 07 15:17:26 crc kubenswrapper[4943]: E0307 15:17:26.416881 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f78aedf3-b272-440c-8adf-8df4a828022a" containerName="swift-ring-rebalance" Mar 07 15:17:26 crc kubenswrapper[4943]: I0307 15:17:26.416954 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="f78aedf3-b272-440c-8adf-8df4a828022a" containerName="swift-ring-rebalance" Mar 07 15:17:26 crc kubenswrapper[4943]: I0307 15:17:26.417149 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c107b9f-c2c1-4261-8c88-bf93b6752bca" containerName="registry-server" Mar 07 15:17:26 crc kubenswrapper[4943]: I0307 15:17:26.417223 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="f78aedf3-b272-440c-8adf-8df4a828022a" containerName="swift-ring-rebalance" Mar 07 15:17:26 crc kubenswrapper[4943]: I0307 15:17:26.417718 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j84nx" Mar 07 15:17:26 crc kubenswrapper[4943]: I0307 15:17:26.419812 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:17:26 crc kubenswrapper[4943]: I0307 15:17:26.420345 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:17:26 crc kubenswrapper[4943]: I0307 15:17:26.427628 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-j84nx"] Mar 07 15:17:26 crc kubenswrapper[4943]: I0307 15:17:26.571121 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/397919e0-109d-45b9-a465-8b527d67c7a7-dispersionconf\") pod \"swift-ring-rebalance-debug-j84nx\" (UID: \"397919e0-109d-45b9-a465-8b527d67c7a7\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j84nx" Mar 07 15:17:26 crc kubenswrapper[4943]: I0307 15:17:26.571158 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfw6n\" (UniqueName: \"kubernetes.io/projected/397919e0-109d-45b9-a465-8b527d67c7a7-kube-api-access-vfw6n\") pod \"swift-ring-rebalance-debug-j84nx\" (UID: \"397919e0-109d-45b9-a465-8b527d67c7a7\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j84nx" Mar 07 15:17:26 crc kubenswrapper[4943]: I0307 15:17:26.571198 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/397919e0-109d-45b9-a465-8b527d67c7a7-etc-swift\") pod \"swift-ring-rebalance-debug-j84nx\" (UID: \"397919e0-109d-45b9-a465-8b527d67c7a7\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j84nx" Mar 07 15:17:26 crc kubenswrapper[4943]: I0307 15:17:26.571241 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/397919e0-109d-45b9-a465-8b527d67c7a7-scripts\") pod \"swift-ring-rebalance-debug-j84nx\" (UID: \"397919e0-109d-45b9-a465-8b527d67c7a7\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j84nx" Mar 07 15:17:26 crc kubenswrapper[4943]: I0307 15:17:26.571278 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/397919e0-109d-45b9-a465-8b527d67c7a7-swiftconf\") pod \"swift-ring-rebalance-debug-j84nx\" (UID: \"397919e0-109d-45b9-a465-8b527d67c7a7\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j84nx" Mar 07 15:17:26 crc kubenswrapper[4943]: I0307 15:17:26.571334 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/397919e0-109d-45b9-a465-8b527d67c7a7-ring-data-devices\") pod \"swift-ring-rebalance-debug-j84nx\" (UID: \"397919e0-109d-45b9-a465-8b527d67c7a7\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j84nx" Mar 07 15:17:26 crc kubenswrapper[4943]: I0307 15:17:26.672548 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/397919e0-109d-45b9-a465-8b527d67c7a7-etc-swift\") pod \"swift-ring-rebalance-debug-j84nx\" (UID: \"397919e0-109d-45b9-a465-8b527d67c7a7\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j84nx" Mar 07 15:17:26 crc kubenswrapper[4943]: I0307 15:17:26.672623 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/397919e0-109d-45b9-a465-8b527d67c7a7-scripts\") pod \"swift-ring-rebalance-debug-j84nx\" (UID: \"397919e0-109d-45b9-a465-8b527d67c7a7\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j84nx" Mar 07 15:17:26 crc kubenswrapper[4943]: I0307 15:17:26.672683 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/397919e0-109d-45b9-a465-8b527d67c7a7-swiftconf\") pod \"swift-ring-rebalance-debug-j84nx\" (UID: \"397919e0-109d-45b9-a465-8b527d67c7a7\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j84nx" Mar 07 15:17:26 crc kubenswrapper[4943]: I0307 15:17:26.672764 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/397919e0-109d-45b9-a465-8b527d67c7a7-ring-data-devices\") pod \"swift-ring-rebalance-debug-j84nx\" (UID: \"397919e0-109d-45b9-a465-8b527d67c7a7\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j84nx" Mar 07 15:17:26 crc kubenswrapper[4943]: I0307 15:17:26.672858 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfw6n\" (UniqueName: \"kubernetes.io/projected/397919e0-109d-45b9-a465-8b527d67c7a7-kube-api-access-vfw6n\") pod \"swift-ring-rebalance-debug-j84nx\" (UID: \"397919e0-109d-45b9-a465-8b527d67c7a7\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j84nx" Mar 07 15:17:26 crc kubenswrapper[4943]: I0307 15:17:26.672890 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/397919e0-109d-45b9-a465-8b527d67c7a7-dispersionconf\") pod \"swift-ring-rebalance-debug-j84nx\" (UID: \"397919e0-109d-45b9-a465-8b527d67c7a7\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j84nx" Mar 07 15:17:26 crc kubenswrapper[4943]: I0307 15:17:26.673011 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/397919e0-109d-45b9-a465-8b527d67c7a7-etc-swift\") pod \"swift-ring-rebalance-debug-j84nx\" (UID: \"397919e0-109d-45b9-a465-8b527d67c7a7\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j84nx" Mar 07 15:17:26 crc kubenswrapper[4943]: I0307 15:17:26.673608 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/397919e0-109d-45b9-a465-8b527d67c7a7-ring-data-devices\") pod \"swift-ring-rebalance-debug-j84nx\" (UID: \"397919e0-109d-45b9-a465-8b527d67c7a7\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j84nx" Mar 07 15:17:26 crc kubenswrapper[4943]: I0307 15:17:26.674043 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/397919e0-109d-45b9-a465-8b527d67c7a7-scripts\") pod \"swift-ring-rebalance-debug-j84nx\" (UID: \"397919e0-109d-45b9-a465-8b527d67c7a7\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j84nx" Mar 07 15:17:26 crc kubenswrapper[4943]: I0307 15:17:26.681375 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/397919e0-109d-45b9-a465-8b527d67c7a7-dispersionconf\") pod \"swift-ring-rebalance-debug-j84nx\" (UID: \"397919e0-109d-45b9-a465-8b527d67c7a7\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j84nx" Mar 07 15:17:26 crc kubenswrapper[4943]: I0307 15:17:26.689257 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/397919e0-109d-45b9-a465-8b527d67c7a7-swiftconf\") pod \"swift-ring-rebalance-debug-j84nx\" (UID: \"397919e0-109d-45b9-a465-8b527d67c7a7\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j84nx" Mar 07 15:17:26 crc kubenswrapper[4943]: I0307 15:17:26.706603 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfw6n\" (UniqueName: \"kubernetes.io/projected/397919e0-109d-45b9-a465-8b527d67c7a7-kube-api-access-vfw6n\") pod \"swift-ring-rebalance-debug-j84nx\" (UID: \"397919e0-109d-45b9-a465-8b527d67c7a7\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j84nx" Mar 07 15:17:26 crc kubenswrapper[4943]: I0307 15:17:26.745329 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j84nx" Mar 07 15:17:26 crc kubenswrapper[4943]: I0307 15:17:26.755599 4943 scope.go:117] "RemoveContainer" containerID="448308ad6e34ac922d66b3c90e66d0927407a94dd28633ec350502d69e4ca8aa" Mar 07 15:17:26 crc kubenswrapper[4943]: E0307 15:17:26.755828 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cpdmm_openshift-machine-config-operator(4c1f18ea-e557-41fc-95dd-ce4e1722a61e)\"" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" Mar 07 15:17:26 crc kubenswrapper[4943]: I0307 15:17:26.768704 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c107b9f-c2c1-4261-8c88-bf93b6752bca" path="/var/lib/kubelet/pods/6c107b9f-c2c1-4261-8c88-bf93b6752bca/volumes" Mar 07 15:17:26 crc kubenswrapper[4943]: I0307 15:17:26.770297 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f78aedf3-b272-440c-8adf-8df4a828022a" path="/var/lib/kubelet/pods/f78aedf3-b272-440c-8adf-8df4a828022a/volumes" Mar 07 15:17:27 crc kubenswrapper[4943]: I0307 15:17:27.074024 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-j84nx"] Mar 07 15:17:27 crc kubenswrapper[4943]: W0307 15:17:27.076445 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod397919e0_109d_45b9_a465_8b527d67c7a7.slice/crio-bd97e5893e87538bdf63d9f45c65f330467232e69bd4445adb6d85fc5fdeac73 WatchSource:0}: Error finding container bd97e5893e87538bdf63d9f45c65f330467232e69bd4445adb6d85fc5fdeac73: Status 404 returned error can't find the container with id bd97e5893e87538bdf63d9f45c65f330467232e69bd4445adb6d85fc5fdeac73 Mar 07 15:17:28 crc kubenswrapper[4943]: I0307 15:17:28.067807 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j84nx" event={"ID":"397919e0-109d-45b9-a465-8b527d67c7a7","Type":"ContainerStarted","Data":"6f02b1dd9fbdb7122d13c97d606c9cd5dda38164092e8957783ef2912d0d3ec0"} Mar 07 15:17:28 crc kubenswrapper[4943]: I0307 15:17:28.068568 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j84nx" event={"ID":"397919e0-109d-45b9-a465-8b527d67c7a7","Type":"ContainerStarted","Data":"bd97e5893e87538bdf63d9f45c65f330467232e69bd4445adb6d85fc5fdeac73"} Mar 07 15:17:28 crc kubenswrapper[4943]: I0307 15:17:28.087252 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j84nx" podStartSLOduration=2.087220845 podStartE2EDuration="2.087220845s" podCreationTimestamp="2026-03-07 15:17:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:17:28.084276473 +0000 UTC m=+2290.036412981" watchObservedRunningTime="2026-03-07 15:17:28.087220845 +0000 UTC m=+2290.039357383" Mar 07 15:17:29 crc kubenswrapper[4943]: I0307 15:17:29.083022 4943 generic.go:334] "Generic (PLEG): container finished" podID="397919e0-109d-45b9-a465-8b527d67c7a7" containerID="6f02b1dd9fbdb7122d13c97d606c9cd5dda38164092e8957783ef2912d0d3ec0" exitCode=0 Mar 07 15:17:29 crc kubenswrapper[4943]: I0307 15:17:29.083110 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j84nx" event={"ID":"397919e0-109d-45b9-a465-8b527d67c7a7","Type":"ContainerDied","Data":"6f02b1dd9fbdb7122d13c97d606c9cd5dda38164092e8957783ef2912d0d3ec0"} Mar 07 15:17:30 crc kubenswrapper[4943]: I0307 15:17:30.365670 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j84nx" Mar 07 15:17:30 crc kubenswrapper[4943]: I0307 15:17:30.405565 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-j84nx"] Mar 07 15:17:30 crc kubenswrapper[4943]: I0307 15:17:30.412335 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-j84nx"] Mar 07 15:17:30 crc kubenswrapper[4943]: I0307 15:17:30.537599 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/397919e0-109d-45b9-a465-8b527d67c7a7-etc-swift\") pod \"397919e0-109d-45b9-a465-8b527d67c7a7\" (UID: \"397919e0-109d-45b9-a465-8b527d67c7a7\") " Mar 07 15:17:30 crc kubenswrapper[4943]: I0307 15:17:30.537668 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/397919e0-109d-45b9-a465-8b527d67c7a7-swiftconf\") pod \"397919e0-109d-45b9-a465-8b527d67c7a7\" (UID: \"397919e0-109d-45b9-a465-8b527d67c7a7\") " Mar 07 15:17:30 crc kubenswrapper[4943]: I0307 15:17:30.537768 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/397919e0-109d-45b9-a465-8b527d67c7a7-dispersionconf\") pod \"397919e0-109d-45b9-a465-8b527d67c7a7\" (UID: \"397919e0-109d-45b9-a465-8b527d67c7a7\") " Mar 07 15:17:30 crc kubenswrapper[4943]: I0307 15:17:30.537801 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/397919e0-109d-45b9-a465-8b527d67c7a7-ring-data-devices\") pod \"397919e0-109d-45b9-a465-8b527d67c7a7\" (UID: \"397919e0-109d-45b9-a465-8b527d67c7a7\") " Mar 07 15:17:30 crc kubenswrapper[4943]: I0307 15:17:30.537872 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/397919e0-109d-45b9-a465-8b527d67c7a7-scripts\") pod \"397919e0-109d-45b9-a465-8b527d67c7a7\" (UID: \"397919e0-109d-45b9-a465-8b527d67c7a7\") " Mar 07 15:17:30 crc kubenswrapper[4943]: I0307 15:17:30.537921 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vfw6n\" (UniqueName: \"kubernetes.io/projected/397919e0-109d-45b9-a465-8b527d67c7a7-kube-api-access-vfw6n\") pod \"397919e0-109d-45b9-a465-8b527d67c7a7\" (UID: \"397919e0-109d-45b9-a465-8b527d67c7a7\") " Mar 07 15:17:30 crc kubenswrapper[4943]: I0307 15:17:30.538639 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/397919e0-109d-45b9-a465-8b527d67c7a7-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "397919e0-109d-45b9-a465-8b527d67c7a7" (UID: "397919e0-109d-45b9-a465-8b527d67c7a7"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:17:30 crc kubenswrapper[4943]: I0307 15:17:30.538682 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/397919e0-109d-45b9-a465-8b527d67c7a7-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "397919e0-109d-45b9-a465-8b527d67c7a7" (UID: "397919e0-109d-45b9-a465-8b527d67c7a7"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:17:30 crc kubenswrapper[4943]: I0307 15:17:30.547192 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/397919e0-109d-45b9-a465-8b527d67c7a7-kube-api-access-vfw6n" (OuterVolumeSpecName: "kube-api-access-vfw6n") pod "397919e0-109d-45b9-a465-8b527d67c7a7" (UID: "397919e0-109d-45b9-a465-8b527d67c7a7"). InnerVolumeSpecName "kube-api-access-vfw6n". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:17:30 crc kubenswrapper[4943]: I0307 15:17:30.571296 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/397919e0-109d-45b9-a465-8b527d67c7a7-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "397919e0-109d-45b9-a465-8b527d67c7a7" (UID: "397919e0-109d-45b9-a465-8b527d67c7a7"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:17:30 crc kubenswrapper[4943]: I0307 15:17:30.579375 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/397919e0-109d-45b9-a465-8b527d67c7a7-scripts" (OuterVolumeSpecName: "scripts") pod "397919e0-109d-45b9-a465-8b527d67c7a7" (UID: "397919e0-109d-45b9-a465-8b527d67c7a7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:17:30 crc kubenswrapper[4943]: I0307 15:17:30.590160 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/397919e0-109d-45b9-a465-8b527d67c7a7-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "397919e0-109d-45b9-a465-8b527d67c7a7" (UID: "397919e0-109d-45b9-a465-8b527d67c7a7"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:17:30 crc kubenswrapper[4943]: I0307 15:17:30.640540 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/397919e0-109d-45b9-a465-8b527d67c7a7-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:30 crc kubenswrapper[4943]: I0307 15:17:30.640595 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/397919e0-109d-45b9-a465-8b527d67c7a7-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:30 crc kubenswrapper[4943]: I0307 15:17:30.640616 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/397919e0-109d-45b9-a465-8b527d67c7a7-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:30 crc kubenswrapper[4943]: I0307 15:17:30.640659 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vfw6n\" (UniqueName: \"kubernetes.io/projected/397919e0-109d-45b9-a465-8b527d67c7a7-kube-api-access-vfw6n\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:30 crc kubenswrapper[4943]: I0307 15:17:30.640682 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/397919e0-109d-45b9-a465-8b527d67c7a7-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:30 crc kubenswrapper[4943]: I0307 15:17:30.640702 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/397919e0-109d-45b9-a465-8b527d67c7a7-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:30 crc kubenswrapper[4943]: I0307 15:17:30.773450 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="397919e0-109d-45b9-a465-8b527d67c7a7" path="/var/lib/kubelet/pods/397919e0-109d-45b9-a465-8b527d67c7a7/volumes" Mar 07 15:17:31 crc kubenswrapper[4943]: I0307 15:17:31.109283 4943 scope.go:117] "RemoveContainer" containerID="6f02b1dd9fbdb7122d13c97d606c9cd5dda38164092e8957783ef2912d0d3ec0" Mar 07 15:17:31 crc kubenswrapper[4943]: I0307 15:17:31.109321 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j84nx" Mar 07 15:17:31 crc kubenswrapper[4943]: I0307 15:17:31.537914 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-t7564"] Mar 07 15:17:31 crc kubenswrapper[4943]: E0307 15:17:31.541053 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="397919e0-109d-45b9-a465-8b527d67c7a7" containerName="swift-ring-rebalance" Mar 07 15:17:31 crc kubenswrapper[4943]: I0307 15:17:31.541104 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="397919e0-109d-45b9-a465-8b527d67c7a7" containerName="swift-ring-rebalance" Mar 07 15:17:31 crc kubenswrapper[4943]: I0307 15:17:31.556495 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="397919e0-109d-45b9-a465-8b527d67c7a7" containerName="swift-ring-rebalance" Mar 07 15:17:31 crc kubenswrapper[4943]: I0307 15:17:31.557674 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-t7564" Mar 07 15:17:31 crc kubenswrapper[4943]: I0307 15:17:31.558478 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-t7564"] Mar 07 15:17:31 crc kubenswrapper[4943]: I0307 15:17:31.561339 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:17:31 crc kubenswrapper[4943]: I0307 15:17:31.561568 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:17:31 crc kubenswrapper[4943]: I0307 15:17:31.658629 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e3716c1c-ce34-48ae-952f-cf16c10baf9b-swiftconf\") pod \"swift-ring-rebalance-debug-t7564\" (UID: \"e3716c1c-ce34-48ae-952f-cf16c10baf9b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-t7564" Mar 07 15:17:31 crc kubenswrapper[4943]: I0307 15:17:31.658675 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e3716c1c-ce34-48ae-952f-cf16c10baf9b-etc-swift\") pod \"swift-ring-rebalance-debug-t7564\" (UID: \"e3716c1c-ce34-48ae-952f-cf16c10baf9b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-t7564" Mar 07 15:17:31 crc kubenswrapper[4943]: I0307 15:17:31.658710 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e3716c1c-ce34-48ae-952f-cf16c10baf9b-scripts\") pod \"swift-ring-rebalance-debug-t7564\" (UID: \"e3716c1c-ce34-48ae-952f-cf16c10baf9b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-t7564" Mar 07 15:17:31 crc kubenswrapper[4943]: I0307 15:17:31.658897 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e3716c1c-ce34-48ae-952f-cf16c10baf9b-dispersionconf\") pod \"swift-ring-rebalance-debug-t7564\" (UID: \"e3716c1c-ce34-48ae-952f-cf16c10baf9b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-t7564" Mar 07 15:17:31 crc kubenswrapper[4943]: I0307 15:17:31.659001 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e3716c1c-ce34-48ae-952f-cf16c10baf9b-ring-data-devices\") pod \"swift-ring-rebalance-debug-t7564\" (UID: \"e3716c1c-ce34-48ae-952f-cf16c10baf9b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-t7564" Mar 07 15:17:31 crc kubenswrapper[4943]: I0307 15:17:31.659266 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxzmv\" (UniqueName: \"kubernetes.io/projected/e3716c1c-ce34-48ae-952f-cf16c10baf9b-kube-api-access-fxzmv\") pod \"swift-ring-rebalance-debug-t7564\" (UID: \"e3716c1c-ce34-48ae-952f-cf16c10baf9b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-t7564" Mar 07 15:17:31 crc kubenswrapper[4943]: I0307 15:17:31.760586 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e3716c1c-ce34-48ae-952f-cf16c10baf9b-ring-data-devices\") pod \"swift-ring-rebalance-debug-t7564\" (UID: \"e3716c1c-ce34-48ae-952f-cf16c10baf9b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-t7564" Mar 07 15:17:31 crc kubenswrapper[4943]: I0307 15:17:31.760701 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxzmv\" (UniqueName: \"kubernetes.io/projected/e3716c1c-ce34-48ae-952f-cf16c10baf9b-kube-api-access-fxzmv\") pod \"swift-ring-rebalance-debug-t7564\" (UID: \"e3716c1c-ce34-48ae-952f-cf16c10baf9b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-t7564" Mar 07 15:17:31 crc kubenswrapper[4943]: I0307 15:17:31.760795 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e3716c1c-ce34-48ae-952f-cf16c10baf9b-swiftconf\") pod \"swift-ring-rebalance-debug-t7564\" (UID: \"e3716c1c-ce34-48ae-952f-cf16c10baf9b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-t7564" Mar 07 15:17:31 crc kubenswrapper[4943]: I0307 15:17:31.760835 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e3716c1c-ce34-48ae-952f-cf16c10baf9b-etc-swift\") pod \"swift-ring-rebalance-debug-t7564\" (UID: \"e3716c1c-ce34-48ae-952f-cf16c10baf9b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-t7564" Mar 07 15:17:31 crc kubenswrapper[4943]: I0307 15:17:31.760891 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e3716c1c-ce34-48ae-952f-cf16c10baf9b-scripts\") pod \"swift-ring-rebalance-debug-t7564\" (UID: \"e3716c1c-ce34-48ae-952f-cf16c10baf9b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-t7564" Mar 07 15:17:31 crc kubenswrapper[4943]: I0307 15:17:31.761014 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e3716c1c-ce34-48ae-952f-cf16c10baf9b-dispersionconf\") pod \"swift-ring-rebalance-debug-t7564\" (UID: \"e3716c1c-ce34-48ae-952f-cf16c10baf9b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-t7564" Mar 07 15:17:31 crc kubenswrapper[4943]: I0307 15:17:31.761511 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e3716c1c-ce34-48ae-952f-cf16c10baf9b-etc-swift\") pod \"swift-ring-rebalance-debug-t7564\" (UID: \"e3716c1c-ce34-48ae-952f-cf16c10baf9b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-t7564" Mar 07 15:17:31 crc kubenswrapper[4943]: I0307 15:17:31.761529 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e3716c1c-ce34-48ae-952f-cf16c10baf9b-scripts\") pod \"swift-ring-rebalance-debug-t7564\" (UID: \"e3716c1c-ce34-48ae-952f-cf16c10baf9b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-t7564" Mar 07 15:17:31 crc kubenswrapper[4943]: I0307 15:17:31.761639 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e3716c1c-ce34-48ae-952f-cf16c10baf9b-ring-data-devices\") pod \"swift-ring-rebalance-debug-t7564\" (UID: \"e3716c1c-ce34-48ae-952f-cf16c10baf9b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-t7564" Mar 07 15:17:31 crc kubenswrapper[4943]: I0307 15:17:31.765912 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e3716c1c-ce34-48ae-952f-cf16c10baf9b-swiftconf\") pod \"swift-ring-rebalance-debug-t7564\" (UID: \"e3716c1c-ce34-48ae-952f-cf16c10baf9b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-t7564" Mar 07 15:17:31 crc kubenswrapper[4943]: I0307 15:17:31.766781 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e3716c1c-ce34-48ae-952f-cf16c10baf9b-dispersionconf\") pod \"swift-ring-rebalance-debug-t7564\" (UID: \"e3716c1c-ce34-48ae-952f-cf16c10baf9b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-t7564" Mar 07 15:17:31 crc kubenswrapper[4943]: I0307 15:17:31.779864 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxzmv\" (UniqueName: \"kubernetes.io/projected/e3716c1c-ce34-48ae-952f-cf16c10baf9b-kube-api-access-fxzmv\") pod \"swift-ring-rebalance-debug-t7564\" (UID: \"e3716c1c-ce34-48ae-952f-cf16c10baf9b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-t7564" Mar 07 15:17:31 crc kubenswrapper[4943]: I0307 15:17:31.901974 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-t7564" Mar 07 15:17:32 crc kubenswrapper[4943]: I0307 15:17:32.142949 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-t7564"] Mar 07 15:17:33 crc kubenswrapper[4943]: I0307 15:17:33.152284 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-t7564" event={"ID":"e3716c1c-ce34-48ae-952f-cf16c10baf9b","Type":"ContainerStarted","Data":"d29f789a22230e7ae99d1c296f7d812000d0d2ffafa7c90a27fd516111a961e0"} Mar 07 15:17:33 crc kubenswrapper[4943]: I0307 15:17:33.152644 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-t7564" event={"ID":"e3716c1c-ce34-48ae-952f-cf16c10baf9b","Type":"ContainerStarted","Data":"a5ea3a2971232efb83738b6041356610b37faaaedab261b2beefa11924d8e73d"} Mar 07 15:17:33 crc kubenswrapper[4943]: I0307 15:17:33.192382 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-t7564" podStartSLOduration=2.192356622 podStartE2EDuration="2.192356622s" podCreationTimestamp="2026-03-07 15:17:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:17:33.174711071 +0000 UTC m=+2295.126847619" watchObservedRunningTime="2026-03-07 15:17:33.192356622 +0000 UTC m=+2295.144493160" Mar 07 15:17:34 crc kubenswrapper[4943]: I0307 15:17:34.167709 4943 generic.go:334] "Generic (PLEG): container finished" podID="e3716c1c-ce34-48ae-952f-cf16c10baf9b" containerID="d29f789a22230e7ae99d1c296f7d812000d0d2ffafa7c90a27fd516111a961e0" exitCode=0 Mar 07 15:17:34 crc kubenswrapper[4943]: I0307 15:17:34.167784 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-t7564" event={"ID":"e3716c1c-ce34-48ae-952f-cf16c10baf9b","Type":"ContainerDied","Data":"d29f789a22230e7ae99d1c296f7d812000d0d2ffafa7c90a27fd516111a961e0"} Mar 07 15:17:35 crc kubenswrapper[4943]: I0307 15:17:35.500790 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-t7564" Mar 07 15:17:35 crc kubenswrapper[4943]: I0307 15:17:35.533877 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-t7564"] Mar 07 15:17:35 crc kubenswrapper[4943]: I0307 15:17:35.538945 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-t7564"] Mar 07 15:17:35 crc kubenswrapper[4943]: I0307 15:17:35.620203 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e3716c1c-ce34-48ae-952f-cf16c10baf9b-ring-data-devices\") pod \"e3716c1c-ce34-48ae-952f-cf16c10baf9b\" (UID: \"e3716c1c-ce34-48ae-952f-cf16c10baf9b\") " Mar 07 15:17:35 crc kubenswrapper[4943]: I0307 15:17:35.620306 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e3716c1c-ce34-48ae-952f-cf16c10baf9b-etc-swift\") pod \"e3716c1c-ce34-48ae-952f-cf16c10baf9b\" (UID: \"e3716c1c-ce34-48ae-952f-cf16c10baf9b\") " Mar 07 15:17:35 crc kubenswrapper[4943]: I0307 15:17:35.620362 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e3716c1c-ce34-48ae-952f-cf16c10baf9b-swiftconf\") pod \"e3716c1c-ce34-48ae-952f-cf16c10baf9b\" (UID: \"e3716c1c-ce34-48ae-952f-cf16c10baf9b\") " Mar 07 15:17:35 crc kubenswrapper[4943]: I0307 15:17:35.620444 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e3716c1c-ce34-48ae-952f-cf16c10baf9b-scripts\") pod \"e3716c1c-ce34-48ae-952f-cf16c10baf9b\" (UID: \"e3716c1c-ce34-48ae-952f-cf16c10baf9b\") " Mar 07 15:17:35 crc kubenswrapper[4943]: I0307 15:17:35.621106 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fxzmv\" (UniqueName: \"kubernetes.io/projected/e3716c1c-ce34-48ae-952f-cf16c10baf9b-kube-api-access-fxzmv\") pod \"e3716c1c-ce34-48ae-952f-cf16c10baf9b\" (UID: \"e3716c1c-ce34-48ae-952f-cf16c10baf9b\") " Mar 07 15:17:35 crc kubenswrapper[4943]: I0307 15:17:35.621156 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3716c1c-ce34-48ae-952f-cf16c10baf9b-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "e3716c1c-ce34-48ae-952f-cf16c10baf9b" (UID: "e3716c1c-ce34-48ae-952f-cf16c10baf9b"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:17:35 crc kubenswrapper[4943]: I0307 15:17:35.621253 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e3716c1c-ce34-48ae-952f-cf16c10baf9b-dispersionconf\") pod \"e3716c1c-ce34-48ae-952f-cf16c10baf9b\" (UID: \"e3716c1c-ce34-48ae-952f-cf16c10baf9b\") " Mar 07 15:17:35 crc kubenswrapper[4943]: I0307 15:17:35.621328 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3716c1c-ce34-48ae-952f-cf16c10baf9b-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "e3716c1c-ce34-48ae-952f-cf16c10baf9b" (UID: "e3716c1c-ce34-48ae-952f-cf16c10baf9b"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:17:35 crc kubenswrapper[4943]: I0307 15:17:35.621793 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e3716c1c-ce34-48ae-952f-cf16c10baf9b-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:35 crc kubenswrapper[4943]: I0307 15:17:35.621817 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e3716c1c-ce34-48ae-952f-cf16c10baf9b-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:35 crc kubenswrapper[4943]: I0307 15:17:35.626034 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3716c1c-ce34-48ae-952f-cf16c10baf9b-kube-api-access-fxzmv" (OuterVolumeSpecName: "kube-api-access-fxzmv") pod "e3716c1c-ce34-48ae-952f-cf16c10baf9b" (UID: "e3716c1c-ce34-48ae-952f-cf16c10baf9b"). InnerVolumeSpecName "kube-api-access-fxzmv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:17:35 crc kubenswrapper[4943]: I0307 15:17:35.647537 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3716c1c-ce34-48ae-952f-cf16c10baf9b-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "e3716c1c-ce34-48ae-952f-cf16c10baf9b" (UID: "e3716c1c-ce34-48ae-952f-cf16c10baf9b"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:17:35 crc kubenswrapper[4943]: I0307 15:17:35.655295 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3716c1c-ce34-48ae-952f-cf16c10baf9b-scripts" (OuterVolumeSpecName: "scripts") pod "e3716c1c-ce34-48ae-952f-cf16c10baf9b" (UID: "e3716c1c-ce34-48ae-952f-cf16c10baf9b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:17:35 crc kubenswrapper[4943]: I0307 15:17:35.664340 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3716c1c-ce34-48ae-952f-cf16c10baf9b-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "e3716c1c-ce34-48ae-952f-cf16c10baf9b" (UID: "e3716c1c-ce34-48ae-952f-cf16c10baf9b"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:17:35 crc kubenswrapper[4943]: I0307 15:17:35.723806 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e3716c1c-ce34-48ae-952f-cf16c10baf9b-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:35 crc kubenswrapper[4943]: I0307 15:17:35.723894 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e3716c1c-ce34-48ae-952f-cf16c10baf9b-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:35 crc kubenswrapper[4943]: I0307 15:17:35.723913 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e3716c1c-ce34-48ae-952f-cf16c10baf9b-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:35 crc kubenswrapper[4943]: I0307 15:17:35.723954 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fxzmv\" (UniqueName: \"kubernetes.io/projected/e3716c1c-ce34-48ae-952f-cf16c10baf9b-kube-api-access-fxzmv\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:36 crc kubenswrapper[4943]: I0307 15:17:36.191429 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a5ea3a2971232efb83738b6041356610b37faaaedab261b2beefa11924d8e73d" Mar 07 15:17:36 crc kubenswrapper[4943]: I0307 15:17:36.191463 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-t7564" Mar 07 15:17:36 crc kubenswrapper[4943]: I0307 15:17:36.719152 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-ddm9z"] Mar 07 15:17:36 crc kubenswrapper[4943]: E0307 15:17:36.719606 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3716c1c-ce34-48ae-952f-cf16c10baf9b" containerName="swift-ring-rebalance" Mar 07 15:17:36 crc kubenswrapper[4943]: I0307 15:17:36.719629 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3716c1c-ce34-48ae-952f-cf16c10baf9b" containerName="swift-ring-rebalance" Mar 07 15:17:36 crc kubenswrapper[4943]: I0307 15:17:36.719882 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3716c1c-ce34-48ae-952f-cf16c10baf9b" containerName="swift-ring-rebalance" Mar 07 15:17:36 crc kubenswrapper[4943]: I0307 15:17:36.720678 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ddm9z" Mar 07 15:17:36 crc kubenswrapper[4943]: I0307 15:17:36.729386 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:17:36 crc kubenswrapper[4943]: I0307 15:17:36.729727 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:17:36 crc kubenswrapper[4943]: I0307 15:17:36.741864 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-ddm9z"] Mar 07 15:17:36 crc kubenswrapper[4943]: I0307 15:17:36.790771 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3716c1c-ce34-48ae-952f-cf16c10baf9b" path="/var/lib/kubelet/pods/e3716c1c-ce34-48ae-952f-cf16c10baf9b/volumes" Mar 07 15:17:36 crc kubenswrapper[4943]: I0307 15:17:36.841692 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f5b9c76c-3d6c-41cd-aa1f-73a6024051a4-dispersionconf\") pod \"swift-ring-rebalance-debug-ddm9z\" (UID: \"f5b9c76c-3d6c-41cd-aa1f-73a6024051a4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ddm9z" Mar 07 15:17:36 crc kubenswrapper[4943]: I0307 15:17:36.841784 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f5b9c76c-3d6c-41cd-aa1f-73a6024051a4-scripts\") pod \"swift-ring-rebalance-debug-ddm9z\" (UID: \"f5b9c76c-3d6c-41cd-aa1f-73a6024051a4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ddm9z" Mar 07 15:17:36 crc kubenswrapper[4943]: I0307 15:17:36.841814 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f5b9c76c-3d6c-41cd-aa1f-73a6024051a4-ring-data-devices\") pod \"swift-ring-rebalance-debug-ddm9z\" (UID: \"f5b9c76c-3d6c-41cd-aa1f-73a6024051a4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ddm9z" Mar 07 15:17:36 crc kubenswrapper[4943]: I0307 15:17:36.842093 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78pjw\" (UniqueName: \"kubernetes.io/projected/f5b9c76c-3d6c-41cd-aa1f-73a6024051a4-kube-api-access-78pjw\") pod \"swift-ring-rebalance-debug-ddm9z\" (UID: \"f5b9c76c-3d6c-41cd-aa1f-73a6024051a4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ddm9z" Mar 07 15:17:36 crc kubenswrapper[4943]: I0307 15:17:36.842147 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f5b9c76c-3d6c-41cd-aa1f-73a6024051a4-swiftconf\") pod \"swift-ring-rebalance-debug-ddm9z\" (UID: \"f5b9c76c-3d6c-41cd-aa1f-73a6024051a4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ddm9z" Mar 07 15:17:36 crc kubenswrapper[4943]: I0307 15:17:36.842245 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f5b9c76c-3d6c-41cd-aa1f-73a6024051a4-etc-swift\") pod \"swift-ring-rebalance-debug-ddm9z\" (UID: \"f5b9c76c-3d6c-41cd-aa1f-73a6024051a4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ddm9z" Mar 07 15:17:36 crc kubenswrapper[4943]: I0307 15:17:36.943836 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f5b9c76c-3d6c-41cd-aa1f-73a6024051a4-dispersionconf\") pod \"swift-ring-rebalance-debug-ddm9z\" (UID: \"f5b9c76c-3d6c-41cd-aa1f-73a6024051a4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ddm9z" Mar 07 15:17:36 crc kubenswrapper[4943]: I0307 15:17:36.944002 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f5b9c76c-3d6c-41cd-aa1f-73a6024051a4-scripts\") pod \"swift-ring-rebalance-debug-ddm9z\" (UID: \"f5b9c76c-3d6c-41cd-aa1f-73a6024051a4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ddm9z" Mar 07 15:17:36 crc kubenswrapper[4943]: I0307 15:17:36.944052 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f5b9c76c-3d6c-41cd-aa1f-73a6024051a4-ring-data-devices\") pod \"swift-ring-rebalance-debug-ddm9z\" (UID: \"f5b9c76c-3d6c-41cd-aa1f-73a6024051a4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ddm9z" Mar 07 15:17:36 crc kubenswrapper[4943]: I0307 15:17:36.944132 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78pjw\" (UniqueName: \"kubernetes.io/projected/f5b9c76c-3d6c-41cd-aa1f-73a6024051a4-kube-api-access-78pjw\") pod \"swift-ring-rebalance-debug-ddm9z\" (UID: \"f5b9c76c-3d6c-41cd-aa1f-73a6024051a4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ddm9z" Mar 07 15:17:36 crc kubenswrapper[4943]: I0307 15:17:36.944167 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f5b9c76c-3d6c-41cd-aa1f-73a6024051a4-swiftconf\") pod \"swift-ring-rebalance-debug-ddm9z\" (UID: \"f5b9c76c-3d6c-41cd-aa1f-73a6024051a4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ddm9z" Mar 07 15:17:36 crc kubenswrapper[4943]: I0307 15:17:36.944266 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f5b9c76c-3d6c-41cd-aa1f-73a6024051a4-etc-swift\") pod \"swift-ring-rebalance-debug-ddm9z\" (UID: \"f5b9c76c-3d6c-41cd-aa1f-73a6024051a4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ddm9z" Mar 07 15:17:36 crc kubenswrapper[4943]: I0307 15:17:36.945719 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f5b9c76c-3d6c-41cd-aa1f-73a6024051a4-etc-swift\") pod \"swift-ring-rebalance-debug-ddm9z\" (UID: \"f5b9c76c-3d6c-41cd-aa1f-73a6024051a4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ddm9z" Mar 07 15:17:36 crc kubenswrapper[4943]: I0307 15:17:36.945792 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f5b9c76c-3d6c-41cd-aa1f-73a6024051a4-ring-data-devices\") pod \"swift-ring-rebalance-debug-ddm9z\" (UID: \"f5b9c76c-3d6c-41cd-aa1f-73a6024051a4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ddm9z" Mar 07 15:17:36 crc kubenswrapper[4943]: I0307 15:17:36.946815 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f5b9c76c-3d6c-41cd-aa1f-73a6024051a4-scripts\") pod \"swift-ring-rebalance-debug-ddm9z\" (UID: \"f5b9c76c-3d6c-41cd-aa1f-73a6024051a4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ddm9z" Mar 07 15:17:36 crc kubenswrapper[4943]: I0307 15:17:36.952708 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f5b9c76c-3d6c-41cd-aa1f-73a6024051a4-dispersionconf\") pod \"swift-ring-rebalance-debug-ddm9z\" (UID: \"f5b9c76c-3d6c-41cd-aa1f-73a6024051a4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ddm9z" Mar 07 15:17:36 crc kubenswrapper[4943]: I0307 15:17:36.952717 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f5b9c76c-3d6c-41cd-aa1f-73a6024051a4-swiftconf\") pod \"swift-ring-rebalance-debug-ddm9z\" (UID: \"f5b9c76c-3d6c-41cd-aa1f-73a6024051a4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ddm9z" Mar 07 15:17:36 crc kubenswrapper[4943]: I0307 15:17:36.969868 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78pjw\" (UniqueName: \"kubernetes.io/projected/f5b9c76c-3d6c-41cd-aa1f-73a6024051a4-kube-api-access-78pjw\") pod \"swift-ring-rebalance-debug-ddm9z\" (UID: \"f5b9c76c-3d6c-41cd-aa1f-73a6024051a4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ddm9z" Mar 07 15:17:37 crc kubenswrapper[4943]: I0307 15:17:37.080525 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ddm9z" Mar 07 15:17:37 crc kubenswrapper[4943]: I0307 15:17:37.539212 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-ddm9z"] Mar 07 15:17:37 crc kubenswrapper[4943]: W0307 15:17:37.545254 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf5b9c76c_3d6c_41cd_aa1f_73a6024051a4.slice/crio-47c5e47beb32d6ac2e6cdb5992fefe3cef70ddf9b96c7158d219a424b1a99740 WatchSource:0}: Error finding container 47c5e47beb32d6ac2e6cdb5992fefe3cef70ddf9b96c7158d219a424b1a99740: Status 404 returned error can't find the container with id 47c5e47beb32d6ac2e6cdb5992fefe3cef70ddf9b96c7158d219a424b1a99740 Mar 07 15:17:38 crc kubenswrapper[4943]: I0307 15:17:38.218960 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ddm9z" event={"ID":"f5b9c76c-3d6c-41cd-aa1f-73a6024051a4","Type":"ContainerStarted","Data":"686d607903b0b70fa2d01d64ef014be2fdca50ec8450aabe5b3cd97228b3570e"} Mar 07 15:17:38 crc kubenswrapper[4943]: I0307 15:17:38.219451 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ddm9z" event={"ID":"f5b9c76c-3d6c-41cd-aa1f-73a6024051a4","Type":"ContainerStarted","Data":"47c5e47beb32d6ac2e6cdb5992fefe3cef70ddf9b96c7158d219a424b1a99740"} Mar 07 15:17:38 crc kubenswrapper[4943]: I0307 15:17:38.247776 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ddm9z" podStartSLOduration=2.247751475 podStartE2EDuration="2.247751475s" podCreationTimestamp="2026-03-07 15:17:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:17:38.2397823 +0000 UTC m=+2300.191918828" watchObservedRunningTime="2026-03-07 15:17:38.247751475 +0000 UTC m=+2300.199887993" Mar 07 15:17:38 crc kubenswrapper[4943]: I0307 15:17:38.768403 4943 scope.go:117] "RemoveContainer" containerID="448308ad6e34ac922d66b3c90e66d0927407a94dd28633ec350502d69e4ca8aa" Mar 07 15:17:38 crc kubenswrapper[4943]: E0307 15:17:38.768698 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cpdmm_openshift-machine-config-operator(4c1f18ea-e557-41fc-95dd-ce4e1722a61e)\"" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" Mar 07 15:17:39 crc kubenswrapper[4943]: I0307 15:17:39.233401 4943 generic.go:334] "Generic (PLEG): container finished" podID="f5b9c76c-3d6c-41cd-aa1f-73a6024051a4" containerID="686d607903b0b70fa2d01d64ef014be2fdca50ec8450aabe5b3cd97228b3570e" exitCode=0 Mar 07 15:17:39 crc kubenswrapper[4943]: I0307 15:17:39.233469 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ddm9z" event={"ID":"f5b9c76c-3d6c-41cd-aa1f-73a6024051a4","Type":"ContainerDied","Data":"686d607903b0b70fa2d01d64ef014be2fdca50ec8450aabe5b3cd97228b3570e"} Mar 07 15:17:40 crc kubenswrapper[4943]: I0307 15:17:40.639507 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ddm9z" Mar 07 15:17:40 crc kubenswrapper[4943]: I0307 15:17:40.682435 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-ddm9z"] Mar 07 15:17:40 crc kubenswrapper[4943]: I0307 15:17:40.687358 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-ddm9z"] Mar 07 15:17:40 crc kubenswrapper[4943]: I0307 15:17:40.708139 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f5b9c76c-3d6c-41cd-aa1f-73a6024051a4-etc-swift\") pod \"f5b9c76c-3d6c-41cd-aa1f-73a6024051a4\" (UID: \"f5b9c76c-3d6c-41cd-aa1f-73a6024051a4\") " Mar 07 15:17:40 crc kubenswrapper[4943]: I0307 15:17:40.708248 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f5b9c76c-3d6c-41cd-aa1f-73a6024051a4-swiftconf\") pod \"f5b9c76c-3d6c-41cd-aa1f-73a6024051a4\" (UID: \"f5b9c76c-3d6c-41cd-aa1f-73a6024051a4\") " Mar 07 15:17:40 crc kubenswrapper[4943]: I0307 15:17:40.708449 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-78pjw\" (UniqueName: \"kubernetes.io/projected/f5b9c76c-3d6c-41cd-aa1f-73a6024051a4-kube-api-access-78pjw\") pod \"f5b9c76c-3d6c-41cd-aa1f-73a6024051a4\" (UID: \"f5b9c76c-3d6c-41cd-aa1f-73a6024051a4\") " Mar 07 15:17:40 crc kubenswrapper[4943]: I0307 15:17:40.708485 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f5b9c76c-3d6c-41cd-aa1f-73a6024051a4-scripts\") pod \"f5b9c76c-3d6c-41cd-aa1f-73a6024051a4\" (UID: \"f5b9c76c-3d6c-41cd-aa1f-73a6024051a4\") " Mar 07 15:17:40 crc kubenswrapper[4943]: I0307 15:17:40.708520 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f5b9c76c-3d6c-41cd-aa1f-73a6024051a4-ring-data-devices\") pod \"f5b9c76c-3d6c-41cd-aa1f-73a6024051a4\" (UID: \"f5b9c76c-3d6c-41cd-aa1f-73a6024051a4\") " Mar 07 15:17:40 crc kubenswrapper[4943]: I0307 15:17:40.708594 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f5b9c76c-3d6c-41cd-aa1f-73a6024051a4-dispersionconf\") pod \"f5b9c76c-3d6c-41cd-aa1f-73a6024051a4\" (UID: \"f5b9c76c-3d6c-41cd-aa1f-73a6024051a4\") " Mar 07 15:17:40 crc kubenswrapper[4943]: I0307 15:17:40.708874 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5b9c76c-3d6c-41cd-aa1f-73a6024051a4-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "f5b9c76c-3d6c-41cd-aa1f-73a6024051a4" (UID: "f5b9c76c-3d6c-41cd-aa1f-73a6024051a4"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:17:40 crc kubenswrapper[4943]: I0307 15:17:40.709843 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5b9c76c-3d6c-41cd-aa1f-73a6024051a4-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "f5b9c76c-3d6c-41cd-aa1f-73a6024051a4" (UID: "f5b9c76c-3d6c-41cd-aa1f-73a6024051a4"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:17:40 crc kubenswrapper[4943]: I0307 15:17:40.714515 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5b9c76c-3d6c-41cd-aa1f-73a6024051a4-kube-api-access-78pjw" (OuterVolumeSpecName: "kube-api-access-78pjw") pod "f5b9c76c-3d6c-41cd-aa1f-73a6024051a4" (UID: "f5b9c76c-3d6c-41cd-aa1f-73a6024051a4"). InnerVolumeSpecName "kube-api-access-78pjw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:17:40 crc kubenswrapper[4943]: I0307 15:17:40.732308 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5b9c76c-3d6c-41cd-aa1f-73a6024051a4-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "f5b9c76c-3d6c-41cd-aa1f-73a6024051a4" (UID: "f5b9c76c-3d6c-41cd-aa1f-73a6024051a4"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:17:40 crc kubenswrapper[4943]: I0307 15:17:40.742100 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5b9c76c-3d6c-41cd-aa1f-73a6024051a4-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "f5b9c76c-3d6c-41cd-aa1f-73a6024051a4" (UID: "f5b9c76c-3d6c-41cd-aa1f-73a6024051a4"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:17:40 crc kubenswrapper[4943]: I0307 15:17:40.747378 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5b9c76c-3d6c-41cd-aa1f-73a6024051a4-scripts" (OuterVolumeSpecName: "scripts") pod "f5b9c76c-3d6c-41cd-aa1f-73a6024051a4" (UID: "f5b9c76c-3d6c-41cd-aa1f-73a6024051a4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:17:40 crc kubenswrapper[4943]: I0307 15:17:40.767725 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5b9c76c-3d6c-41cd-aa1f-73a6024051a4" path="/var/lib/kubelet/pods/f5b9c76c-3d6c-41cd-aa1f-73a6024051a4/volumes" Mar 07 15:17:40 crc kubenswrapper[4943]: I0307 15:17:40.810138 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f5b9c76c-3d6c-41cd-aa1f-73a6024051a4-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:40 crc kubenswrapper[4943]: I0307 15:17:40.810171 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-78pjw\" (UniqueName: \"kubernetes.io/projected/f5b9c76c-3d6c-41cd-aa1f-73a6024051a4-kube-api-access-78pjw\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:40 crc kubenswrapper[4943]: I0307 15:17:40.810183 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f5b9c76c-3d6c-41cd-aa1f-73a6024051a4-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:40 crc kubenswrapper[4943]: I0307 15:17:40.810195 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f5b9c76c-3d6c-41cd-aa1f-73a6024051a4-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:40 crc kubenswrapper[4943]: I0307 15:17:40.810207 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f5b9c76c-3d6c-41cd-aa1f-73a6024051a4-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:40 crc kubenswrapper[4943]: I0307 15:17:40.810218 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f5b9c76c-3d6c-41cd-aa1f-73a6024051a4-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:41 crc kubenswrapper[4943]: I0307 15:17:41.259428 4943 scope.go:117] "RemoveContainer" containerID="686d607903b0b70fa2d01d64ef014be2fdca50ec8450aabe5b3cd97228b3570e" Mar 07 15:17:41 crc kubenswrapper[4943]: I0307 15:17:41.259495 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ddm9z" Mar 07 15:17:41 crc kubenswrapper[4943]: I0307 15:17:41.881205 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7sdpk"] Mar 07 15:17:41 crc kubenswrapper[4943]: E0307 15:17:41.882400 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5b9c76c-3d6c-41cd-aa1f-73a6024051a4" containerName="swift-ring-rebalance" Mar 07 15:17:41 crc kubenswrapper[4943]: I0307 15:17:41.882422 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5b9c76c-3d6c-41cd-aa1f-73a6024051a4" containerName="swift-ring-rebalance" Mar 07 15:17:41 crc kubenswrapper[4943]: I0307 15:17:41.882650 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5b9c76c-3d6c-41cd-aa1f-73a6024051a4" containerName="swift-ring-rebalance" Mar 07 15:17:41 crc kubenswrapper[4943]: I0307 15:17:41.883457 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7sdpk" Mar 07 15:17:41 crc kubenswrapper[4943]: I0307 15:17:41.887523 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:17:41 crc kubenswrapper[4943]: I0307 15:17:41.887690 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:17:41 crc kubenswrapper[4943]: I0307 15:17:41.895319 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7sdpk"] Mar 07 15:17:41 crc kubenswrapper[4943]: I0307 15:17:41.927144 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1f579f99-a0a1-42c4-bc99-b64bdbe3dd59-etc-swift\") pod \"swift-ring-rebalance-debug-7sdpk\" (UID: \"1f579f99-a0a1-42c4-bc99-b64bdbe3dd59\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7sdpk" Mar 07 15:17:41 crc kubenswrapper[4943]: I0307 15:17:41.927211 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1f579f99-a0a1-42c4-bc99-b64bdbe3dd59-ring-data-devices\") pod \"swift-ring-rebalance-debug-7sdpk\" (UID: \"1f579f99-a0a1-42c4-bc99-b64bdbe3dd59\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7sdpk" Mar 07 15:17:41 crc kubenswrapper[4943]: I0307 15:17:41.927287 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xnhr\" (UniqueName: \"kubernetes.io/projected/1f579f99-a0a1-42c4-bc99-b64bdbe3dd59-kube-api-access-2xnhr\") pod \"swift-ring-rebalance-debug-7sdpk\" (UID: \"1f579f99-a0a1-42c4-bc99-b64bdbe3dd59\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7sdpk" Mar 07 15:17:41 crc kubenswrapper[4943]: I0307 15:17:41.927348 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1f579f99-a0a1-42c4-bc99-b64bdbe3dd59-scripts\") pod \"swift-ring-rebalance-debug-7sdpk\" (UID: \"1f579f99-a0a1-42c4-bc99-b64bdbe3dd59\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7sdpk" Mar 07 15:17:41 crc kubenswrapper[4943]: I0307 15:17:41.927415 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1f579f99-a0a1-42c4-bc99-b64bdbe3dd59-swiftconf\") pod \"swift-ring-rebalance-debug-7sdpk\" (UID: \"1f579f99-a0a1-42c4-bc99-b64bdbe3dd59\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7sdpk" Mar 07 15:17:41 crc kubenswrapper[4943]: I0307 15:17:41.927447 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1f579f99-a0a1-42c4-bc99-b64bdbe3dd59-dispersionconf\") pod \"swift-ring-rebalance-debug-7sdpk\" (UID: \"1f579f99-a0a1-42c4-bc99-b64bdbe3dd59\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7sdpk" Mar 07 15:17:42 crc kubenswrapper[4943]: I0307 15:17:42.028848 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1f579f99-a0a1-42c4-bc99-b64bdbe3dd59-swiftconf\") pod \"swift-ring-rebalance-debug-7sdpk\" (UID: \"1f579f99-a0a1-42c4-bc99-b64bdbe3dd59\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7sdpk" Mar 07 15:17:42 crc kubenswrapper[4943]: I0307 15:17:42.028898 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1f579f99-a0a1-42c4-bc99-b64bdbe3dd59-dispersionconf\") pod \"swift-ring-rebalance-debug-7sdpk\" (UID: \"1f579f99-a0a1-42c4-bc99-b64bdbe3dd59\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7sdpk" Mar 07 15:17:42 crc kubenswrapper[4943]: I0307 15:17:42.029021 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1f579f99-a0a1-42c4-bc99-b64bdbe3dd59-etc-swift\") pod \"swift-ring-rebalance-debug-7sdpk\" (UID: \"1f579f99-a0a1-42c4-bc99-b64bdbe3dd59\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7sdpk" Mar 07 15:17:42 crc kubenswrapper[4943]: I0307 15:17:42.029065 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1f579f99-a0a1-42c4-bc99-b64bdbe3dd59-ring-data-devices\") pod \"swift-ring-rebalance-debug-7sdpk\" (UID: \"1f579f99-a0a1-42c4-bc99-b64bdbe3dd59\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7sdpk" Mar 07 15:17:42 crc kubenswrapper[4943]: I0307 15:17:42.029134 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xnhr\" (UniqueName: \"kubernetes.io/projected/1f579f99-a0a1-42c4-bc99-b64bdbe3dd59-kube-api-access-2xnhr\") pod \"swift-ring-rebalance-debug-7sdpk\" (UID: \"1f579f99-a0a1-42c4-bc99-b64bdbe3dd59\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7sdpk" Mar 07 15:17:42 crc kubenswrapper[4943]: I0307 15:17:42.029333 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1f579f99-a0a1-42c4-bc99-b64bdbe3dd59-scripts\") pod \"swift-ring-rebalance-debug-7sdpk\" (UID: \"1f579f99-a0a1-42c4-bc99-b64bdbe3dd59\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7sdpk" Mar 07 15:17:42 crc kubenswrapper[4943]: I0307 15:17:42.030391 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1f579f99-a0a1-42c4-bc99-b64bdbe3dd59-etc-swift\") pod \"swift-ring-rebalance-debug-7sdpk\" (UID: \"1f579f99-a0a1-42c4-bc99-b64bdbe3dd59\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7sdpk" Mar 07 15:17:42 crc kubenswrapper[4943]: I0307 15:17:42.030640 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1f579f99-a0a1-42c4-bc99-b64bdbe3dd59-ring-data-devices\") pod \"swift-ring-rebalance-debug-7sdpk\" (UID: \"1f579f99-a0a1-42c4-bc99-b64bdbe3dd59\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7sdpk" Mar 07 15:17:42 crc kubenswrapper[4943]: I0307 15:17:42.031250 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1f579f99-a0a1-42c4-bc99-b64bdbe3dd59-scripts\") pod \"swift-ring-rebalance-debug-7sdpk\" (UID: \"1f579f99-a0a1-42c4-bc99-b64bdbe3dd59\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7sdpk" Mar 07 15:17:42 crc kubenswrapper[4943]: I0307 15:17:42.038767 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1f579f99-a0a1-42c4-bc99-b64bdbe3dd59-swiftconf\") pod \"swift-ring-rebalance-debug-7sdpk\" (UID: \"1f579f99-a0a1-42c4-bc99-b64bdbe3dd59\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7sdpk" Mar 07 15:17:42 crc kubenswrapper[4943]: I0307 15:17:42.049596 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1f579f99-a0a1-42c4-bc99-b64bdbe3dd59-dispersionconf\") pod \"swift-ring-rebalance-debug-7sdpk\" (UID: \"1f579f99-a0a1-42c4-bc99-b64bdbe3dd59\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7sdpk" Mar 07 15:17:42 crc kubenswrapper[4943]: I0307 15:17:42.068238 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xnhr\" (UniqueName: \"kubernetes.io/projected/1f579f99-a0a1-42c4-bc99-b64bdbe3dd59-kube-api-access-2xnhr\") pod \"swift-ring-rebalance-debug-7sdpk\" (UID: \"1f579f99-a0a1-42c4-bc99-b64bdbe3dd59\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7sdpk" Mar 07 15:17:42 crc kubenswrapper[4943]: I0307 15:17:42.213522 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7sdpk" Mar 07 15:17:42 crc kubenswrapper[4943]: I0307 15:17:42.748562 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7sdpk"] Mar 07 15:17:43 crc kubenswrapper[4943]: I0307 15:17:43.293606 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7sdpk" event={"ID":"1f579f99-a0a1-42c4-bc99-b64bdbe3dd59","Type":"ContainerStarted","Data":"dd79e8e409342a91fdea9c5771d4f3e26221f83877ff8eb377eda62cf2a22a49"} Mar 07 15:17:43 crc kubenswrapper[4943]: I0307 15:17:43.293939 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7sdpk" event={"ID":"1f579f99-a0a1-42c4-bc99-b64bdbe3dd59","Type":"ContainerStarted","Data":"272e65f4853f4288e82a83334f1b99e63a2fc0c99b09b756386681fae2f135b8"} Mar 07 15:17:45 crc kubenswrapper[4943]: I0307 15:17:45.319495 4943 generic.go:334] "Generic (PLEG): container finished" podID="1f579f99-a0a1-42c4-bc99-b64bdbe3dd59" containerID="dd79e8e409342a91fdea9c5771d4f3e26221f83877ff8eb377eda62cf2a22a49" exitCode=0 Mar 07 15:17:45 crc kubenswrapper[4943]: I0307 15:17:45.319627 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7sdpk" event={"ID":"1f579f99-a0a1-42c4-bc99-b64bdbe3dd59","Type":"ContainerDied","Data":"dd79e8e409342a91fdea9c5771d4f3e26221f83877ff8eb377eda62cf2a22a49"} Mar 07 15:17:46 crc kubenswrapper[4943]: I0307 15:17:46.700497 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7sdpk" Mar 07 15:17:46 crc kubenswrapper[4943]: I0307 15:17:46.740489 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7sdpk"] Mar 07 15:17:46 crc kubenswrapper[4943]: I0307 15:17:46.749107 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7sdpk"] Mar 07 15:17:46 crc kubenswrapper[4943]: I0307 15:17:46.807099 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1f579f99-a0a1-42c4-bc99-b64bdbe3dd59-etc-swift\") pod \"1f579f99-a0a1-42c4-bc99-b64bdbe3dd59\" (UID: \"1f579f99-a0a1-42c4-bc99-b64bdbe3dd59\") " Mar 07 15:17:46 crc kubenswrapper[4943]: I0307 15:17:46.807161 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1f579f99-a0a1-42c4-bc99-b64bdbe3dd59-dispersionconf\") pod \"1f579f99-a0a1-42c4-bc99-b64bdbe3dd59\" (UID: \"1f579f99-a0a1-42c4-bc99-b64bdbe3dd59\") " Mar 07 15:17:46 crc kubenswrapper[4943]: I0307 15:17:46.807200 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1f579f99-a0a1-42c4-bc99-b64bdbe3dd59-swiftconf\") pod \"1f579f99-a0a1-42c4-bc99-b64bdbe3dd59\" (UID: \"1f579f99-a0a1-42c4-bc99-b64bdbe3dd59\") " Mar 07 15:17:46 crc kubenswrapper[4943]: I0307 15:17:46.807274 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1f579f99-a0a1-42c4-bc99-b64bdbe3dd59-scripts\") pod \"1f579f99-a0a1-42c4-bc99-b64bdbe3dd59\" (UID: \"1f579f99-a0a1-42c4-bc99-b64bdbe3dd59\") " Mar 07 15:17:46 crc kubenswrapper[4943]: I0307 15:17:46.807310 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1f579f99-a0a1-42c4-bc99-b64bdbe3dd59-ring-data-devices\") pod \"1f579f99-a0a1-42c4-bc99-b64bdbe3dd59\" (UID: \"1f579f99-a0a1-42c4-bc99-b64bdbe3dd59\") " Mar 07 15:17:46 crc kubenswrapper[4943]: I0307 15:17:46.807340 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2xnhr\" (UniqueName: \"kubernetes.io/projected/1f579f99-a0a1-42c4-bc99-b64bdbe3dd59-kube-api-access-2xnhr\") pod \"1f579f99-a0a1-42c4-bc99-b64bdbe3dd59\" (UID: \"1f579f99-a0a1-42c4-bc99-b64bdbe3dd59\") " Mar 07 15:17:46 crc kubenswrapper[4943]: I0307 15:17:46.807972 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f579f99-a0a1-42c4-bc99-b64bdbe3dd59-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "1f579f99-a0a1-42c4-bc99-b64bdbe3dd59" (UID: "1f579f99-a0a1-42c4-bc99-b64bdbe3dd59"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:17:46 crc kubenswrapper[4943]: I0307 15:17:46.808535 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f579f99-a0a1-42c4-bc99-b64bdbe3dd59-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "1f579f99-a0a1-42c4-bc99-b64bdbe3dd59" (UID: "1f579f99-a0a1-42c4-bc99-b64bdbe3dd59"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:17:46 crc kubenswrapper[4943]: I0307 15:17:46.812598 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f579f99-a0a1-42c4-bc99-b64bdbe3dd59-kube-api-access-2xnhr" (OuterVolumeSpecName: "kube-api-access-2xnhr") pod "1f579f99-a0a1-42c4-bc99-b64bdbe3dd59" (UID: "1f579f99-a0a1-42c4-bc99-b64bdbe3dd59"). InnerVolumeSpecName "kube-api-access-2xnhr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:17:46 crc kubenswrapper[4943]: I0307 15:17:46.827779 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f579f99-a0a1-42c4-bc99-b64bdbe3dd59-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "1f579f99-a0a1-42c4-bc99-b64bdbe3dd59" (UID: "1f579f99-a0a1-42c4-bc99-b64bdbe3dd59"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:17:46 crc kubenswrapper[4943]: I0307 15:17:46.832828 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f579f99-a0a1-42c4-bc99-b64bdbe3dd59-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "1f579f99-a0a1-42c4-bc99-b64bdbe3dd59" (UID: "1f579f99-a0a1-42c4-bc99-b64bdbe3dd59"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:17:46 crc kubenswrapper[4943]: I0307 15:17:46.843260 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f579f99-a0a1-42c4-bc99-b64bdbe3dd59-scripts" (OuterVolumeSpecName: "scripts") pod "1f579f99-a0a1-42c4-bc99-b64bdbe3dd59" (UID: "1f579f99-a0a1-42c4-bc99-b64bdbe3dd59"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:17:46 crc kubenswrapper[4943]: I0307 15:17:46.909419 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1f579f99-a0a1-42c4-bc99-b64bdbe3dd59-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:46 crc kubenswrapper[4943]: I0307 15:17:46.909453 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1f579f99-a0a1-42c4-bc99-b64bdbe3dd59-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:46 crc kubenswrapper[4943]: I0307 15:17:46.909468 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2xnhr\" (UniqueName: \"kubernetes.io/projected/1f579f99-a0a1-42c4-bc99-b64bdbe3dd59-kube-api-access-2xnhr\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:46 crc kubenswrapper[4943]: I0307 15:17:46.909481 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1f579f99-a0a1-42c4-bc99-b64bdbe3dd59-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:46 crc kubenswrapper[4943]: I0307 15:17:46.909491 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1f579f99-a0a1-42c4-bc99-b64bdbe3dd59-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:46 crc kubenswrapper[4943]: I0307 15:17:46.909511 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1f579f99-a0a1-42c4-bc99-b64bdbe3dd59-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:47 crc kubenswrapper[4943]: I0307 15:17:47.365006 4943 scope.go:117] "RemoveContainer" containerID="dd79e8e409342a91fdea9c5771d4f3e26221f83877ff8eb377eda62cf2a22a49" Mar 07 15:17:47 crc kubenswrapper[4943]: I0307 15:17:47.365158 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7sdpk" Mar 07 15:17:47 crc kubenswrapper[4943]: I0307 15:17:47.902443 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-6gp7d"] Mar 07 15:17:47 crc kubenswrapper[4943]: E0307 15:17:47.903189 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f579f99-a0a1-42c4-bc99-b64bdbe3dd59" containerName="swift-ring-rebalance" Mar 07 15:17:47 crc kubenswrapper[4943]: I0307 15:17:47.903209 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f579f99-a0a1-42c4-bc99-b64bdbe3dd59" containerName="swift-ring-rebalance" Mar 07 15:17:47 crc kubenswrapper[4943]: I0307 15:17:47.903478 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f579f99-a0a1-42c4-bc99-b64bdbe3dd59" containerName="swift-ring-rebalance" Mar 07 15:17:47 crc kubenswrapper[4943]: I0307 15:17:47.904255 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6gp7d" Mar 07 15:17:47 crc kubenswrapper[4943]: I0307 15:17:47.907448 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:17:47 crc kubenswrapper[4943]: I0307 15:17:47.907968 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:17:47 crc kubenswrapper[4943]: I0307 15:17:47.922062 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-6gp7d"] Mar 07 15:17:47 crc kubenswrapper[4943]: I0307 15:17:47.926293 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/77a21bfc-ec08-4161-800e-334616554c0f-scripts\") pod \"swift-ring-rebalance-debug-6gp7d\" (UID: \"77a21bfc-ec08-4161-800e-334616554c0f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6gp7d" Mar 07 15:17:47 crc kubenswrapper[4943]: I0307 15:17:47.926380 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/77a21bfc-ec08-4161-800e-334616554c0f-ring-data-devices\") pod \"swift-ring-rebalance-debug-6gp7d\" (UID: \"77a21bfc-ec08-4161-800e-334616554c0f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6gp7d" Mar 07 15:17:47 crc kubenswrapper[4943]: I0307 15:17:47.926457 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2272\" (UniqueName: \"kubernetes.io/projected/77a21bfc-ec08-4161-800e-334616554c0f-kube-api-access-f2272\") pod \"swift-ring-rebalance-debug-6gp7d\" (UID: \"77a21bfc-ec08-4161-800e-334616554c0f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6gp7d" Mar 07 15:17:47 crc kubenswrapper[4943]: I0307 15:17:47.926516 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/77a21bfc-ec08-4161-800e-334616554c0f-dispersionconf\") pod \"swift-ring-rebalance-debug-6gp7d\" (UID: \"77a21bfc-ec08-4161-800e-334616554c0f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6gp7d" Mar 07 15:17:47 crc kubenswrapper[4943]: I0307 15:17:47.926591 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/77a21bfc-ec08-4161-800e-334616554c0f-swiftconf\") pod \"swift-ring-rebalance-debug-6gp7d\" (UID: \"77a21bfc-ec08-4161-800e-334616554c0f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6gp7d" Mar 07 15:17:47 crc kubenswrapper[4943]: I0307 15:17:47.926694 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/77a21bfc-ec08-4161-800e-334616554c0f-etc-swift\") pod \"swift-ring-rebalance-debug-6gp7d\" (UID: \"77a21bfc-ec08-4161-800e-334616554c0f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6gp7d" Mar 07 15:17:48 crc kubenswrapper[4943]: I0307 15:17:48.027811 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/77a21bfc-ec08-4161-800e-334616554c0f-etc-swift\") pod \"swift-ring-rebalance-debug-6gp7d\" (UID: \"77a21bfc-ec08-4161-800e-334616554c0f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6gp7d" Mar 07 15:17:48 crc kubenswrapper[4943]: I0307 15:17:48.028077 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/77a21bfc-ec08-4161-800e-334616554c0f-scripts\") pod \"swift-ring-rebalance-debug-6gp7d\" (UID: \"77a21bfc-ec08-4161-800e-334616554c0f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6gp7d" Mar 07 15:17:48 crc kubenswrapper[4943]: I0307 15:17:48.028116 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/77a21bfc-ec08-4161-800e-334616554c0f-ring-data-devices\") pod \"swift-ring-rebalance-debug-6gp7d\" (UID: \"77a21bfc-ec08-4161-800e-334616554c0f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6gp7d" Mar 07 15:17:48 crc kubenswrapper[4943]: I0307 15:17:48.028162 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2272\" (UniqueName: \"kubernetes.io/projected/77a21bfc-ec08-4161-800e-334616554c0f-kube-api-access-f2272\") pod \"swift-ring-rebalance-debug-6gp7d\" (UID: \"77a21bfc-ec08-4161-800e-334616554c0f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6gp7d" Mar 07 15:17:48 crc kubenswrapper[4943]: I0307 15:17:48.028204 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/77a21bfc-ec08-4161-800e-334616554c0f-dispersionconf\") pod \"swift-ring-rebalance-debug-6gp7d\" (UID: \"77a21bfc-ec08-4161-800e-334616554c0f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6gp7d" Mar 07 15:17:48 crc kubenswrapper[4943]: I0307 15:17:48.028253 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/77a21bfc-ec08-4161-800e-334616554c0f-swiftconf\") pod \"swift-ring-rebalance-debug-6gp7d\" (UID: \"77a21bfc-ec08-4161-800e-334616554c0f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6gp7d" Mar 07 15:17:48 crc kubenswrapper[4943]: I0307 15:17:48.029891 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/77a21bfc-ec08-4161-800e-334616554c0f-etc-swift\") pod \"swift-ring-rebalance-debug-6gp7d\" (UID: \"77a21bfc-ec08-4161-800e-334616554c0f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6gp7d" Mar 07 15:17:48 crc kubenswrapper[4943]: I0307 15:17:48.030430 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/77a21bfc-ec08-4161-800e-334616554c0f-ring-data-devices\") pod \"swift-ring-rebalance-debug-6gp7d\" (UID: \"77a21bfc-ec08-4161-800e-334616554c0f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6gp7d" Mar 07 15:17:48 crc kubenswrapper[4943]: I0307 15:17:48.030836 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/77a21bfc-ec08-4161-800e-334616554c0f-scripts\") pod \"swift-ring-rebalance-debug-6gp7d\" (UID: \"77a21bfc-ec08-4161-800e-334616554c0f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6gp7d" Mar 07 15:17:48 crc kubenswrapper[4943]: I0307 15:17:48.035728 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/77a21bfc-ec08-4161-800e-334616554c0f-swiftconf\") pod \"swift-ring-rebalance-debug-6gp7d\" (UID: \"77a21bfc-ec08-4161-800e-334616554c0f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6gp7d" Mar 07 15:17:48 crc kubenswrapper[4943]: I0307 15:17:48.043155 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/77a21bfc-ec08-4161-800e-334616554c0f-dispersionconf\") pod \"swift-ring-rebalance-debug-6gp7d\" (UID: \"77a21bfc-ec08-4161-800e-334616554c0f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6gp7d" Mar 07 15:17:48 crc kubenswrapper[4943]: I0307 15:17:48.059897 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2272\" (UniqueName: \"kubernetes.io/projected/77a21bfc-ec08-4161-800e-334616554c0f-kube-api-access-f2272\") pod \"swift-ring-rebalance-debug-6gp7d\" (UID: \"77a21bfc-ec08-4161-800e-334616554c0f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6gp7d" Mar 07 15:17:48 crc kubenswrapper[4943]: I0307 15:17:48.232603 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6gp7d" Mar 07 15:17:48 crc kubenswrapper[4943]: I0307 15:17:48.772546 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f579f99-a0a1-42c4-bc99-b64bdbe3dd59" path="/var/lib/kubelet/pods/1f579f99-a0a1-42c4-bc99-b64bdbe3dd59/volumes" Mar 07 15:17:48 crc kubenswrapper[4943]: W0307 15:17:48.773027 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod77a21bfc_ec08_4161_800e_334616554c0f.slice/crio-54c3b6df4cfb34d4d786e20a4cc74f512baa19a80e28af1bd35fdb993b1c55e5 WatchSource:0}: Error finding container 54c3b6df4cfb34d4d786e20a4cc74f512baa19a80e28af1bd35fdb993b1c55e5: Status 404 returned error can't find the container with id 54c3b6df4cfb34d4d786e20a4cc74f512baa19a80e28af1bd35fdb993b1c55e5 Mar 07 15:17:48 crc kubenswrapper[4943]: I0307 15:17:48.773198 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-6gp7d"] Mar 07 15:17:49 crc kubenswrapper[4943]: I0307 15:17:49.394221 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6gp7d" event={"ID":"77a21bfc-ec08-4161-800e-334616554c0f","Type":"ContainerStarted","Data":"1c4cc644d6231f8c43333e6d4105684175b14eebf2a5269c88dcd41990a21f54"} Mar 07 15:17:49 crc kubenswrapper[4943]: I0307 15:17:49.394662 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6gp7d" event={"ID":"77a21bfc-ec08-4161-800e-334616554c0f","Type":"ContainerStarted","Data":"54c3b6df4cfb34d4d786e20a4cc74f512baa19a80e28af1bd35fdb993b1c55e5"} Mar 07 15:17:49 crc kubenswrapper[4943]: I0307 15:17:49.429727 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6gp7d" podStartSLOduration=2.4297031860000002 podStartE2EDuration="2.429703186s" podCreationTimestamp="2026-03-07 15:17:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:17:49.420365468 +0000 UTC m=+2311.372502006" watchObservedRunningTime="2026-03-07 15:17:49.429703186 +0000 UTC m=+2311.381839694" Mar 07 15:17:50 crc kubenswrapper[4943]: I0307 15:17:50.408030 4943 generic.go:334] "Generic (PLEG): container finished" podID="77a21bfc-ec08-4161-800e-334616554c0f" containerID="1c4cc644d6231f8c43333e6d4105684175b14eebf2a5269c88dcd41990a21f54" exitCode=0 Mar 07 15:17:50 crc kubenswrapper[4943]: I0307 15:17:50.408088 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6gp7d" event={"ID":"77a21bfc-ec08-4161-800e-334616554c0f","Type":"ContainerDied","Data":"1c4cc644d6231f8c43333e6d4105684175b14eebf2a5269c88dcd41990a21f54"} Mar 07 15:17:51 crc kubenswrapper[4943]: I0307 15:17:51.739654 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6gp7d" Mar 07 15:17:51 crc kubenswrapper[4943]: I0307 15:17:51.792360 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-6gp7d"] Mar 07 15:17:51 crc kubenswrapper[4943]: I0307 15:17:51.798084 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-6gp7d"] Mar 07 15:17:51 crc kubenswrapper[4943]: I0307 15:17:51.912180 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/77a21bfc-ec08-4161-800e-334616554c0f-dispersionconf\") pod \"77a21bfc-ec08-4161-800e-334616554c0f\" (UID: \"77a21bfc-ec08-4161-800e-334616554c0f\") " Mar 07 15:17:51 crc kubenswrapper[4943]: I0307 15:17:51.912252 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/77a21bfc-ec08-4161-800e-334616554c0f-swiftconf\") pod \"77a21bfc-ec08-4161-800e-334616554c0f\" (UID: \"77a21bfc-ec08-4161-800e-334616554c0f\") " Mar 07 15:17:51 crc kubenswrapper[4943]: I0307 15:17:51.912319 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/77a21bfc-ec08-4161-800e-334616554c0f-etc-swift\") pod \"77a21bfc-ec08-4161-800e-334616554c0f\" (UID: \"77a21bfc-ec08-4161-800e-334616554c0f\") " Mar 07 15:17:51 crc kubenswrapper[4943]: I0307 15:17:51.912475 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/77a21bfc-ec08-4161-800e-334616554c0f-ring-data-devices\") pod \"77a21bfc-ec08-4161-800e-334616554c0f\" (UID: \"77a21bfc-ec08-4161-800e-334616554c0f\") " Mar 07 15:17:51 crc kubenswrapper[4943]: I0307 15:17:51.912523 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/77a21bfc-ec08-4161-800e-334616554c0f-scripts\") pod \"77a21bfc-ec08-4161-800e-334616554c0f\" (UID: \"77a21bfc-ec08-4161-800e-334616554c0f\") " Mar 07 15:17:51 crc kubenswrapper[4943]: I0307 15:17:51.912559 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f2272\" (UniqueName: \"kubernetes.io/projected/77a21bfc-ec08-4161-800e-334616554c0f-kube-api-access-f2272\") pod \"77a21bfc-ec08-4161-800e-334616554c0f\" (UID: \"77a21bfc-ec08-4161-800e-334616554c0f\") " Mar 07 15:17:51 crc kubenswrapper[4943]: I0307 15:17:51.914395 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77a21bfc-ec08-4161-800e-334616554c0f-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "77a21bfc-ec08-4161-800e-334616554c0f" (UID: "77a21bfc-ec08-4161-800e-334616554c0f"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:17:51 crc kubenswrapper[4943]: I0307 15:17:51.914801 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/77a21bfc-ec08-4161-800e-334616554c0f-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "77a21bfc-ec08-4161-800e-334616554c0f" (UID: "77a21bfc-ec08-4161-800e-334616554c0f"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:17:51 crc kubenswrapper[4943]: I0307 15:17:51.919132 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77a21bfc-ec08-4161-800e-334616554c0f-kube-api-access-f2272" (OuterVolumeSpecName: "kube-api-access-f2272") pod "77a21bfc-ec08-4161-800e-334616554c0f" (UID: "77a21bfc-ec08-4161-800e-334616554c0f"). InnerVolumeSpecName "kube-api-access-f2272". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:17:51 crc kubenswrapper[4943]: I0307 15:17:51.941429 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77a21bfc-ec08-4161-800e-334616554c0f-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "77a21bfc-ec08-4161-800e-334616554c0f" (UID: "77a21bfc-ec08-4161-800e-334616554c0f"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:17:51 crc kubenswrapper[4943]: I0307 15:17:51.952118 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77a21bfc-ec08-4161-800e-334616554c0f-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "77a21bfc-ec08-4161-800e-334616554c0f" (UID: "77a21bfc-ec08-4161-800e-334616554c0f"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:17:51 crc kubenswrapper[4943]: I0307 15:17:51.952380 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/77a21bfc-ec08-4161-800e-334616554c0f-scripts" (OuterVolumeSpecName: "scripts") pod "77a21bfc-ec08-4161-800e-334616554c0f" (UID: "77a21bfc-ec08-4161-800e-334616554c0f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:17:52 crc kubenswrapper[4943]: I0307 15:17:52.014642 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/77a21bfc-ec08-4161-800e-334616554c0f-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:52 crc kubenswrapper[4943]: I0307 15:17:52.014695 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/77a21bfc-ec08-4161-800e-334616554c0f-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:52 crc kubenswrapper[4943]: I0307 15:17:52.014715 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/77a21bfc-ec08-4161-800e-334616554c0f-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:52 crc kubenswrapper[4943]: I0307 15:17:52.014733 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/77a21bfc-ec08-4161-800e-334616554c0f-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:52 crc kubenswrapper[4943]: I0307 15:17:52.014751 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/77a21bfc-ec08-4161-800e-334616554c0f-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:52 crc kubenswrapper[4943]: I0307 15:17:52.014769 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f2272\" (UniqueName: \"kubernetes.io/projected/77a21bfc-ec08-4161-800e-334616554c0f-kube-api-access-f2272\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:52 crc kubenswrapper[4943]: I0307 15:17:52.435476 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="54c3b6df4cfb34d4d786e20a4cc74f512baa19a80e28af1bd35fdb993b1c55e5" Mar 07 15:17:52 crc kubenswrapper[4943]: I0307 15:17:52.435577 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6gp7d" Mar 07 15:17:52 crc kubenswrapper[4943]: I0307 15:17:52.771976 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77a21bfc-ec08-4161-800e-334616554c0f" path="/var/lib/kubelet/pods/77a21bfc-ec08-4161-800e-334616554c0f/volumes" Mar 07 15:17:52 crc kubenswrapper[4943]: I0307 15:17:52.982244 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-kkbm5"] Mar 07 15:17:52 crc kubenswrapper[4943]: E0307 15:17:52.982574 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77a21bfc-ec08-4161-800e-334616554c0f" containerName="swift-ring-rebalance" Mar 07 15:17:52 crc kubenswrapper[4943]: I0307 15:17:52.982599 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="77a21bfc-ec08-4161-800e-334616554c0f" containerName="swift-ring-rebalance" Mar 07 15:17:52 crc kubenswrapper[4943]: I0307 15:17:52.982819 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="77a21bfc-ec08-4161-800e-334616554c0f" containerName="swift-ring-rebalance" Mar 07 15:17:52 crc kubenswrapper[4943]: I0307 15:17:52.983443 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kkbm5" Mar 07 15:17:52 crc kubenswrapper[4943]: I0307 15:17:52.985841 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:17:52 crc kubenswrapper[4943]: I0307 15:17:52.985889 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:17:53 crc kubenswrapper[4943]: I0307 15:17:52.997679 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-kkbm5"] Mar 07 15:17:53 crc kubenswrapper[4943]: I0307 15:17:53.134279 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bfedd98f-1d06-49dd-94ac-3bc66c1bf603-swiftconf\") pod \"swift-ring-rebalance-debug-kkbm5\" (UID: \"bfedd98f-1d06-49dd-94ac-3bc66c1bf603\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kkbm5" Mar 07 15:17:53 crc kubenswrapper[4943]: I0307 15:17:53.134361 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bfedd98f-1d06-49dd-94ac-3bc66c1bf603-etc-swift\") pod \"swift-ring-rebalance-debug-kkbm5\" (UID: \"bfedd98f-1d06-49dd-94ac-3bc66c1bf603\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kkbm5" Mar 07 15:17:53 crc kubenswrapper[4943]: I0307 15:17:53.134403 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvpzg\" (UniqueName: \"kubernetes.io/projected/bfedd98f-1d06-49dd-94ac-3bc66c1bf603-kube-api-access-tvpzg\") pod \"swift-ring-rebalance-debug-kkbm5\" (UID: \"bfedd98f-1d06-49dd-94ac-3bc66c1bf603\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kkbm5" Mar 07 15:17:53 crc kubenswrapper[4943]: I0307 15:17:53.134464 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bfedd98f-1d06-49dd-94ac-3bc66c1bf603-ring-data-devices\") pod \"swift-ring-rebalance-debug-kkbm5\" (UID: \"bfedd98f-1d06-49dd-94ac-3bc66c1bf603\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kkbm5" Mar 07 15:17:53 crc kubenswrapper[4943]: I0307 15:17:53.134492 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bfedd98f-1d06-49dd-94ac-3bc66c1bf603-dispersionconf\") pod \"swift-ring-rebalance-debug-kkbm5\" (UID: \"bfedd98f-1d06-49dd-94ac-3bc66c1bf603\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kkbm5" Mar 07 15:17:53 crc kubenswrapper[4943]: I0307 15:17:53.134544 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bfedd98f-1d06-49dd-94ac-3bc66c1bf603-scripts\") pod \"swift-ring-rebalance-debug-kkbm5\" (UID: \"bfedd98f-1d06-49dd-94ac-3bc66c1bf603\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kkbm5" Mar 07 15:17:53 crc kubenswrapper[4943]: I0307 15:17:53.235901 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bfedd98f-1d06-49dd-94ac-3bc66c1bf603-swiftconf\") pod \"swift-ring-rebalance-debug-kkbm5\" (UID: \"bfedd98f-1d06-49dd-94ac-3bc66c1bf603\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kkbm5" Mar 07 15:17:53 crc kubenswrapper[4943]: I0307 15:17:53.235972 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bfedd98f-1d06-49dd-94ac-3bc66c1bf603-etc-swift\") pod \"swift-ring-rebalance-debug-kkbm5\" (UID: \"bfedd98f-1d06-49dd-94ac-3bc66c1bf603\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kkbm5" Mar 07 15:17:53 crc kubenswrapper[4943]: I0307 15:17:53.236009 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvpzg\" (UniqueName: \"kubernetes.io/projected/bfedd98f-1d06-49dd-94ac-3bc66c1bf603-kube-api-access-tvpzg\") pod \"swift-ring-rebalance-debug-kkbm5\" (UID: \"bfedd98f-1d06-49dd-94ac-3bc66c1bf603\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kkbm5" Mar 07 15:17:53 crc kubenswrapper[4943]: I0307 15:17:53.236041 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bfedd98f-1d06-49dd-94ac-3bc66c1bf603-ring-data-devices\") pod \"swift-ring-rebalance-debug-kkbm5\" (UID: \"bfedd98f-1d06-49dd-94ac-3bc66c1bf603\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kkbm5" Mar 07 15:17:53 crc kubenswrapper[4943]: I0307 15:17:53.236062 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bfedd98f-1d06-49dd-94ac-3bc66c1bf603-dispersionconf\") pod \"swift-ring-rebalance-debug-kkbm5\" (UID: \"bfedd98f-1d06-49dd-94ac-3bc66c1bf603\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kkbm5" Mar 07 15:17:53 crc kubenswrapper[4943]: I0307 15:17:53.236094 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bfedd98f-1d06-49dd-94ac-3bc66c1bf603-scripts\") pod \"swift-ring-rebalance-debug-kkbm5\" (UID: \"bfedd98f-1d06-49dd-94ac-3bc66c1bf603\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kkbm5" Mar 07 15:17:53 crc kubenswrapper[4943]: I0307 15:17:53.237075 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bfedd98f-1d06-49dd-94ac-3bc66c1bf603-scripts\") pod \"swift-ring-rebalance-debug-kkbm5\" (UID: \"bfedd98f-1d06-49dd-94ac-3bc66c1bf603\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kkbm5" Mar 07 15:17:53 crc kubenswrapper[4943]: I0307 15:17:53.237722 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bfedd98f-1d06-49dd-94ac-3bc66c1bf603-ring-data-devices\") pod \"swift-ring-rebalance-debug-kkbm5\" (UID: \"bfedd98f-1d06-49dd-94ac-3bc66c1bf603\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kkbm5" Mar 07 15:17:53 crc kubenswrapper[4943]: I0307 15:17:53.238244 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bfedd98f-1d06-49dd-94ac-3bc66c1bf603-etc-swift\") pod \"swift-ring-rebalance-debug-kkbm5\" (UID: \"bfedd98f-1d06-49dd-94ac-3bc66c1bf603\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kkbm5" Mar 07 15:17:53 crc kubenswrapper[4943]: I0307 15:17:53.240769 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bfedd98f-1d06-49dd-94ac-3bc66c1bf603-swiftconf\") pod \"swift-ring-rebalance-debug-kkbm5\" (UID: \"bfedd98f-1d06-49dd-94ac-3bc66c1bf603\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kkbm5" Mar 07 15:17:53 crc kubenswrapper[4943]: I0307 15:17:53.241722 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bfedd98f-1d06-49dd-94ac-3bc66c1bf603-dispersionconf\") pod \"swift-ring-rebalance-debug-kkbm5\" (UID: \"bfedd98f-1d06-49dd-94ac-3bc66c1bf603\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kkbm5" Mar 07 15:17:53 crc kubenswrapper[4943]: I0307 15:17:53.266949 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvpzg\" (UniqueName: \"kubernetes.io/projected/bfedd98f-1d06-49dd-94ac-3bc66c1bf603-kube-api-access-tvpzg\") pod \"swift-ring-rebalance-debug-kkbm5\" (UID: \"bfedd98f-1d06-49dd-94ac-3bc66c1bf603\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kkbm5" Mar 07 15:17:53 crc kubenswrapper[4943]: I0307 15:17:53.316600 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kkbm5" Mar 07 15:17:53 crc kubenswrapper[4943]: I0307 15:17:53.658534 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-kkbm5"] Mar 07 15:17:53 crc kubenswrapper[4943]: I0307 15:17:53.755596 4943 scope.go:117] "RemoveContainer" containerID="448308ad6e34ac922d66b3c90e66d0927407a94dd28633ec350502d69e4ca8aa" Mar 07 15:17:53 crc kubenswrapper[4943]: E0307 15:17:53.755889 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cpdmm_openshift-machine-config-operator(4c1f18ea-e557-41fc-95dd-ce4e1722a61e)\"" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" Mar 07 15:17:54 crc kubenswrapper[4943]: I0307 15:17:54.497221 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kkbm5" event={"ID":"bfedd98f-1d06-49dd-94ac-3bc66c1bf603","Type":"ContainerStarted","Data":"947668c3e7b1fdf5ff2eff97ead24071b8d501ac7fc0fdcbd0a21eb8e20c7eeb"} Mar 07 15:17:54 crc kubenswrapper[4943]: I0307 15:17:54.497715 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kkbm5" event={"ID":"bfedd98f-1d06-49dd-94ac-3bc66c1bf603","Type":"ContainerStarted","Data":"d84cacb872060332df3d064bc367c35c779f67c4b732f1c9049320ed338ed0ca"} Mar 07 15:17:54 crc kubenswrapper[4943]: I0307 15:17:54.523633 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kkbm5" podStartSLOduration=2.5236071989999997 podStartE2EDuration="2.523607199s" podCreationTimestamp="2026-03-07 15:17:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:17:54.513113033 +0000 UTC m=+2316.465249531" watchObservedRunningTime="2026-03-07 15:17:54.523607199 +0000 UTC m=+2316.475743727" Mar 07 15:17:55 crc kubenswrapper[4943]: I0307 15:17:55.515191 4943 generic.go:334] "Generic (PLEG): container finished" podID="bfedd98f-1d06-49dd-94ac-3bc66c1bf603" containerID="947668c3e7b1fdf5ff2eff97ead24071b8d501ac7fc0fdcbd0a21eb8e20c7eeb" exitCode=0 Mar 07 15:17:55 crc kubenswrapper[4943]: I0307 15:17:55.515301 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kkbm5" event={"ID":"bfedd98f-1d06-49dd-94ac-3bc66c1bf603","Type":"ContainerDied","Data":"947668c3e7b1fdf5ff2eff97ead24071b8d501ac7fc0fdcbd0a21eb8e20c7eeb"} Mar 07 15:17:56 crc kubenswrapper[4943]: I0307 15:17:56.814612 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kkbm5" Mar 07 15:17:56 crc kubenswrapper[4943]: I0307 15:17:56.851100 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-kkbm5"] Mar 07 15:17:56 crc kubenswrapper[4943]: I0307 15:17:56.859470 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-kkbm5"] Mar 07 15:17:56 crc kubenswrapper[4943]: I0307 15:17:56.993882 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bfedd98f-1d06-49dd-94ac-3bc66c1bf603-ring-data-devices\") pod \"bfedd98f-1d06-49dd-94ac-3bc66c1bf603\" (UID: \"bfedd98f-1d06-49dd-94ac-3bc66c1bf603\") " Mar 07 15:17:56 crc kubenswrapper[4943]: I0307 15:17:56.993957 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tvpzg\" (UniqueName: \"kubernetes.io/projected/bfedd98f-1d06-49dd-94ac-3bc66c1bf603-kube-api-access-tvpzg\") pod \"bfedd98f-1d06-49dd-94ac-3bc66c1bf603\" (UID: \"bfedd98f-1d06-49dd-94ac-3bc66c1bf603\") " Mar 07 15:17:56 crc kubenswrapper[4943]: I0307 15:17:56.994051 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bfedd98f-1d06-49dd-94ac-3bc66c1bf603-etc-swift\") pod \"bfedd98f-1d06-49dd-94ac-3bc66c1bf603\" (UID: \"bfedd98f-1d06-49dd-94ac-3bc66c1bf603\") " Mar 07 15:17:56 crc kubenswrapper[4943]: I0307 15:17:56.994105 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bfedd98f-1d06-49dd-94ac-3bc66c1bf603-scripts\") pod \"bfedd98f-1d06-49dd-94ac-3bc66c1bf603\" (UID: \"bfedd98f-1d06-49dd-94ac-3bc66c1bf603\") " Mar 07 15:17:56 crc kubenswrapper[4943]: I0307 15:17:56.994156 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bfedd98f-1d06-49dd-94ac-3bc66c1bf603-dispersionconf\") pod \"bfedd98f-1d06-49dd-94ac-3bc66c1bf603\" (UID: \"bfedd98f-1d06-49dd-94ac-3bc66c1bf603\") " Mar 07 15:17:56 crc kubenswrapper[4943]: I0307 15:17:56.994211 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bfedd98f-1d06-49dd-94ac-3bc66c1bf603-swiftconf\") pod \"bfedd98f-1d06-49dd-94ac-3bc66c1bf603\" (UID: \"bfedd98f-1d06-49dd-94ac-3bc66c1bf603\") " Mar 07 15:17:56 crc kubenswrapper[4943]: I0307 15:17:56.994910 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bfedd98f-1d06-49dd-94ac-3bc66c1bf603-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "bfedd98f-1d06-49dd-94ac-3bc66c1bf603" (UID: "bfedd98f-1d06-49dd-94ac-3bc66c1bf603"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:17:56 crc kubenswrapper[4943]: I0307 15:17:56.994997 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bfedd98f-1d06-49dd-94ac-3bc66c1bf603-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "bfedd98f-1d06-49dd-94ac-3bc66c1bf603" (UID: "bfedd98f-1d06-49dd-94ac-3bc66c1bf603"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:17:57 crc kubenswrapper[4943]: I0307 15:17:57.003586 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfedd98f-1d06-49dd-94ac-3bc66c1bf603-kube-api-access-tvpzg" (OuterVolumeSpecName: "kube-api-access-tvpzg") pod "bfedd98f-1d06-49dd-94ac-3bc66c1bf603" (UID: "bfedd98f-1d06-49dd-94ac-3bc66c1bf603"). InnerVolumeSpecName "kube-api-access-tvpzg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:17:57 crc kubenswrapper[4943]: I0307 15:17:57.022068 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfedd98f-1d06-49dd-94ac-3bc66c1bf603-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "bfedd98f-1d06-49dd-94ac-3bc66c1bf603" (UID: "bfedd98f-1d06-49dd-94ac-3bc66c1bf603"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:17:57 crc kubenswrapper[4943]: I0307 15:17:57.025667 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bfedd98f-1d06-49dd-94ac-3bc66c1bf603-scripts" (OuterVolumeSpecName: "scripts") pod "bfedd98f-1d06-49dd-94ac-3bc66c1bf603" (UID: "bfedd98f-1d06-49dd-94ac-3bc66c1bf603"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:17:57 crc kubenswrapper[4943]: I0307 15:17:57.039853 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfedd98f-1d06-49dd-94ac-3bc66c1bf603-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "bfedd98f-1d06-49dd-94ac-3bc66c1bf603" (UID: "bfedd98f-1d06-49dd-94ac-3bc66c1bf603"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:17:57 crc kubenswrapper[4943]: I0307 15:17:57.096139 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bfedd98f-1d06-49dd-94ac-3bc66c1bf603-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:57 crc kubenswrapper[4943]: I0307 15:17:57.096174 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bfedd98f-1d06-49dd-94ac-3bc66c1bf603-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:57 crc kubenswrapper[4943]: I0307 15:17:57.096182 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bfedd98f-1d06-49dd-94ac-3bc66c1bf603-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:57 crc kubenswrapper[4943]: I0307 15:17:57.096192 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bfedd98f-1d06-49dd-94ac-3bc66c1bf603-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:57 crc kubenswrapper[4943]: I0307 15:17:57.096200 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bfedd98f-1d06-49dd-94ac-3bc66c1bf603-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:57 crc kubenswrapper[4943]: I0307 15:17:57.096209 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tvpzg\" (UniqueName: \"kubernetes.io/projected/bfedd98f-1d06-49dd-94ac-3bc66c1bf603-kube-api-access-tvpzg\") on node \"crc\" DevicePath \"\"" Mar 07 15:17:57 crc kubenswrapper[4943]: I0307 15:17:57.542613 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d84cacb872060332df3d064bc367c35c779f67c4b732f1c9049320ed338ed0ca" Mar 07 15:17:57 crc kubenswrapper[4943]: I0307 15:17:57.542691 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kkbm5" Mar 07 15:17:58 crc kubenswrapper[4943]: I0307 15:17:58.332632 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-8457s"] Mar 07 15:17:58 crc kubenswrapper[4943]: E0307 15:17:58.333082 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfedd98f-1d06-49dd-94ac-3bc66c1bf603" containerName="swift-ring-rebalance" Mar 07 15:17:58 crc kubenswrapper[4943]: I0307 15:17:58.333103 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfedd98f-1d06-49dd-94ac-3bc66c1bf603" containerName="swift-ring-rebalance" Mar 07 15:17:58 crc kubenswrapper[4943]: I0307 15:17:58.333374 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfedd98f-1d06-49dd-94ac-3bc66c1bf603" containerName="swift-ring-rebalance" Mar 07 15:17:58 crc kubenswrapper[4943]: I0307 15:17:58.334165 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8457s" Mar 07 15:17:58 crc kubenswrapper[4943]: I0307 15:17:58.338027 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:17:58 crc kubenswrapper[4943]: I0307 15:17:58.341304 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:17:58 crc kubenswrapper[4943]: I0307 15:17:58.344747 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-8457s"] Mar 07 15:17:58 crc kubenswrapper[4943]: I0307 15:17:58.521596 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e13ed315-34fd-4f9e-aff2-6347ee26d2c9-ring-data-devices\") pod \"swift-ring-rebalance-debug-8457s\" (UID: \"e13ed315-34fd-4f9e-aff2-6347ee26d2c9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8457s" Mar 07 15:17:58 crc kubenswrapper[4943]: I0307 15:17:58.521726 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e13ed315-34fd-4f9e-aff2-6347ee26d2c9-dispersionconf\") pod \"swift-ring-rebalance-debug-8457s\" (UID: \"e13ed315-34fd-4f9e-aff2-6347ee26d2c9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8457s" Mar 07 15:17:58 crc kubenswrapper[4943]: I0307 15:17:58.521771 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e13ed315-34fd-4f9e-aff2-6347ee26d2c9-scripts\") pod \"swift-ring-rebalance-debug-8457s\" (UID: \"e13ed315-34fd-4f9e-aff2-6347ee26d2c9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8457s" Mar 07 15:17:58 crc kubenswrapper[4943]: I0307 15:17:58.521840 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e13ed315-34fd-4f9e-aff2-6347ee26d2c9-swiftconf\") pod \"swift-ring-rebalance-debug-8457s\" (UID: \"e13ed315-34fd-4f9e-aff2-6347ee26d2c9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8457s" Mar 07 15:17:58 crc kubenswrapper[4943]: I0307 15:17:58.522013 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5mtl\" (UniqueName: \"kubernetes.io/projected/e13ed315-34fd-4f9e-aff2-6347ee26d2c9-kube-api-access-t5mtl\") pod \"swift-ring-rebalance-debug-8457s\" (UID: \"e13ed315-34fd-4f9e-aff2-6347ee26d2c9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8457s" Mar 07 15:17:58 crc kubenswrapper[4943]: I0307 15:17:58.522091 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e13ed315-34fd-4f9e-aff2-6347ee26d2c9-etc-swift\") pod \"swift-ring-rebalance-debug-8457s\" (UID: \"e13ed315-34fd-4f9e-aff2-6347ee26d2c9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8457s" Mar 07 15:17:58 crc kubenswrapper[4943]: I0307 15:17:58.623896 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e13ed315-34fd-4f9e-aff2-6347ee26d2c9-ring-data-devices\") pod \"swift-ring-rebalance-debug-8457s\" (UID: \"e13ed315-34fd-4f9e-aff2-6347ee26d2c9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8457s" Mar 07 15:17:58 crc kubenswrapper[4943]: I0307 15:17:58.623990 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e13ed315-34fd-4f9e-aff2-6347ee26d2c9-dispersionconf\") pod \"swift-ring-rebalance-debug-8457s\" (UID: \"e13ed315-34fd-4f9e-aff2-6347ee26d2c9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8457s" Mar 07 15:17:58 crc kubenswrapper[4943]: I0307 15:17:58.624013 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e13ed315-34fd-4f9e-aff2-6347ee26d2c9-scripts\") pod \"swift-ring-rebalance-debug-8457s\" (UID: \"e13ed315-34fd-4f9e-aff2-6347ee26d2c9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8457s" Mar 07 15:17:58 crc kubenswrapper[4943]: I0307 15:17:58.624045 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e13ed315-34fd-4f9e-aff2-6347ee26d2c9-swiftconf\") pod \"swift-ring-rebalance-debug-8457s\" (UID: \"e13ed315-34fd-4f9e-aff2-6347ee26d2c9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8457s" Mar 07 15:17:58 crc kubenswrapper[4943]: I0307 15:17:58.624073 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5mtl\" (UniqueName: \"kubernetes.io/projected/e13ed315-34fd-4f9e-aff2-6347ee26d2c9-kube-api-access-t5mtl\") pod \"swift-ring-rebalance-debug-8457s\" (UID: \"e13ed315-34fd-4f9e-aff2-6347ee26d2c9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8457s" Mar 07 15:17:58 crc kubenswrapper[4943]: I0307 15:17:58.624093 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e13ed315-34fd-4f9e-aff2-6347ee26d2c9-etc-swift\") pod \"swift-ring-rebalance-debug-8457s\" (UID: \"e13ed315-34fd-4f9e-aff2-6347ee26d2c9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8457s" Mar 07 15:17:58 crc kubenswrapper[4943]: I0307 15:17:58.624536 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e13ed315-34fd-4f9e-aff2-6347ee26d2c9-etc-swift\") pod \"swift-ring-rebalance-debug-8457s\" (UID: \"e13ed315-34fd-4f9e-aff2-6347ee26d2c9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8457s" Mar 07 15:17:58 crc kubenswrapper[4943]: I0307 15:17:58.625742 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e13ed315-34fd-4f9e-aff2-6347ee26d2c9-scripts\") pod \"swift-ring-rebalance-debug-8457s\" (UID: \"e13ed315-34fd-4f9e-aff2-6347ee26d2c9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8457s" Mar 07 15:17:58 crc kubenswrapper[4943]: I0307 15:17:58.626333 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e13ed315-34fd-4f9e-aff2-6347ee26d2c9-ring-data-devices\") pod \"swift-ring-rebalance-debug-8457s\" (UID: \"e13ed315-34fd-4f9e-aff2-6347ee26d2c9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8457s" Mar 07 15:17:58 crc kubenswrapper[4943]: I0307 15:17:58.636745 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e13ed315-34fd-4f9e-aff2-6347ee26d2c9-swiftconf\") pod \"swift-ring-rebalance-debug-8457s\" (UID: \"e13ed315-34fd-4f9e-aff2-6347ee26d2c9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8457s" Mar 07 15:17:58 crc kubenswrapper[4943]: I0307 15:17:58.637985 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e13ed315-34fd-4f9e-aff2-6347ee26d2c9-dispersionconf\") pod \"swift-ring-rebalance-debug-8457s\" (UID: \"e13ed315-34fd-4f9e-aff2-6347ee26d2c9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8457s" Mar 07 15:17:58 crc kubenswrapper[4943]: I0307 15:17:58.655410 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5mtl\" (UniqueName: \"kubernetes.io/projected/e13ed315-34fd-4f9e-aff2-6347ee26d2c9-kube-api-access-t5mtl\") pod \"swift-ring-rebalance-debug-8457s\" (UID: \"e13ed315-34fd-4f9e-aff2-6347ee26d2c9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8457s" Mar 07 15:17:58 crc kubenswrapper[4943]: I0307 15:17:58.659809 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8457s" Mar 07 15:17:58 crc kubenswrapper[4943]: I0307 15:17:58.773806 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bfedd98f-1d06-49dd-94ac-3bc66c1bf603" path="/var/lib/kubelet/pods/bfedd98f-1d06-49dd-94ac-3bc66c1bf603/volumes" Mar 07 15:17:59 crc kubenswrapper[4943]: I0307 15:17:59.143250 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-8457s"] Mar 07 15:17:59 crc kubenswrapper[4943]: I0307 15:17:59.581734 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8457s" event={"ID":"e13ed315-34fd-4f9e-aff2-6347ee26d2c9","Type":"ContainerStarted","Data":"87f5dcfec89a023c445838e4a2e2d9c91f334a96f940176f3451c0ee88f4f88a"} Mar 07 15:18:00 crc kubenswrapper[4943]: I0307 15:18:00.142307 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29548278-f5c2x"] Mar 07 15:18:00 crc kubenswrapper[4943]: I0307 15:18:00.143349 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548278-f5c2x" Mar 07 15:18:00 crc kubenswrapper[4943]: I0307 15:18:00.145100 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 07 15:18:00 crc kubenswrapper[4943]: I0307 15:18:00.146214 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 07 15:18:00 crc kubenswrapper[4943]: I0307 15:18:00.147220 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-qqlk5" Mar 07 15:18:00 crc kubenswrapper[4943]: I0307 15:18:00.148504 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6xlv\" (UniqueName: \"kubernetes.io/projected/fe264b14-43cf-4101-9c71-ae8a5c77a3e5-kube-api-access-m6xlv\") pod \"auto-csr-approver-29548278-f5c2x\" (UID: \"fe264b14-43cf-4101-9c71-ae8a5c77a3e5\") " pod="openshift-infra/auto-csr-approver-29548278-f5c2x" Mar 07 15:18:00 crc kubenswrapper[4943]: I0307 15:18:00.157939 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29548278-f5c2x"] Mar 07 15:18:00 crc kubenswrapper[4943]: I0307 15:18:00.249989 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6xlv\" (UniqueName: \"kubernetes.io/projected/fe264b14-43cf-4101-9c71-ae8a5c77a3e5-kube-api-access-m6xlv\") pod \"auto-csr-approver-29548278-f5c2x\" (UID: \"fe264b14-43cf-4101-9c71-ae8a5c77a3e5\") " pod="openshift-infra/auto-csr-approver-29548278-f5c2x" Mar 07 15:18:00 crc kubenswrapper[4943]: I0307 15:18:00.271574 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6xlv\" (UniqueName: \"kubernetes.io/projected/fe264b14-43cf-4101-9c71-ae8a5c77a3e5-kube-api-access-m6xlv\") pod \"auto-csr-approver-29548278-f5c2x\" (UID: \"fe264b14-43cf-4101-9c71-ae8a5c77a3e5\") " pod="openshift-infra/auto-csr-approver-29548278-f5c2x" Mar 07 15:18:00 crc kubenswrapper[4943]: I0307 15:18:00.466613 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548278-f5c2x" Mar 07 15:18:00 crc kubenswrapper[4943]: I0307 15:18:00.596904 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8457s" event={"ID":"e13ed315-34fd-4f9e-aff2-6347ee26d2c9","Type":"ContainerStarted","Data":"9cee618806c4822eec6839068f839c17ac0b9228435ce5aabeb5cf561850d0fe"} Mar 07 15:18:00 crc kubenswrapper[4943]: I0307 15:18:00.631324 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8457s" podStartSLOduration=2.631303077 podStartE2EDuration="2.631303077s" podCreationTimestamp="2026-03-07 15:17:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:18:00.621365134 +0000 UTC m=+2322.573501642" watchObservedRunningTime="2026-03-07 15:18:00.631303077 +0000 UTC m=+2322.583439595" Mar 07 15:18:00 crc kubenswrapper[4943]: I0307 15:18:00.769234 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29548278-f5c2x"] Mar 07 15:18:00 crc kubenswrapper[4943]: W0307 15:18:00.774788 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfe264b14_43cf_4101_9c71_ae8a5c77a3e5.slice/crio-a4cdb52839851c364593297630808299086a9dc912ad3aebd5b1f031359be548 WatchSource:0}: Error finding container a4cdb52839851c364593297630808299086a9dc912ad3aebd5b1f031359be548: Status 404 returned error can't find the container with id a4cdb52839851c364593297630808299086a9dc912ad3aebd5b1f031359be548 Mar 07 15:18:01 crc kubenswrapper[4943]: I0307 15:18:01.610133 4943 generic.go:334] "Generic (PLEG): container finished" podID="e13ed315-34fd-4f9e-aff2-6347ee26d2c9" containerID="9cee618806c4822eec6839068f839c17ac0b9228435ce5aabeb5cf561850d0fe" exitCode=0 Mar 07 15:18:01 crc kubenswrapper[4943]: I0307 15:18:01.610238 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8457s" event={"ID":"e13ed315-34fd-4f9e-aff2-6347ee26d2c9","Type":"ContainerDied","Data":"9cee618806c4822eec6839068f839c17ac0b9228435ce5aabeb5cf561850d0fe"} Mar 07 15:18:01 crc kubenswrapper[4943]: I0307 15:18:01.612299 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548278-f5c2x" event={"ID":"fe264b14-43cf-4101-9c71-ae8a5c77a3e5","Type":"ContainerStarted","Data":"a4cdb52839851c364593297630808299086a9dc912ad3aebd5b1f031359be548"} Mar 07 15:18:02 crc kubenswrapper[4943]: I0307 15:18:02.391475 4943 prober.go:107] "Probe failed" probeType="Readiness" pod="swift-kuttl-tests/swift-proxy-76c998454c-kngkw" podUID="32f9bc3c-6fe2-42ab-8485-d7fda4d10c78" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Mar 07 15:18:02 crc kubenswrapper[4943]: I0307 15:18:02.624133 4943 generic.go:334] "Generic (PLEG): container finished" podID="fe264b14-43cf-4101-9c71-ae8a5c77a3e5" containerID="58716c0f983cc31149b498b43d095ecc0c1da9ab57bc83f61ceeeb033fbe320e" exitCode=0 Mar 07 15:18:02 crc kubenswrapper[4943]: I0307 15:18:02.624225 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548278-f5c2x" event={"ID":"fe264b14-43cf-4101-9c71-ae8a5c77a3e5","Type":"ContainerDied","Data":"58716c0f983cc31149b498b43d095ecc0c1da9ab57bc83f61ceeeb033fbe320e"} Mar 07 15:18:02 crc kubenswrapper[4943]: I0307 15:18:02.900894 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8457s" Mar 07 15:18:02 crc kubenswrapper[4943]: I0307 15:18:02.931649 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-8457s"] Mar 07 15:18:02 crc kubenswrapper[4943]: I0307 15:18:02.939902 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-8457s"] Mar 07 15:18:02 crc kubenswrapper[4943]: I0307 15:18:02.992279 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e13ed315-34fd-4f9e-aff2-6347ee26d2c9-etc-swift\") pod \"e13ed315-34fd-4f9e-aff2-6347ee26d2c9\" (UID: \"e13ed315-34fd-4f9e-aff2-6347ee26d2c9\") " Mar 07 15:18:02 crc kubenswrapper[4943]: I0307 15:18:02.992359 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e13ed315-34fd-4f9e-aff2-6347ee26d2c9-dispersionconf\") pod \"e13ed315-34fd-4f9e-aff2-6347ee26d2c9\" (UID: \"e13ed315-34fd-4f9e-aff2-6347ee26d2c9\") " Mar 07 15:18:02 crc kubenswrapper[4943]: I0307 15:18:02.992435 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e13ed315-34fd-4f9e-aff2-6347ee26d2c9-scripts\") pod \"e13ed315-34fd-4f9e-aff2-6347ee26d2c9\" (UID: \"e13ed315-34fd-4f9e-aff2-6347ee26d2c9\") " Mar 07 15:18:02 crc kubenswrapper[4943]: I0307 15:18:02.992535 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t5mtl\" (UniqueName: \"kubernetes.io/projected/e13ed315-34fd-4f9e-aff2-6347ee26d2c9-kube-api-access-t5mtl\") pod \"e13ed315-34fd-4f9e-aff2-6347ee26d2c9\" (UID: \"e13ed315-34fd-4f9e-aff2-6347ee26d2c9\") " Mar 07 15:18:02 crc kubenswrapper[4943]: I0307 15:18:02.992569 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e13ed315-34fd-4f9e-aff2-6347ee26d2c9-swiftconf\") pod \"e13ed315-34fd-4f9e-aff2-6347ee26d2c9\" (UID: \"e13ed315-34fd-4f9e-aff2-6347ee26d2c9\") " Mar 07 15:18:02 crc kubenswrapper[4943]: I0307 15:18:02.992632 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e13ed315-34fd-4f9e-aff2-6347ee26d2c9-ring-data-devices\") pod \"e13ed315-34fd-4f9e-aff2-6347ee26d2c9\" (UID: \"e13ed315-34fd-4f9e-aff2-6347ee26d2c9\") " Mar 07 15:18:02 crc kubenswrapper[4943]: I0307 15:18:02.993413 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e13ed315-34fd-4f9e-aff2-6347ee26d2c9-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "e13ed315-34fd-4f9e-aff2-6347ee26d2c9" (UID: "e13ed315-34fd-4f9e-aff2-6347ee26d2c9"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:18:02 crc kubenswrapper[4943]: I0307 15:18:02.993987 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e13ed315-34fd-4f9e-aff2-6347ee26d2c9-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "e13ed315-34fd-4f9e-aff2-6347ee26d2c9" (UID: "e13ed315-34fd-4f9e-aff2-6347ee26d2c9"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:18:03 crc kubenswrapper[4943]: I0307 15:18:03.020180 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e13ed315-34fd-4f9e-aff2-6347ee26d2c9-kube-api-access-t5mtl" (OuterVolumeSpecName: "kube-api-access-t5mtl") pod "e13ed315-34fd-4f9e-aff2-6347ee26d2c9" (UID: "e13ed315-34fd-4f9e-aff2-6347ee26d2c9"). InnerVolumeSpecName "kube-api-access-t5mtl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:18:03 crc kubenswrapper[4943]: I0307 15:18:03.036282 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e13ed315-34fd-4f9e-aff2-6347ee26d2c9-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "e13ed315-34fd-4f9e-aff2-6347ee26d2c9" (UID: "e13ed315-34fd-4f9e-aff2-6347ee26d2c9"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:18:03 crc kubenswrapper[4943]: I0307 15:18:03.038517 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e13ed315-34fd-4f9e-aff2-6347ee26d2c9-scripts" (OuterVolumeSpecName: "scripts") pod "e13ed315-34fd-4f9e-aff2-6347ee26d2c9" (UID: "e13ed315-34fd-4f9e-aff2-6347ee26d2c9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:18:03 crc kubenswrapper[4943]: I0307 15:18:03.042034 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e13ed315-34fd-4f9e-aff2-6347ee26d2c9-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "e13ed315-34fd-4f9e-aff2-6347ee26d2c9" (UID: "e13ed315-34fd-4f9e-aff2-6347ee26d2c9"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:18:03 crc kubenswrapper[4943]: I0307 15:18:03.094010 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e13ed315-34fd-4f9e-aff2-6347ee26d2c9-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:03 crc kubenswrapper[4943]: I0307 15:18:03.094044 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e13ed315-34fd-4f9e-aff2-6347ee26d2c9-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:03 crc kubenswrapper[4943]: I0307 15:18:03.094054 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t5mtl\" (UniqueName: \"kubernetes.io/projected/e13ed315-34fd-4f9e-aff2-6347ee26d2c9-kube-api-access-t5mtl\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:03 crc kubenswrapper[4943]: I0307 15:18:03.094064 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e13ed315-34fd-4f9e-aff2-6347ee26d2c9-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:03 crc kubenswrapper[4943]: I0307 15:18:03.094072 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e13ed315-34fd-4f9e-aff2-6347ee26d2c9-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:03 crc kubenswrapper[4943]: I0307 15:18:03.094079 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e13ed315-34fd-4f9e-aff2-6347ee26d2c9-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:03 crc kubenswrapper[4943]: I0307 15:18:03.640197 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="87f5dcfec89a023c445838e4a2e2d9c91f334a96f940176f3451c0ee88f4f88a" Mar 07 15:18:03 crc kubenswrapper[4943]: I0307 15:18:03.640242 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8457s" Mar 07 15:18:03 crc kubenswrapper[4943]: I0307 15:18:03.916421 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548278-f5c2x" Mar 07 15:18:04 crc kubenswrapper[4943]: I0307 15:18:04.007281 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m6xlv\" (UniqueName: \"kubernetes.io/projected/fe264b14-43cf-4101-9c71-ae8a5c77a3e5-kube-api-access-m6xlv\") pod \"fe264b14-43cf-4101-9c71-ae8a5c77a3e5\" (UID: \"fe264b14-43cf-4101-9c71-ae8a5c77a3e5\") " Mar 07 15:18:04 crc kubenswrapper[4943]: I0307 15:18:04.012428 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe264b14-43cf-4101-9c71-ae8a5c77a3e5-kube-api-access-m6xlv" (OuterVolumeSpecName: "kube-api-access-m6xlv") pod "fe264b14-43cf-4101-9c71-ae8a5c77a3e5" (UID: "fe264b14-43cf-4101-9c71-ae8a5c77a3e5"). InnerVolumeSpecName "kube-api-access-m6xlv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:18:04 crc kubenswrapper[4943]: I0307 15:18:04.090629 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-hwcxp"] Mar 07 15:18:04 crc kubenswrapper[4943]: E0307 15:18:04.090890 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe264b14-43cf-4101-9c71-ae8a5c77a3e5" containerName="oc" Mar 07 15:18:04 crc kubenswrapper[4943]: I0307 15:18:04.090906 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe264b14-43cf-4101-9c71-ae8a5c77a3e5" containerName="oc" Mar 07 15:18:04 crc kubenswrapper[4943]: E0307 15:18:04.090951 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e13ed315-34fd-4f9e-aff2-6347ee26d2c9" containerName="swift-ring-rebalance" Mar 07 15:18:04 crc kubenswrapper[4943]: I0307 15:18:04.090958 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="e13ed315-34fd-4f9e-aff2-6347ee26d2c9" containerName="swift-ring-rebalance" Mar 07 15:18:04 crc kubenswrapper[4943]: I0307 15:18:04.091086 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe264b14-43cf-4101-9c71-ae8a5c77a3e5" containerName="oc" Mar 07 15:18:04 crc kubenswrapper[4943]: I0307 15:18:04.091100 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="e13ed315-34fd-4f9e-aff2-6347ee26d2c9" containerName="swift-ring-rebalance" Mar 07 15:18:04 crc kubenswrapper[4943]: I0307 15:18:04.091536 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-hwcxp" Mar 07 15:18:04 crc kubenswrapper[4943]: I0307 15:18:04.093474 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:18:04 crc kubenswrapper[4943]: I0307 15:18:04.093545 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:18:04 crc kubenswrapper[4943]: I0307 15:18:04.109031 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m6xlv\" (UniqueName: \"kubernetes.io/projected/fe264b14-43cf-4101-9c71-ae8a5c77a3e5-kube-api-access-m6xlv\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:04 crc kubenswrapper[4943]: I0307 15:18:04.116642 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-hwcxp"] Mar 07 15:18:04 crc kubenswrapper[4943]: I0307 15:18:04.209846 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6f6ch\" (UniqueName: \"kubernetes.io/projected/ccc1f850-fe24-475f-9abf-e8805f8fe126-kube-api-access-6f6ch\") pod \"swift-ring-rebalance-debug-hwcxp\" (UID: \"ccc1f850-fe24-475f-9abf-e8805f8fe126\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hwcxp" Mar 07 15:18:04 crc kubenswrapper[4943]: I0307 15:18:04.209891 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ccc1f850-fe24-475f-9abf-e8805f8fe126-etc-swift\") pod \"swift-ring-rebalance-debug-hwcxp\" (UID: \"ccc1f850-fe24-475f-9abf-e8805f8fe126\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hwcxp" Mar 07 15:18:04 crc kubenswrapper[4943]: I0307 15:18:04.210172 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ccc1f850-fe24-475f-9abf-e8805f8fe126-swiftconf\") pod \"swift-ring-rebalance-debug-hwcxp\" (UID: \"ccc1f850-fe24-475f-9abf-e8805f8fe126\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hwcxp" Mar 07 15:18:04 crc kubenswrapper[4943]: I0307 15:18:04.210243 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ccc1f850-fe24-475f-9abf-e8805f8fe126-dispersionconf\") pod \"swift-ring-rebalance-debug-hwcxp\" (UID: \"ccc1f850-fe24-475f-9abf-e8805f8fe126\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hwcxp" Mar 07 15:18:04 crc kubenswrapper[4943]: I0307 15:18:04.210368 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ccc1f850-fe24-475f-9abf-e8805f8fe126-ring-data-devices\") pod \"swift-ring-rebalance-debug-hwcxp\" (UID: \"ccc1f850-fe24-475f-9abf-e8805f8fe126\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hwcxp" Mar 07 15:18:04 crc kubenswrapper[4943]: I0307 15:18:04.210473 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ccc1f850-fe24-475f-9abf-e8805f8fe126-scripts\") pod \"swift-ring-rebalance-debug-hwcxp\" (UID: \"ccc1f850-fe24-475f-9abf-e8805f8fe126\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hwcxp" Mar 07 15:18:04 crc kubenswrapper[4943]: I0307 15:18:04.312187 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ccc1f850-fe24-475f-9abf-e8805f8fe126-scripts\") pod \"swift-ring-rebalance-debug-hwcxp\" (UID: \"ccc1f850-fe24-475f-9abf-e8805f8fe126\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hwcxp" Mar 07 15:18:04 crc kubenswrapper[4943]: I0307 15:18:04.312340 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6f6ch\" (UniqueName: \"kubernetes.io/projected/ccc1f850-fe24-475f-9abf-e8805f8fe126-kube-api-access-6f6ch\") pod \"swift-ring-rebalance-debug-hwcxp\" (UID: \"ccc1f850-fe24-475f-9abf-e8805f8fe126\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hwcxp" Mar 07 15:18:04 crc kubenswrapper[4943]: I0307 15:18:04.312399 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ccc1f850-fe24-475f-9abf-e8805f8fe126-etc-swift\") pod \"swift-ring-rebalance-debug-hwcxp\" (UID: \"ccc1f850-fe24-475f-9abf-e8805f8fe126\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hwcxp" Mar 07 15:18:04 crc kubenswrapper[4943]: I0307 15:18:04.312477 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ccc1f850-fe24-475f-9abf-e8805f8fe126-swiftconf\") pod \"swift-ring-rebalance-debug-hwcxp\" (UID: \"ccc1f850-fe24-475f-9abf-e8805f8fe126\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hwcxp" Mar 07 15:18:04 crc kubenswrapper[4943]: I0307 15:18:04.312511 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ccc1f850-fe24-475f-9abf-e8805f8fe126-dispersionconf\") pod \"swift-ring-rebalance-debug-hwcxp\" (UID: \"ccc1f850-fe24-475f-9abf-e8805f8fe126\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hwcxp" Mar 07 15:18:04 crc kubenswrapper[4943]: I0307 15:18:04.312571 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ccc1f850-fe24-475f-9abf-e8805f8fe126-ring-data-devices\") pod \"swift-ring-rebalance-debug-hwcxp\" (UID: \"ccc1f850-fe24-475f-9abf-e8805f8fe126\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hwcxp" Mar 07 15:18:04 crc kubenswrapper[4943]: I0307 15:18:04.313288 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ccc1f850-fe24-475f-9abf-e8805f8fe126-etc-swift\") pod \"swift-ring-rebalance-debug-hwcxp\" (UID: \"ccc1f850-fe24-475f-9abf-e8805f8fe126\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hwcxp" Mar 07 15:18:04 crc kubenswrapper[4943]: I0307 15:18:04.313788 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ccc1f850-fe24-475f-9abf-e8805f8fe126-ring-data-devices\") pod \"swift-ring-rebalance-debug-hwcxp\" (UID: \"ccc1f850-fe24-475f-9abf-e8805f8fe126\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hwcxp" Mar 07 15:18:04 crc kubenswrapper[4943]: I0307 15:18:04.313786 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ccc1f850-fe24-475f-9abf-e8805f8fe126-scripts\") pod \"swift-ring-rebalance-debug-hwcxp\" (UID: \"ccc1f850-fe24-475f-9abf-e8805f8fe126\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hwcxp" Mar 07 15:18:04 crc kubenswrapper[4943]: I0307 15:18:04.316118 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ccc1f850-fe24-475f-9abf-e8805f8fe126-swiftconf\") pod \"swift-ring-rebalance-debug-hwcxp\" (UID: \"ccc1f850-fe24-475f-9abf-e8805f8fe126\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hwcxp" Mar 07 15:18:04 crc kubenswrapper[4943]: I0307 15:18:04.316400 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ccc1f850-fe24-475f-9abf-e8805f8fe126-dispersionconf\") pod \"swift-ring-rebalance-debug-hwcxp\" (UID: \"ccc1f850-fe24-475f-9abf-e8805f8fe126\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hwcxp" Mar 07 15:18:04 crc kubenswrapper[4943]: I0307 15:18:04.344058 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6f6ch\" (UniqueName: \"kubernetes.io/projected/ccc1f850-fe24-475f-9abf-e8805f8fe126-kube-api-access-6f6ch\") pod \"swift-ring-rebalance-debug-hwcxp\" (UID: \"ccc1f850-fe24-475f-9abf-e8805f8fe126\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hwcxp" Mar 07 15:18:04 crc kubenswrapper[4943]: I0307 15:18:04.417619 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-hwcxp" Mar 07 15:18:04 crc kubenswrapper[4943]: I0307 15:18:04.655266 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548278-f5c2x" event={"ID":"fe264b14-43cf-4101-9c71-ae8a5c77a3e5","Type":"ContainerDied","Data":"a4cdb52839851c364593297630808299086a9dc912ad3aebd5b1f031359be548"} Mar 07 15:18:04 crc kubenswrapper[4943]: I0307 15:18:04.655655 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a4cdb52839851c364593297630808299086a9dc912ad3aebd5b1f031359be548" Mar 07 15:18:04 crc kubenswrapper[4943]: I0307 15:18:04.655325 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548278-f5c2x" Mar 07 15:18:04 crc kubenswrapper[4943]: I0307 15:18:04.763279 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e13ed315-34fd-4f9e-aff2-6347ee26d2c9" path="/var/lib/kubelet/pods/e13ed315-34fd-4f9e-aff2-6347ee26d2c9/volumes" Mar 07 15:18:04 crc kubenswrapper[4943]: W0307 15:18:04.892265 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podccc1f850_fe24_475f_9abf_e8805f8fe126.slice/crio-52270c24b888efeb4cd72984f91b92a3e3a65a57e0a09ed0078b17e54f0a77c2 WatchSource:0}: Error finding container 52270c24b888efeb4cd72984f91b92a3e3a65a57e0a09ed0078b17e54f0a77c2: Status 404 returned error can't find the container with id 52270c24b888efeb4cd72984f91b92a3e3a65a57e0a09ed0078b17e54f0a77c2 Mar 07 15:18:04 crc kubenswrapper[4943]: I0307 15:18:04.898904 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-hwcxp"] Mar 07 15:18:05 crc kubenswrapper[4943]: I0307 15:18:05.052329 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29548272-qdhrf"] Mar 07 15:18:05 crc kubenswrapper[4943]: I0307 15:18:05.063380 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29548272-qdhrf"] Mar 07 15:18:05 crc kubenswrapper[4943]: I0307 15:18:05.686304 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-hwcxp" event={"ID":"ccc1f850-fe24-475f-9abf-e8805f8fe126","Type":"ContainerStarted","Data":"d1ae9074ee81b7864d70c49b679509bbf9bf02c3d8d0fe7b9fac37707ed78614"} Mar 07 15:18:05 crc kubenswrapper[4943]: I0307 15:18:05.686693 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-hwcxp" event={"ID":"ccc1f850-fe24-475f-9abf-e8805f8fe126","Type":"ContainerStarted","Data":"52270c24b888efeb4cd72984f91b92a3e3a65a57e0a09ed0078b17e54f0a77c2"} Mar 07 15:18:05 crc kubenswrapper[4943]: I0307 15:18:05.708875 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-hwcxp" podStartSLOduration=1.708858881 podStartE2EDuration="1.708858881s" podCreationTimestamp="2026-03-07 15:18:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:18:05.705473499 +0000 UTC m=+2327.657610017" watchObservedRunningTime="2026-03-07 15:18:05.708858881 +0000 UTC m=+2327.660995389" Mar 07 15:18:05 crc kubenswrapper[4943]: I0307 15:18:05.755688 4943 scope.go:117] "RemoveContainer" containerID="448308ad6e34ac922d66b3c90e66d0927407a94dd28633ec350502d69e4ca8aa" Mar 07 15:18:05 crc kubenswrapper[4943]: E0307 15:18:05.755974 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cpdmm_openshift-machine-config-operator(4c1f18ea-e557-41fc-95dd-ce4e1722a61e)\"" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" Mar 07 15:18:06 crc kubenswrapper[4943]: I0307 15:18:06.713718 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-hwcxp" event={"ID":"ccc1f850-fe24-475f-9abf-e8805f8fe126","Type":"ContainerDied","Data":"d1ae9074ee81b7864d70c49b679509bbf9bf02c3d8d0fe7b9fac37707ed78614"} Mar 07 15:18:06 crc kubenswrapper[4943]: I0307 15:18:06.713784 4943 generic.go:334] "Generic (PLEG): container finished" podID="ccc1f850-fe24-475f-9abf-e8805f8fe126" containerID="d1ae9074ee81b7864d70c49b679509bbf9bf02c3d8d0fe7b9fac37707ed78614" exitCode=0 Mar 07 15:18:06 crc kubenswrapper[4943]: I0307 15:18:06.770264 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3920f3e-037a-4678-b3dd-9da204486979" path="/var/lib/kubelet/pods/f3920f3e-037a-4678-b3dd-9da204486979/volumes" Mar 07 15:18:08 crc kubenswrapper[4943]: I0307 15:18:08.038737 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-hwcxp" Mar 07 15:18:08 crc kubenswrapper[4943]: I0307 15:18:08.104418 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-hwcxp"] Mar 07 15:18:08 crc kubenswrapper[4943]: I0307 15:18:08.125049 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-hwcxp"] Mar 07 15:18:08 crc kubenswrapper[4943]: I0307 15:18:08.178186 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ccc1f850-fe24-475f-9abf-e8805f8fe126-swiftconf\") pod \"ccc1f850-fe24-475f-9abf-e8805f8fe126\" (UID: \"ccc1f850-fe24-475f-9abf-e8805f8fe126\") " Mar 07 15:18:08 crc kubenswrapper[4943]: I0307 15:18:08.178312 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6f6ch\" (UniqueName: \"kubernetes.io/projected/ccc1f850-fe24-475f-9abf-e8805f8fe126-kube-api-access-6f6ch\") pod \"ccc1f850-fe24-475f-9abf-e8805f8fe126\" (UID: \"ccc1f850-fe24-475f-9abf-e8805f8fe126\") " Mar 07 15:18:08 crc kubenswrapper[4943]: I0307 15:18:08.178346 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ccc1f850-fe24-475f-9abf-e8805f8fe126-ring-data-devices\") pod \"ccc1f850-fe24-475f-9abf-e8805f8fe126\" (UID: \"ccc1f850-fe24-475f-9abf-e8805f8fe126\") " Mar 07 15:18:08 crc kubenswrapper[4943]: I0307 15:18:08.178368 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ccc1f850-fe24-475f-9abf-e8805f8fe126-scripts\") pod \"ccc1f850-fe24-475f-9abf-e8805f8fe126\" (UID: \"ccc1f850-fe24-475f-9abf-e8805f8fe126\") " Mar 07 15:18:08 crc kubenswrapper[4943]: I0307 15:18:08.178409 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ccc1f850-fe24-475f-9abf-e8805f8fe126-etc-swift\") pod \"ccc1f850-fe24-475f-9abf-e8805f8fe126\" (UID: \"ccc1f850-fe24-475f-9abf-e8805f8fe126\") " Mar 07 15:18:08 crc kubenswrapper[4943]: I0307 15:18:08.178432 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ccc1f850-fe24-475f-9abf-e8805f8fe126-dispersionconf\") pod \"ccc1f850-fe24-475f-9abf-e8805f8fe126\" (UID: \"ccc1f850-fe24-475f-9abf-e8805f8fe126\") " Mar 07 15:18:08 crc kubenswrapper[4943]: I0307 15:18:08.179728 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccc1f850-fe24-475f-9abf-e8805f8fe126-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "ccc1f850-fe24-475f-9abf-e8805f8fe126" (UID: "ccc1f850-fe24-475f-9abf-e8805f8fe126"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:18:08 crc kubenswrapper[4943]: I0307 15:18:08.189736 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ccc1f850-fe24-475f-9abf-e8805f8fe126-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "ccc1f850-fe24-475f-9abf-e8805f8fe126" (UID: "ccc1f850-fe24-475f-9abf-e8805f8fe126"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:18:08 crc kubenswrapper[4943]: I0307 15:18:08.195160 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccc1f850-fe24-475f-9abf-e8805f8fe126-kube-api-access-6f6ch" (OuterVolumeSpecName: "kube-api-access-6f6ch") pod "ccc1f850-fe24-475f-9abf-e8805f8fe126" (UID: "ccc1f850-fe24-475f-9abf-e8805f8fe126"). InnerVolumeSpecName "kube-api-access-6f6ch". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:18:08 crc kubenswrapper[4943]: I0307 15:18:08.207673 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccc1f850-fe24-475f-9abf-e8805f8fe126-scripts" (OuterVolumeSpecName: "scripts") pod "ccc1f850-fe24-475f-9abf-e8805f8fe126" (UID: "ccc1f850-fe24-475f-9abf-e8805f8fe126"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:18:08 crc kubenswrapper[4943]: I0307 15:18:08.207787 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccc1f850-fe24-475f-9abf-e8805f8fe126-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "ccc1f850-fe24-475f-9abf-e8805f8fe126" (UID: "ccc1f850-fe24-475f-9abf-e8805f8fe126"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:18:08 crc kubenswrapper[4943]: I0307 15:18:08.209704 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccc1f850-fe24-475f-9abf-e8805f8fe126-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "ccc1f850-fe24-475f-9abf-e8805f8fe126" (UID: "ccc1f850-fe24-475f-9abf-e8805f8fe126"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:18:08 crc kubenswrapper[4943]: I0307 15:18:08.279979 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6f6ch\" (UniqueName: \"kubernetes.io/projected/ccc1f850-fe24-475f-9abf-e8805f8fe126-kube-api-access-6f6ch\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:08 crc kubenswrapper[4943]: I0307 15:18:08.280032 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ccc1f850-fe24-475f-9abf-e8805f8fe126-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:08 crc kubenswrapper[4943]: I0307 15:18:08.280052 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ccc1f850-fe24-475f-9abf-e8805f8fe126-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:08 crc kubenswrapper[4943]: I0307 15:18:08.280066 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ccc1f850-fe24-475f-9abf-e8805f8fe126-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:08 crc kubenswrapper[4943]: I0307 15:18:08.280082 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ccc1f850-fe24-475f-9abf-e8805f8fe126-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:08 crc kubenswrapper[4943]: I0307 15:18:08.280095 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ccc1f850-fe24-475f-9abf-e8805f8fe126-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:08 crc kubenswrapper[4943]: I0307 15:18:08.741144 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="52270c24b888efeb4cd72984f91b92a3e3a65a57e0a09ed0078b17e54f0a77c2" Mar 07 15:18:08 crc kubenswrapper[4943]: I0307 15:18:08.741250 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-hwcxp" Mar 07 15:18:08 crc kubenswrapper[4943]: I0307 15:18:08.774297 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ccc1f850-fe24-475f-9abf-e8805f8fe126" path="/var/lib/kubelet/pods/ccc1f850-fe24-475f-9abf-e8805f8fe126/volumes" Mar 07 15:18:09 crc kubenswrapper[4943]: I0307 15:18:09.227129 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-pbcsx"] Mar 07 15:18:09 crc kubenswrapper[4943]: E0307 15:18:09.227500 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccc1f850-fe24-475f-9abf-e8805f8fe126" containerName="swift-ring-rebalance" Mar 07 15:18:09 crc kubenswrapper[4943]: I0307 15:18:09.227514 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccc1f850-fe24-475f-9abf-e8805f8fe126" containerName="swift-ring-rebalance" Mar 07 15:18:09 crc kubenswrapper[4943]: I0307 15:18:09.227712 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccc1f850-fe24-475f-9abf-e8805f8fe126" containerName="swift-ring-rebalance" Mar 07 15:18:09 crc kubenswrapper[4943]: I0307 15:18:09.228440 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pbcsx" Mar 07 15:18:09 crc kubenswrapper[4943]: I0307 15:18:09.236277 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:18:09 crc kubenswrapper[4943]: I0307 15:18:09.238297 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-pbcsx"] Mar 07 15:18:09 crc kubenswrapper[4943]: I0307 15:18:09.242408 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:18:09 crc kubenswrapper[4943]: I0307 15:18:09.400111 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/cb6a3d56-01ef-47fc-a633-6f7c4f69f27f-etc-swift\") pod \"swift-ring-rebalance-debug-pbcsx\" (UID: \"cb6a3d56-01ef-47fc-a633-6f7c4f69f27f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pbcsx" Mar 07 15:18:09 crc kubenswrapper[4943]: I0307 15:18:09.400202 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/cb6a3d56-01ef-47fc-a633-6f7c4f69f27f-swiftconf\") pod \"swift-ring-rebalance-debug-pbcsx\" (UID: \"cb6a3d56-01ef-47fc-a633-6f7c4f69f27f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pbcsx" Mar 07 15:18:09 crc kubenswrapper[4943]: I0307 15:18:09.400518 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f74tz\" (UniqueName: \"kubernetes.io/projected/cb6a3d56-01ef-47fc-a633-6f7c4f69f27f-kube-api-access-f74tz\") pod \"swift-ring-rebalance-debug-pbcsx\" (UID: \"cb6a3d56-01ef-47fc-a633-6f7c4f69f27f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pbcsx" Mar 07 15:18:09 crc kubenswrapper[4943]: I0307 15:18:09.400619 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/cb6a3d56-01ef-47fc-a633-6f7c4f69f27f-dispersionconf\") pod \"swift-ring-rebalance-debug-pbcsx\" (UID: \"cb6a3d56-01ef-47fc-a633-6f7c4f69f27f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pbcsx" Mar 07 15:18:09 crc kubenswrapper[4943]: I0307 15:18:09.400775 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cb6a3d56-01ef-47fc-a633-6f7c4f69f27f-scripts\") pod \"swift-ring-rebalance-debug-pbcsx\" (UID: \"cb6a3d56-01ef-47fc-a633-6f7c4f69f27f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pbcsx" Mar 07 15:18:09 crc kubenswrapper[4943]: I0307 15:18:09.400848 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/cb6a3d56-01ef-47fc-a633-6f7c4f69f27f-ring-data-devices\") pod \"swift-ring-rebalance-debug-pbcsx\" (UID: \"cb6a3d56-01ef-47fc-a633-6f7c4f69f27f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pbcsx" Mar 07 15:18:09 crc kubenswrapper[4943]: I0307 15:18:09.502265 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/cb6a3d56-01ef-47fc-a633-6f7c4f69f27f-swiftconf\") pod \"swift-ring-rebalance-debug-pbcsx\" (UID: \"cb6a3d56-01ef-47fc-a633-6f7c4f69f27f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pbcsx" Mar 07 15:18:09 crc kubenswrapper[4943]: I0307 15:18:09.502371 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f74tz\" (UniqueName: \"kubernetes.io/projected/cb6a3d56-01ef-47fc-a633-6f7c4f69f27f-kube-api-access-f74tz\") pod \"swift-ring-rebalance-debug-pbcsx\" (UID: \"cb6a3d56-01ef-47fc-a633-6f7c4f69f27f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pbcsx" Mar 07 15:18:09 crc kubenswrapper[4943]: I0307 15:18:09.502416 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/cb6a3d56-01ef-47fc-a633-6f7c4f69f27f-dispersionconf\") pod \"swift-ring-rebalance-debug-pbcsx\" (UID: \"cb6a3d56-01ef-47fc-a633-6f7c4f69f27f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pbcsx" Mar 07 15:18:09 crc kubenswrapper[4943]: I0307 15:18:09.502470 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cb6a3d56-01ef-47fc-a633-6f7c4f69f27f-scripts\") pod \"swift-ring-rebalance-debug-pbcsx\" (UID: \"cb6a3d56-01ef-47fc-a633-6f7c4f69f27f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pbcsx" Mar 07 15:18:09 crc kubenswrapper[4943]: I0307 15:18:09.502509 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/cb6a3d56-01ef-47fc-a633-6f7c4f69f27f-ring-data-devices\") pod \"swift-ring-rebalance-debug-pbcsx\" (UID: \"cb6a3d56-01ef-47fc-a633-6f7c4f69f27f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pbcsx" Mar 07 15:18:09 crc kubenswrapper[4943]: I0307 15:18:09.502634 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/cb6a3d56-01ef-47fc-a633-6f7c4f69f27f-etc-swift\") pod \"swift-ring-rebalance-debug-pbcsx\" (UID: \"cb6a3d56-01ef-47fc-a633-6f7c4f69f27f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pbcsx" Mar 07 15:18:09 crc kubenswrapper[4943]: I0307 15:18:09.503695 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/cb6a3d56-01ef-47fc-a633-6f7c4f69f27f-etc-swift\") pod \"swift-ring-rebalance-debug-pbcsx\" (UID: \"cb6a3d56-01ef-47fc-a633-6f7c4f69f27f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pbcsx" Mar 07 15:18:09 crc kubenswrapper[4943]: I0307 15:18:09.504087 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cb6a3d56-01ef-47fc-a633-6f7c4f69f27f-scripts\") pod \"swift-ring-rebalance-debug-pbcsx\" (UID: \"cb6a3d56-01ef-47fc-a633-6f7c4f69f27f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pbcsx" Mar 07 15:18:09 crc kubenswrapper[4943]: I0307 15:18:09.504131 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/cb6a3d56-01ef-47fc-a633-6f7c4f69f27f-ring-data-devices\") pod \"swift-ring-rebalance-debug-pbcsx\" (UID: \"cb6a3d56-01ef-47fc-a633-6f7c4f69f27f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pbcsx" Mar 07 15:18:09 crc kubenswrapper[4943]: I0307 15:18:09.509009 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/cb6a3d56-01ef-47fc-a633-6f7c4f69f27f-swiftconf\") pod \"swift-ring-rebalance-debug-pbcsx\" (UID: \"cb6a3d56-01ef-47fc-a633-6f7c4f69f27f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pbcsx" Mar 07 15:18:09 crc kubenswrapper[4943]: I0307 15:18:09.509137 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/cb6a3d56-01ef-47fc-a633-6f7c4f69f27f-dispersionconf\") pod \"swift-ring-rebalance-debug-pbcsx\" (UID: \"cb6a3d56-01ef-47fc-a633-6f7c4f69f27f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pbcsx" Mar 07 15:18:09 crc kubenswrapper[4943]: I0307 15:18:09.535762 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f74tz\" (UniqueName: \"kubernetes.io/projected/cb6a3d56-01ef-47fc-a633-6f7c4f69f27f-kube-api-access-f74tz\") pod \"swift-ring-rebalance-debug-pbcsx\" (UID: \"cb6a3d56-01ef-47fc-a633-6f7c4f69f27f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pbcsx" Mar 07 15:18:09 crc kubenswrapper[4943]: I0307 15:18:09.544603 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pbcsx" Mar 07 15:18:10 crc kubenswrapper[4943]: I0307 15:18:10.073678 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-pbcsx"] Mar 07 15:18:10 crc kubenswrapper[4943]: I0307 15:18:10.777602 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pbcsx" event={"ID":"cb6a3d56-01ef-47fc-a633-6f7c4f69f27f","Type":"ContainerStarted","Data":"e61d64cba756b76315104336679b2c2dd348c08d8e95644f67c79dbfd8b55e13"} Mar 07 15:18:10 crc kubenswrapper[4943]: I0307 15:18:10.778021 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pbcsx" event={"ID":"cb6a3d56-01ef-47fc-a633-6f7c4f69f27f","Type":"ContainerStarted","Data":"e181f9819b1ca763a181bcad41f7bb45187fef9eefecc4829a841b324c0873bd"} Mar 07 15:18:10 crc kubenswrapper[4943]: I0307 15:18:10.801308 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pbcsx" podStartSLOduration=1.801283759 podStartE2EDuration="1.801283759s" podCreationTimestamp="2026-03-07 15:18:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:18:10.791327886 +0000 UTC m=+2332.743464424" watchObservedRunningTime="2026-03-07 15:18:10.801283759 +0000 UTC m=+2332.753420287" Mar 07 15:18:11 crc kubenswrapper[4943]: I0307 15:18:11.787396 4943 generic.go:334] "Generic (PLEG): container finished" podID="cb6a3d56-01ef-47fc-a633-6f7c4f69f27f" containerID="e61d64cba756b76315104336679b2c2dd348c08d8e95644f67c79dbfd8b55e13" exitCode=0 Mar 07 15:18:11 crc kubenswrapper[4943]: I0307 15:18:11.787452 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pbcsx" event={"ID":"cb6a3d56-01ef-47fc-a633-6f7c4f69f27f","Type":"ContainerDied","Data":"e61d64cba756b76315104336679b2c2dd348c08d8e95644f67c79dbfd8b55e13"} Mar 07 15:18:13 crc kubenswrapper[4943]: I0307 15:18:13.197176 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pbcsx" Mar 07 15:18:13 crc kubenswrapper[4943]: I0307 15:18:13.227707 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-pbcsx"] Mar 07 15:18:13 crc kubenswrapper[4943]: I0307 15:18:13.234764 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-pbcsx"] Mar 07 15:18:13 crc kubenswrapper[4943]: I0307 15:18:13.392093 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cb6a3d56-01ef-47fc-a633-6f7c4f69f27f-scripts\") pod \"cb6a3d56-01ef-47fc-a633-6f7c4f69f27f\" (UID: \"cb6a3d56-01ef-47fc-a633-6f7c4f69f27f\") " Mar 07 15:18:13 crc kubenswrapper[4943]: I0307 15:18:13.392188 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f74tz\" (UniqueName: \"kubernetes.io/projected/cb6a3d56-01ef-47fc-a633-6f7c4f69f27f-kube-api-access-f74tz\") pod \"cb6a3d56-01ef-47fc-a633-6f7c4f69f27f\" (UID: \"cb6a3d56-01ef-47fc-a633-6f7c4f69f27f\") " Mar 07 15:18:13 crc kubenswrapper[4943]: I0307 15:18:13.392225 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/cb6a3d56-01ef-47fc-a633-6f7c4f69f27f-etc-swift\") pod \"cb6a3d56-01ef-47fc-a633-6f7c4f69f27f\" (UID: \"cb6a3d56-01ef-47fc-a633-6f7c4f69f27f\") " Mar 07 15:18:13 crc kubenswrapper[4943]: I0307 15:18:13.392297 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/cb6a3d56-01ef-47fc-a633-6f7c4f69f27f-ring-data-devices\") pod \"cb6a3d56-01ef-47fc-a633-6f7c4f69f27f\" (UID: \"cb6a3d56-01ef-47fc-a633-6f7c4f69f27f\") " Mar 07 15:18:13 crc kubenswrapper[4943]: I0307 15:18:13.392325 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/cb6a3d56-01ef-47fc-a633-6f7c4f69f27f-swiftconf\") pod \"cb6a3d56-01ef-47fc-a633-6f7c4f69f27f\" (UID: \"cb6a3d56-01ef-47fc-a633-6f7c4f69f27f\") " Mar 07 15:18:13 crc kubenswrapper[4943]: I0307 15:18:13.392350 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/cb6a3d56-01ef-47fc-a633-6f7c4f69f27f-dispersionconf\") pod \"cb6a3d56-01ef-47fc-a633-6f7c4f69f27f\" (UID: \"cb6a3d56-01ef-47fc-a633-6f7c4f69f27f\") " Mar 07 15:18:13 crc kubenswrapper[4943]: I0307 15:18:13.393034 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb6a3d56-01ef-47fc-a633-6f7c4f69f27f-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "cb6a3d56-01ef-47fc-a633-6f7c4f69f27f" (UID: "cb6a3d56-01ef-47fc-a633-6f7c4f69f27f"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:18:13 crc kubenswrapper[4943]: I0307 15:18:13.393579 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cb6a3d56-01ef-47fc-a633-6f7c4f69f27f-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "cb6a3d56-01ef-47fc-a633-6f7c4f69f27f" (UID: "cb6a3d56-01ef-47fc-a633-6f7c4f69f27f"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:18:13 crc kubenswrapper[4943]: I0307 15:18:13.398771 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb6a3d56-01ef-47fc-a633-6f7c4f69f27f-kube-api-access-f74tz" (OuterVolumeSpecName: "kube-api-access-f74tz") pod "cb6a3d56-01ef-47fc-a633-6f7c4f69f27f" (UID: "cb6a3d56-01ef-47fc-a633-6f7c4f69f27f"). InnerVolumeSpecName "kube-api-access-f74tz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:18:13 crc kubenswrapper[4943]: I0307 15:18:13.412052 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cb6a3d56-01ef-47fc-a633-6f7c4f69f27f-scripts" (OuterVolumeSpecName: "scripts") pod "cb6a3d56-01ef-47fc-a633-6f7c4f69f27f" (UID: "cb6a3d56-01ef-47fc-a633-6f7c4f69f27f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:18:13 crc kubenswrapper[4943]: I0307 15:18:13.416865 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb6a3d56-01ef-47fc-a633-6f7c4f69f27f-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "cb6a3d56-01ef-47fc-a633-6f7c4f69f27f" (UID: "cb6a3d56-01ef-47fc-a633-6f7c4f69f27f"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:18:13 crc kubenswrapper[4943]: I0307 15:18:13.417147 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb6a3d56-01ef-47fc-a633-6f7c4f69f27f-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "cb6a3d56-01ef-47fc-a633-6f7c4f69f27f" (UID: "cb6a3d56-01ef-47fc-a633-6f7c4f69f27f"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:18:13 crc kubenswrapper[4943]: I0307 15:18:13.493562 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f74tz\" (UniqueName: \"kubernetes.io/projected/cb6a3d56-01ef-47fc-a633-6f7c4f69f27f-kube-api-access-f74tz\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:13 crc kubenswrapper[4943]: I0307 15:18:13.493590 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/cb6a3d56-01ef-47fc-a633-6f7c4f69f27f-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:13 crc kubenswrapper[4943]: I0307 15:18:13.493599 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/cb6a3d56-01ef-47fc-a633-6f7c4f69f27f-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:13 crc kubenswrapper[4943]: I0307 15:18:13.493608 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/cb6a3d56-01ef-47fc-a633-6f7c4f69f27f-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:13 crc kubenswrapper[4943]: I0307 15:18:13.493616 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/cb6a3d56-01ef-47fc-a633-6f7c4f69f27f-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:13 crc kubenswrapper[4943]: I0307 15:18:13.493627 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cb6a3d56-01ef-47fc-a633-6f7c4f69f27f-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:13 crc kubenswrapper[4943]: I0307 15:18:13.814811 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e181f9819b1ca763a181bcad41f7bb45187fef9eefecc4829a841b324c0873bd" Mar 07 15:18:13 crc kubenswrapper[4943]: I0307 15:18:13.814881 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pbcsx" Mar 07 15:18:14 crc kubenswrapper[4943]: I0307 15:18:14.470722 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-g47xx"] Mar 07 15:18:14 crc kubenswrapper[4943]: E0307 15:18:14.472456 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb6a3d56-01ef-47fc-a633-6f7c4f69f27f" containerName="swift-ring-rebalance" Mar 07 15:18:14 crc kubenswrapper[4943]: I0307 15:18:14.472716 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb6a3d56-01ef-47fc-a633-6f7c4f69f27f" containerName="swift-ring-rebalance" Mar 07 15:18:14 crc kubenswrapper[4943]: I0307 15:18:14.473304 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb6a3d56-01ef-47fc-a633-6f7c4f69f27f" containerName="swift-ring-rebalance" Mar 07 15:18:14 crc kubenswrapper[4943]: I0307 15:18:14.474606 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-g47xx" Mar 07 15:18:14 crc kubenswrapper[4943]: I0307 15:18:14.478393 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:18:14 crc kubenswrapper[4943]: I0307 15:18:14.479846 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-g47xx"] Mar 07 15:18:14 crc kubenswrapper[4943]: I0307 15:18:14.480460 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:18:14 crc kubenswrapper[4943]: I0307 15:18:14.609414 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5c68512c-e384-4823-8f78-9b901192498c-dispersionconf\") pod \"swift-ring-rebalance-debug-g47xx\" (UID: \"5c68512c-e384-4823-8f78-9b901192498c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-g47xx" Mar 07 15:18:14 crc kubenswrapper[4943]: I0307 15:18:14.609860 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5c68512c-e384-4823-8f78-9b901192498c-scripts\") pod \"swift-ring-rebalance-debug-g47xx\" (UID: \"5c68512c-e384-4823-8f78-9b901192498c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-g47xx" Mar 07 15:18:14 crc kubenswrapper[4943]: I0307 15:18:14.609902 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7vsr\" (UniqueName: \"kubernetes.io/projected/5c68512c-e384-4823-8f78-9b901192498c-kube-api-access-v7vsr\") pod \"swift-ring-rebalance-debug-g47xx\" (UID: \"5c68512c-e384-4823-8f78-9b901192498c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-g47xx" Mar 07 15:18:14 crc kubenswrapper[4943]: I0307 15:18:14.610008 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5c68512c-e384-4823-8f78-9b901192498c-swiftconf\") pod \"swift-ring-rebalance-debug-g47xx\" (UID: \"5c68512c-e384-4823-8f78-9b901192498c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-g47xx" Mar 07 15:18:14 crc kubenswrapper[4943]: I0307 15:18:14.610053 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5c68512c-e384-4823-8f78-9b901192498c-etc-swift\") pod \"swift-ring-rebalance-debug-g47xx\" (UID: \"5c68512c-e384-4823-8f78-9b901192498c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-g47xx" Mar 07 15:18:14 crc kubenswrapper[4943]: I0307 15:18:14.610126 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5c68512c-e384-4823-8f78-9b901192498c-ring-data-devices\") pod \"swift-ring-rebalance-debug-g47xx\" (UID: \"5c68512c-e384-4823-8f78-9b901192498c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-g47xx" Mar 07 15:18:14 crc kubenswrapper[4943]: I0307 15:18:14.711187 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5c68512c-e384-4823-8f78-9b901192498c-dispersionconf\") pod \"swift-ring-rebalance-debug-g47xx\" (UID: \"5c68512c-e384-4823-8f78-9b901192498c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-g47xx" Mar 07 15:18:14 crc kubenswrapper[4943]: I0307 15:18:14.711237 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5c68512c-e384-4823-8f78-9b901192498c-scripts\") pod \"swift-ring-rebalance-debug-g47xx\" (UID: \"5c68512c-e384-4823-8f78-9b901192498c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-g47xx" Mar 07 15:18:14 crc kubenswrapper[4943]: I0307 15:18:14.711260 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7vsr\" (UniqueName: \"kubernetes.io/projected/5c68512c-e384-4823-8f78-9b901192498c-kube-api-access-v7vsr\") pod \"swift-ring-rebalance-debug-g47xx\" (UID: \"5c68512c-e384-4823-8f78-9b901192498c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-g47xx" Mar 07 15:18:14 crc kubenswrapper[4943]: I0307 15:18:14.711321 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5c68512c-e384-4823-8f78-9b901192498c-swiftconf\") pod \"swift-ring-rebalance-debug-g47xx\" (UID: \"5c68512c-e384-4823-8f78-9b901192498c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-g47xx" Mar 07 15:18:14 crc kubenswrapper[4943]: I0307 15:18:14.711356 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5c68512c-e384-4823-8f78-9b901192498c-etc-swift\") pod \"swift-ring-rebalance-debug-g47xx\" (UID: \"5c68512c-e384-4823-8f78-9b901192498c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-g47xx" Mar 07 15:18:14 crc kubenswrapper[4943]: I0307 15:18:14.711389 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5c68512c-e384-4823-8f78-9b901192498c-ring-data-devices\") pod \"swift-ring-rebalance-debug-g47xx\" (UID: \"5c68512c-e384-4823-8f78-9b901192498c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-g47xx" Mar 07 15:18:14 crc kubenswrapper[4943]: I0307 15:18:14.712180 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5c68512c-e384-4823-8f78-9b901192498c-ring-data-devices\") pod \"swift-ring-rebalance-debug-g47xx\" (UID: \"5c68512c-e384-4823-8f78-9b901192498c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-g47xx" Mar 07 15:18:14 crc kubenswrapper[4943]: I0307 15:18:14.712426 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5c68512c-e384-4823-8f78-9b901192498c-etc-swift\") pod \"swift-ring-rebalance-debug-g47xx\" (UID: \"5c68512c-e384-4823-8f78-9b901192498c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-g47xx" Mar 07 15:18:14 crc kubenswrapper[4943]: I0307 15:18:14.712811 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5c68512c-e384-4823-8f78-9b901192498c-scripts\") pod \"swift-ring-rebalance-debug-g47xx\" (UID: \"5c68512c-e384-4823-8f78-9b901192498c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-g47xx" Mar 07 15:18:14 crc kubenswrapper[4943]: I0307 15:18:14.716375 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5c68512c-e384-4823-8f78-9b901192498c-swiftconf\") pod \"swift-ring-rebalance-debug-g47xx\" (UID: \"5c68512c-e384-4823-8f78-9b901192498c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-g47xx" Mar 07 15:18:14 crc kubenswrapper[4943]: I0307 15:18:14.716679 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5c68512c-e384-4823-8f78-9b901192498c-dispersionconf\") pod \"swift-ring-rebalance-debug-g47xx\" (UID: \"5c68512c-e384-4823-8f78-9b901192498c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-g47xx" Mar 07 15:18:14 crc kubenswrapper[4943]: I0307 15:18:14.729290 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7vsr\" (UniqueName: \"kubernetes.io/projected/5c68512c-e384-4823-8f78-9b901192498c-kube-api-access-v7vsr\") pod \"swift-ring-rebalance-debug-g47xx\" (UID: \"5c68512c-e384-4823-8f78-9b901192498c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-g47xx" Mar 07 15:18:14 crc kubenswrapper[4943]: I0307 15:18:14.775887 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb6a3d56-01ef-47fc-a633-6f7c4f69f27f" path="/var/lib/kubelet/pods/cb6a3d56-01ef-47fc-a633-6f7c4f69f27f/volumes" Mar 07 15:18:14 crc kubenswrapper[4943]: I0307 15:18:14.799241 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-g47xx" Mar 07 15:18:15 crc kubenswrapper[4943]: I0307 15:18:15.085355 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-g47xx"] Mar 07 15:18:15 crc kubenswrapper[4943]: I0307 15:18:15.842691 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-g47xx" event={"ID":"5c68512c-e384-4823-8f78-9b901192498c","Type":"ContainerStarted","Data":"962b71ed6ff3177f85e874e3b257c8bbe3283bc4efb8679922a61ae0d3afe84a"} Mar 07 15:18:15 crc kubenswrapper[4943]: I0307 15:18:15.843066 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-g47xx" event={"ID":"5c68512c-e384-4823-8f78-9b901192498c","Type":"ContainerStarted","Data":"00c61633ca3b64d22f0e5dbacfdd7626e4da190c0e41cb41ef72509d2d48f859"} Mar 07 15:18:15 crc kubenswrapper[4943]: I0307 15:18:15.865838 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-g47xx" podStartSLOduration=1.8658180039999999 podStartE2EDuration="1.865818004s" podCreationTimestamp="2026-03-07 15:18:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:18:15.865274711 +0000 UTC m=+2337.817411209" watchObservedRunningTime="2026-03-07 15:18:15.865818004 +0000 UTC m=+2337.817954522" Mar 07 15:18:16 crc kubenswrapper[4943]: I0307 15:18:16.855606 4943 generic.go:334] "Generic (PLEG): container finished" podID="5c68512c-e384-4823-8f78-9b901192498c" containerID="962b71ed6ff3177f85e874e3b257c8bbe3283bc4efb8679922a61ae0d3afe84a" exitCode=0 Mar 07 15:18:16 crc kubenswrapper[4943]: I0307 15:18:16.855680 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-g47xx" event={"ID":"5c68512c-e384-4823-8f78-9b901192498c","Type":"ContainerDied","Data":"962b71ed6ff3177f85e874e3b257c8bbe3283bc4efb8679922a61ae0d3afe84a"} Mar 07 15:18:18 crc kubenswrapper[4943]: I0307 15:18:18.225475 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-g47xx" Mar 07 15:18:18 crc kubenswrapper[4943]: I0307 15:18:18.255901 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-g47xx"] Mar 07 15:18:18 crc kubenswrapper[4943]: I0307 15:18:18.261381 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-g47xx"] Mar 07 15:18:18 crc kubenswrapper[4943]: I0307 15:18:18.365984 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5c68512c-e384-4823-8f78-9b901192498c-etc-swift\") pod \"5c68512c-e384-4823-8f78-9b901192498c\" (UID: \"5c68512c-e384-4823-8f78-9b901192498c\") " Mar 07 15:18:18 crc kubenswrapper[4943]: I0307 15:18:18.366061 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5c68512c-e384-4823-8f78-9b901192498c-scripts\") pod \"5c68512c-e384-4823-8f78-9b901192498c\" (UID: \"5c68512c-e384-4823-8f78-9b901192498c\") " Mar 07 15:18:18 crc kubenswrapper[4943]: I0307 15:18:18.366118 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v7vsr\" (UniqueName: \"kubernetes.io/projected/5c68512c-e384-4823-8f78-9b901192498c-kube-api-access-v7vsr\") pod \"5c68512c-e384-4823-8f78-9b901192498c\" (UID: \"5c68512c-e384-4823-8f78-9b901192498c\") " Mar 07 15:18:18 crc kubenswrapper[4943]: I0307 15:18:18.366142 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5c68512c-e384-4823-8f78-9b901192498c-ring-data-devices\") pod \"5c68512c-e384-4823-8f78-9b901192498c\" (UID: \"5c68512c-e384-4823-8f78-9b901192498c\") " Mar 07 15:18:18 crc kubenswrapper[4943]: I0307 15:18:18.366202 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5c68512c-e384-4823-8f78-9b901192498c-swiftconf\") pod \"5c68512c-e384-4823-8f78-9b901192498c\" (UID: \"5c68512c-e384-4823-8f78-9b901192498c\") " Mar 07 15:18:18 crc kubenswrapper[4943]: I0307 15:18:18.366232 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5c68512c-e384-4823-8f78-9b901192498c-dispersionconf\") pod \"5c68512c-e384-4823-8f78-9b901192498c\" (UID: \"5c68512c-e384-4823-8f78-9b901192498c\") " Mar 07 15:18:18 crc kubenswrapper[4943]: I0307 15:18:18.366728 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c68512c-e384-4823-8f78-9b901192498c-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "5c68512c-e384-4823-8f78-9b901192498c" (UID: "5c68512c-e384-4823-8f78-9b901192498c"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:18:18 crc kubenswrapper[4943]: I0307 15:18:18.367073 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c68512c-e384-4823-8f78-9b901192498c-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "5c68512c-e384-4823-8f78-9b901192498c" (UID: "5c68512c-e384-4823-8f78-9b901192498c"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:18:18 crc kubenswrapper[4943]: I0307 15:18:18.371648 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c68512c-e384-4823-8f78-9b901192498c-kube-api-access-v7vsr" (OuterVolumeSpecName: "kube-api-access-v7vsr") pod "5c68512c-e384-4823-8f78-9b901192498c" (UID: "5c68512c-e384-4823-8f78-9b901192498c"). InnerVolumeSpecName "kube-api-access-v7vsr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:18:18 crc kubenswrapper[4943]: I0307 15:18:18.391196 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c68512c-e384-4823-8f78-9b901192498c-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "5c68512c-e384-4823-8f78-9b901192498c" (UID: "5c68512c-e384-4823-8f78-9b901192498c"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:18:18 crc kubenswrapper[4943]: I0307 15:18:18.396251 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c68512c-e384-4823-8f78-9b901192498c-scripts" (OuterVolumeSpecName: "scripts") pod "5c68512c-e384-4823-8f78-9b901192498c" (UID: "5c68512c-e384-4823-8f78-9b901192498c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:18:18 crc kubenswrapper[4943]: I0307 15:18:18.402582 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c68512c-e384-4823-8f78-9b901192498c-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "5c68512c-e384-4823-8f78-9b901192498c" (UID: "5c68512c-e384-4823-8f78-9b901192498c"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:18:18 crc kubenswrapper[4943]: I0307 15:18:18.467971 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5c68512c-e384-4823-8f78-9b901192498c-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:18 crc kubenswrapper[4943]: I0307 15:18:18.468009 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5c68512c-e384-4823-8f78-9b901192498c-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:18 crc kubenswrapper[4943]: I0307 15:18:18.468023 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5c68512c-e384-4823-8f78-9b901192498c-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:18 crc kubenswrapper[4943]: I0307 15:18:18.468036 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5c68512c-e384-4823-8f78-9b901192498c-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:18 crc kubenswrapper[4943]: I0307 15:18:18.468050 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v7vsr\" (UniqueName: \"kubernetes.io/projected/5c68512c-e384-4823-8f78-9b901192498c-kube-api-access-v7vsr\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:18 crc kubenswrapper[4943]: I0307 15:18:18.468064 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5c68512c-e384-4823-8f78-9b901192498c-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:18 crc kubenswrapper[4943]: I0307 15:18:18.799678 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c68512c-e384-4823-8f78-9b901192498c" path="/var/lib/kubelet/pods/5c68512c-e384-4823-8f78-9b901192498c/volumes" Mar 07 15:18:18 crc kubenswrapper[4943]: I0307 15:18:18.885000 4943 scope.go:117] "RemoveContainer" containerID="962b71ed6ff3177f85e874e3b257c8bbe3283bc4efb8679922a61ae0d3afe84a" Mar 07 15:18:18 crc kubenswrapper[4943]: I0307 15:18:18.885050 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-g47xx" Mar 07 15:18:19 crc kubenswrapper[4943]: I0307 15:18:19.500887 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-8fhx9"] Mar 07 15:18:19 crc kubenswrapper[4943]: E0307 15:18:19.501815 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c68512c-e384-4823-8f78-9b901192498c" containerName="swift-ring-rebalance" Mar 07 15:18:19 crc kubenswrapper[4943]: I0307 15:18:19.501838 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c68512c-e384-4823-8f78-9b901192498c" containerName="swift-ring-rebalance" Mar 07 15:18:19 crc kubenswrapper[4943]: I0307 15:18:19.502193 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c68512c-e384-4823-8f78-9b901192498c" containerName="swift-ring-rebalance" Mar 07 15:18:19 crc kubenswrapper[4943]: I0307 15:18:19.503071 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8fhx9" Mar 07 15:18:19 crc kubenswrapper[4943]: I0307 15:18:19.505390 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:18:19 crc kubenswrapper[4943]: I0307 15:18:19.505841 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:18:19 crc kubenswrapper[4943]: I0307 15:18:19.520257 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-8fhx9"] Mar 07 15:18:19 crc kubenswrapper[4943]: I0307 15:18:19.692670 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3b59da91-de5f-4e4c-9e49-e14b2954c7f1-dispersionconf\") pod \"swift-ring-rebalance-debug-8fhx9\" (UID: \"3b59da91-de5f-4e4c-9e49-e14b2954c7f1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8fhx9" Mar 07 15:18:19 crc kubenswrapper[4943]: I0307 15:18:19.692729 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3b59da91-de5f-4e4c-9e49-e14b2954c7f1-swiftconf\") pod \"swift-ring-rebalance-debug-8fhx9\" (UID: \"3b59da91-de5f-4e4c-9e49-e14b2954c7f1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8fhx9" Mar 07 15:18:19 crc kubenswrapper[4943]: I0307 15:18:19.692764 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3b59da91-de5f-4e4c-9e49-e14b2954c7f1-scripts\") pod \"swift-ring-rebalance-debug-8fhx9\" (UID: \"3b59da91-de5f-4e4c-9e49-e14b2954c7f1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8fhx9" Mar 07 15:18:19 crc kubenswrapper[4943]: I0307 15:18:19.692797 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3b59da91-de5f-4e4c-9e49-e14b2954c7f1-ring-data-devices\") pod \"swift-ring-rebalance-debug-8fhx9\" (UID: \"3b59da91-de5f-4e4c-9e49-e14b2954c7f1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8fhx9" Mar 07 15:18:19 crc kubenswrapper[4943]: I0307 15:18:19.692894 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3b59da91-de5f-4e4c-9e49-e14b2954c7f1-etc-swift\") pod \"swift-ring-rebalance-debug-8fhx9\" (UID: \"3b59da91-de5f-4e4c-9e49-e14b2954c7f1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8fhx9" Mar 07 15:18:19 crc kubenswrapper[4943]: I0307 15:18:19.692980 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cm9l8\" (UniqueName: \"kubernetes.io/projected/3b59da91-de5f-4e4c-9e49-e14b2954c7f1-kube-api-access-cm9l8\") pod \"swift-ring-rebalance-debug-8fhx9\" (UID: \"3b59da91-de5f-4e4c-9e49-e14b2954c7f1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8fhx9" Mar 07 15:18:19 crc kubenswrapper[4943]: I0307 15:18:19.755501 4943 scope.go:117] "RemoveContainer" containerID="448308ad6e34ac922d66b3c90e66d0927407a94dd28633ec350502d69e4ca8aa" Mar 07 15:18:19 crc kubenswrapper[4943]: E0307 15:18:19.755903 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cpdmm_openshift-machine-config-operator(4c1f18ea-e557-41fc-95dd-ce4e1722a61e)\"" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" Mar 07 15:18:19 crc kubenswrapper[4943]: I0307 15:18:19.794579 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3b59da91-de5f-4e4c-9e49-e14b2954c7f1-etc-swift\") pod \"swift-ring-rebalance-debug-8fhx9\" (UID: \"3b59da91-de5f-4e4c-9e49-e14b2954c7f1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8fhx9" Mar 07 15:18:19 crc kubenswrapper[4943]: I0307 15:18:19.794698 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cm9l8\" (UniqueName: \"kubernetes.io/projected/3b59da91-de5f-4e4c-9e49-e14b2954c7f1-kube-api-access-cm9l8\") pod \"swift-ring-rebalance-debug-8fhx9\" (UID: \"3b59da91-de5f-4e4c-9e49-e14b2954c7f1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8fhx9" Mar 07 15:18:19 crc kubenswrapper[4943]: I0307 15:18:19.794854 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3b59da91-de5f-4e4c-9e49-e14b2954c7f1-dispersionconf\") pod \"swift-ring-rebalance-debug-8fhx9\" (UID: \"3b59da91-de5f-4e4c-9e49-e14b2954c7f1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8fhx9" Mar 07 15:18:19 crc kubenswrapper[4943]: I0307 15:18:19.794897 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3b59da91-de5f-4e4c-9e49-e14b2954c7f1-swiftconf\") pod \"swift-ring-rebalance-debug-8fhx9\" (UID: \"3b59da91-de5f-4e4c-9e49-e14b2954c7f1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8fhx9" Mar 07 15:18:19 crc kubenswrapper[4943]: I0307 15:18:19.794953 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3b59da91-de5f-4e4c-9e49-e14b2954c7f1-scripts\") pod \"swift-ring-rebalance-debug-8fhx9\" (UID: \"3b59da91-de5f-4e4c-9e49-e14b2954c7f1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8fhx9" Mar 07 15:18:19 crc kubenswrapper[4943]: I0307 15:18:19.794991 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3b59da91-de5f-4e4c-9e49-e14b2954c7f1-ring-data-devices\") pod \"swift-ring-rebalance-debug-8fhx9\" (UID: \"3b59da91-de5f-4e4c-9e49-e14b2954c7f1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8fhx9" Mar 07 15:18:19 crc kubenswrapper[4943]: I0307 15:18:19.796176 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3b59da91-de5f-4e4c-9e49-e14b2954c7f1-scripts\") pod \"swift-ring-rebalance-debug-8fhx9\" (UID: \"3b59da91-de5f-4e4c-9e49-e14b2954c7f1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8fhx9" Mar 07 15:18:19 crc kubenswrapper[4943]: I0307 15:18:19.796463 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3b59da91-de5f-4e4c-9e49-e14b2954c7f1-etc-swift\") pod \"swift-ring-rebalance-debug-8fhx9\" (UID: \"3b59da91-de5f-4e4c-9e49-e14b2954c7f1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8fhx9" Mar 07 15:18:19 crc kubenswrapper[4943]: I0307 15:18:19.796526 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3b59da91-de5f-4e4c-9e49-e14b2954c7f1-ring-data-devices\") pod \"swift-ring-rebalance-debug-8fhx9\" (UID: \"3b59da91-de5f-4e4c-9e49-e14b2954c7f1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8fhx9" Mar 07 15:18:19 crc kubenswrapper[4943]: I0307 15:18:19.802829 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3b59da91-de5f-4e4c-9e49-e14b2954c7f1-dispersionconf\") pod \"swift-ring-rebalance-debug-8fhx9\" (UID: \"3b59da91-de5f-4e4c-9e49-e14b2954c7f1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8fhx9" Mar 07 15:18:19 crc kubenswrapper[4943]: I0307 15:18:19.805165 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3b59da91-de5f-4e4c-9e49-e14b2954c7f1-swiftconf\") pod \"swift-ring-rebalance-debug-8fhx9\" (UID: \"3b59da91-de5f-4e4c-9e49-e14b2954c7f1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8fhx9" Mar 07 15:18:19 crc kubenswrapper[4943]: I0307 15:18:19.826801 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cm9l8\" (UniqueName: \"kubernetes.io/projected/3b59da91-de5f-4e4c-9e49-e14b2954c7f1-kube-api-access-cm9l8\") pod \"swift-ring-rebalance-debug-8fhx9\" (UID: \"3b59da91-de5f-4e4c-9e49-e14b2954c7f1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8fhx9" Mar 07 15:18:19 crc kubenswrapper[4943]: I0307 15:18:19.832710 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8fhx9" Mar 07 15:18:20 crc kubenswrapper[4943]: I0307 15:18:20.346919 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-8fhx9"] Mar 07 15:18:20 crc kubenswrapper[4943]: I0307 15:18:20.912290 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8fhx9" event={"ID":"3b59da91-de5f-4e4c-9e49-e14b2954c7f1","Type":"ContainerStarted","Data":"53bd7b59c66104a6413ec02643d74763f6805fd1bea29090d815f0fb26a9908c"} Mar 07 15:18:20 crc kubenswrapper[4943]: I0307 15:18:20.912552 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8fhx9" event={"ID":"3b59da91-de5f-4e4c-9e49-e14b2954c7f1","Type":"ContainerStarted","Data":"92d9c7e4f5bc1d24d164ebc84dffb0dad75964949f1d1ab8fad10e35e3881988"} Mar 07 15:18:21 crc kubenswrapper[4943]: I0307 15:18:21.927053 4943 generic.go:334] "Generic (PLEG): container finished" podID="3b59da91-de5f-4e4c-9e49-e14b2954c7f1" containerID="53bd7b59c66104a6413ec02643d74763f6805fd1bea29090d815f0fb26a9908c" exitCode=0 Mar 07 15:18:21 crc kubenswrapper[4943]: I0307 15:18:21.927099 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8fhx9" event={"ID":"3b59da91-de5f-4e4c-9e49-e14b2954c7f1","Type":"ContainerDied","Data":"53bd7b59c66104a6413ec02643d74763f6805fd1bea29090d815f0fb26a9908c"} Mar 07 15:18:22 crc kubenswrapper[4943]: I0307 15:18:22.914402 4943 scope.go:117] "RemoveContainer" containerID="c1f1429cb6c9119934cd1f73e5430200a689739be373733fd9c4eac0c91fcd3c" Mar 07 15:18:22 crc kubenswrapper[4943]: I0307 15:18:22.945918 4943 scope.go:117] "RemoveContainer" containerID="40334ac85a352f802ad7e5593220d3e8811b459855f74de6c90b7bb16b7077e3" Mar 07 15:18:22 crc kubenswrapper[4943]: I0307 15:18:22.994841 4943 scope.go:117] "RemoveContainer" containerID="001de98844014410af24877aeeb22718020bffccafe294976e9da66e536d469c" Mar 07 15:18:23 crc kubenswrapper[4943]: I0307 15:18:23.042768 4943 scope.go:117] "RemoveContainer" containerID="dd1d4566544d3718923d8b461f0a50398b77407a748b0fab8c31f40fa2ebe405" Mar 07 15:18:23 crc kubenswrapper[4943]: I0307 15:18:23.067168 4943 scope.go:117] "RemoveContainer" containerID="8bdb2c800d876fccdead65d0ca37c85a454cd13b12e2418466d5f9d9667fa317" Mar 07 15:18:23 crc kubenswrapper[4943]: I0307 15:18:23.118713 4943 scope.go:117] "RemoveContainer" containerID="43aa2e98bb78522353924a86521f72f9a0b8723d89b7db6e30ef90d1cb0e943f" Mar 07 15:18:23 crc kubenswrapper[4943]: I0307 15:18:23.154176 4943 scope.go:117] "RemoveContainer" containerID="cd807c94f54160ad472a6127781bd7d1138328a5f52b97d1acfb48a8bfecf050" Mar 07 15:18:23 crc kubenswrapper[4943]: I0307 15:18:23.183347 4943 scope.go:117] "RemoveContainer" containerID="d92a84d136eafdb41db47adc5e8b097377d0e0eb5b3d35e93f3d3b6af516f440" Mar 07 15:18:23 crc kubenswrapper[4943]: I0307 15:18:23.204508 4943 scope.go:117] "RemoveContainer" containerID="c6a6c74af6e301b6fccd01a45947763df009f2d805ee884c52a266e7fe4cdfc0" Mar 07 15:18:23 crc kubenswrapper[4943]: I0307 15:18:23.225868 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8fhx9" Mar 07 15:18:23 crc kubenswrapper[4943]: I0307 15:18:23.249589 4943 scope.go:117] "RemoveContainer" containerID="ead602cb4d8d3d03b05cfcbf8fcb86519261f201698044c7f52a83582c4c3d7e" Mar 07 15:18:23 crc kubenswrapper[4943]: I0307 15:18:23.267535 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-8fhx9"] Mar 07 15:18:23 crc kubenswrapper[4943]: I0307 15:18:23.273457 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-8fhx9"] Mar 07 15:18:23 crc kubenswrapper[4943]: I0307 15:18:23.273942 4943 scope.go:117] "RemoveContainer" containerID="0ef227a2835e6837b378f6818163dba489e1445ccb0365b2e4f3711ce07e264d" Mar 07 15:18:23 crc kubenswrapper[4943]: I0307 15:18:23.357098 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3b59da91-de5f-4e4c-9e49-e14b2954c7f1-dispersionconf\") pod \"3b59da91-de5f-4e4c-9e49-e14b2954c7f1\" (UID: \"3b59da91-de5f-4e4c-9e49-e14b2954c7f1\") " Mar 07 15:18:23 crc kubenswrapper[4943]: I0307 15:18:23.357166 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3b59da91-de5f-4e4c-9e49-e14b2954c7f1-etc-swift\") pod \"3b59da91-de5f-4e4c-9e49-e14b2954c7f1\" (UID: \"3b59da91-de5f-4e4c-9e49-e14b2954c7f1\") " Mar 07 15:18:23 crc kubenswrapper[4943]: I0307 15:18:23.357300 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3b59da91-de5f-4e4c-9e49-e14b2954c7f1-scripts\") pod \"3b59da91-de5f-4e4c-9e49-e14b2954c7f1\" (UID: \"3b59da91-de5f-4e4c-9e49-e14b2954c7f1\") " Mar 07 15:18:23 crc kubenswrapper[4943]: I0307 15:18:23.357430 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3b59da91-de5f-4e4c-9e49-e14b2954c7f1-ring-data-devices\") pod \"3b59da91-de5f-4e4c-9e49-e14b2954c7f1\" (UID: \"3b59da91-de5f-4e4c-9e49-e14b2954c7f1\") " Mar 07 15:18:23 crc kubenswrapper[4943]: I0307 15:18:23.357469 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3b59da91-de5f-4e4c-9e49-e14b2954c7f1-swiftconf\") pod \"3b59da91-de5f-4e4c-9e49-e14b2954c7f1\" (UID: \"3b59da91-de5f-4e4c-9e49-e14b2954c7f1\") " Mar 07 15:18:23 crc kubenswrapper[4943]: I0307 15:18:23.357551 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cm9l8\" (UniqueName: \"kubernetes.io/projected/3b59da91-de5f-4e4c-9e49-e14b2954c7f1-kube-api-access-cm9l8\") pod \"3b59da91-de5f-4e4c-9e49-e14b2954c7f1\" (UID: \"3b59da91-de5f-4e4c-9e49-e14b2954c7f1\") " Mar 07 15:18:23 crc kubenswrapper[4943]: I0307 15:18:23.358205 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b59da91-de5f-4e4c-9e49-e14b2954c7f1-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "3b59da91-de5f-4e4c-9e49-e14b2954c7f1" (UID: "3b59da91-de5f-4e4c-9e49-e14b2954c7f1"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:18:23 crc kubenswrapper[4943]: I0307 15:18:23.358299 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b59da91-de5f-4e4c-9e49-e14b2954c7f1-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "3b59da91-de5f-4e4c-9e49-e14b2954c7f1" (UID: "3b59da91-de5f-4e4c-9e49-e14b2954c7f1"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:18:23 crc kubenswrapper[4943]: I0307 15:18:23.363140 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b59da91-de5f-4e4c-9e49-e14b2954c7f1-kube-api-access-cm9l8" (OuterVolumeSpecName: "kube-api-access-cm9l8") pod "3b59da91-de5f-4e4c-9e49-e14b2954c7f1" (UID: "3b59da91-de5f-4e4c-9e49-e14b2954c7f1"). InnerVolumeSpecName "kube-api-access-cm9l8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:18:23 crc kubenswrapper[4943]: I0307 15:18:23.380592 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b59da91-de5f-4e4c-9e49-e14b2954c7f1-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "3b59da91-de5f-4e4c-9e49-e14b2954c7f1" (UID: "3b59da91-de5f-4e4c-9e49-e14b2954c7f1"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:18:23 crc kubenswrapper[4943]: I0307 15:18:23.385867 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b59da91-de5f-4e4c-9e49-e14b2954c7f1-scripts" (OuterVolumeSpecName: "scripts") pod "3b59da91-de5f-4e4c-9e49-e14b2954c7f1" (UID: "3b59da91-de5f-4e4c-9e49-e14b2954c7f1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:18:23 crc kubenswrapper[4943]: I0307 15:18:23.390178 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b59da91-de5f-4e4c-9e49-e14b2954c7f1-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "3b59da91-de5f-4e4c-9e49-e14b2954c7f1" (UID: "3b59da91-de5f-4e4c-9e49-e14b2954c7f1"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:18:23 crc kubenswrapper[4943]: I0307 15:18:23.460233 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3b59da91-de5f-4e4c-9e49-e14b2954c7f1-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:23 crc kubenswrapper[4943]: I0307 15:18:23.460315 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3b59da91-de5f-4e4c-9e49-e14b2954c7f1-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:23 crc kubenswrapper[4943]: I0307 15:18:23.460337 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3b59da91-de5f-4e4c-9e49-e14b2954c7f1-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:23 crc kubenswrapper[4943]: I0307 15:18:23.460356 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3b59da91-de5f-4e4c-9e49-e14b2954c7f1-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:23 crc kubenswrapper[4943]: I0307 15:18:23.460376 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3b59da91-de5f-4e4c-9e49-e14b2954c7f1-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:23 crc kubenswrapper[4943]: I0307 15:18:23.460394 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cm9l8\" (UniqueName: \"kubernetes.io/projected/3b59da91-de5f-4e4c-9e49-e14b2954c7f1-kube-api-access-cm9l8\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:23 crc kubenswrapper[4943]: I0307 15:18:23.960041 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="92d9c7e4f5bc1d24d164ebc84dffb0dad75964949f1d1ab8fad10e35e3881988" Mar 07 15:18:23 crc kubenswrapper[4943]: I0307 15:18:23.960114 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8fhx9" Mar 07 15:18:24 crc kubenswrapper[4943]: I0307 15:18:24.497789 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-clr4h"] Mar 07 15:18:24 crc kubenswrapper[4943]: E0307 15:18:24.499294 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b59da91-de5f-4e4c-9e49-e14b2954c7f1" containerName="swift-ring-rebalance" Mar 07 15:18:24 crc kubenswrapper[4943]: I0307 15:18:24.499363 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b59da91-de5f-4e4c-9e49-e14b2954c7f1" containerName="swift-ring-rebalance" Mar 07 15:18:24 crc kubenswrapper[4943]: I0307 15:18:24.500021 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b59da91-de5f-4e4c-9e49-e14b2954c7f1" containerName="swift-ring-rebalance" Mar 07 15:18:24 crc kubenswrapper[4943]: I0307 15:18:24.502125 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-clr4h" Mar 07 15:18:24 crc kubenswrapper[4943]: I0307 15:18:24.509278 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:18:24 crc kubenswrapper[4943]: I0307 15:18:24.509540 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:18:24 crc kubenswrapper[4943]: I0307 15:18:24.516520 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-clr4h"] Mar 07 15:18:24 crc kubenswrapper[4943]: I0307 15:18:24.683197 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/48d6f728-3804-4641-83c3-cc03fecf876c-swiftconf\") pod \"swift-ring-rebalance-debug-clr4h\" (UID: \"48d6f728-3804-4641-83c3-cc03fecf876c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-clr4h" Mar 07 15:18:24 crc kubenswrapper[4943]: I0307 15:18:24.683313 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/48d6f728-3804-4641-83c3-cc03fecf876c-scripts\") pod \"swift-ring-rebalance-debug-clr4h\" (UID: \"48d6f728-3804-4641-83c3-cc03fecf876c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-clr4h" Mar 07 15:18:24 crc kubenswrapper[4943]: I0307 15:18:24.683417 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/48d6f728-3804-4641-83c3-cc03fecf876c-dispersionconf\") pod \"swift-ring-rebalance-debug-clr4h\" (UID: \"48d6f728-3804-4641-83c3-cc03fecf876c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-clr4h" Mar 07 15:18:24 crc kubenswrapper[4943]: I0307 15:18:24.683514 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/48d6f728-3804-4641-83c3-cc03fecf876c-etc-swift\") pod \"swift-ring-rebalance-debug-clr4h\" (UID: \"48d6f728-3804-4641-83c3-cc03fecf876c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-clr4h" Mar 07 15:18:24 crc kubenswrapper[4943]: I0307 15:18:24.683547 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkvhn\" (UniqueName: \"kubernetes.io/projected/48d6f728-3804-4641-83c3-cc03fecf876c-kube-api-access-kkvhn\") pod \"swift-ring-rebalance-debug-clr4h\" (UID: \"48d6f728-3804-4641-83c3-cc03fecf876c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-clr4h" Mar 07 15:18:24 crc kubenswrapper[4943]: I0307 15:18:24.683621 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/48d6f728-3804-4641-83c3-cc03fecf876c-ring-data-devices\") pod \"swift-ring-rebalance-debug-clr4h\" (UID: \"48d6f728-3804-4641-83c3-cc03fecf876c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-clr4h" Mar 07 15:18:24 crc kubenswrapper[4943]: I0307 15:18:24.774015 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b59da91-de5f-4e4c-9e49-e14b2954c7f1" path="/var/lib/kubelet/pods/3b59da91-de5f-4e4c-9e49-e14b2954c7f1/volumes" Mar 07 15:18:24 crc kubenswrapper[4943]: I0307 15:18:24.786263 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/48d6f728-3804-4641-83c3-cc03fecf876c-swiftconf\") pod \"swift-ring-rebalance-debug-clr4h\" (UID: \"48d6f728-3804-4641-83c3-cc03fecf876c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-clr4h" Mar 07 15:18:24 crc kubenswrapper[4943]: I0307 15:18:24.786345 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/48d6f728-3804-4641-83c3-cc03fecf876c-scripts\") pod \"swift-ring-rebalance-debug-clr4h\" (UID: \"48d6f728-3804-4641-83c3-cc03fecf876c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-clr4h" Mar 07 15:18:24 crc kubenswrapper[4943]: I0307 15:18:24.786427 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/48d6f728-3804-4641-83c3-cc03fecf876c-dispersionconf\") pod \"swift-ring-rebalance-debug-clr4h\" (UID: \"48d6f728-3804-4641-83c3-cc03fecf876c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-clr4h" Mar 07 15:18:24 crc kubenswrapper[4943]: I0307 15:18:24.786497 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/48d6f728-3804-4641-83c3-cc03fecf876c-etc-swift\") pod \"swift-ring-rebalance-debug-clr4h\" (UID: \"48d6f728-3804-4641-83c3-cc03fecf876c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-clr4h" Mar 07 15:18:24 crc kubenswrapper[4943]: I0307 15:18:24.786567 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkvhn\" (UniqueName: \"kubernetes.io/projected/48d6f728-3804-4641-83c3-cc03fecf876c-kube-api-access-kkvhn\") pod \"swift-ring-rebalance-debug-clr4h\" (UID: \"48d6f728-3804-4641-83c3-cc03fecf876c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-clr4h" Mar 07 15:18:24 crc kubenswrapper[4943]: I0307 15:18:24.787105 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/48d6f728-3804-4641-83c3-cc03fecf876c-ring-data-devices\") pod \"swift-ring-rebalance-debug-clr4h\" (UID: \"48d6f728-3804-4641-83c3-cc03fecf876c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-clr4h" Mar 07 15:18:24 crc kubenswrapper[4943]: I0307 15:18:24.788438 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/48d6f728-3804-4641-83c3-cc03fecf876c-scripts\") pod \"swift-ring-rebalance-debug-clr4h\" (UID: \"48d6f728-3804-4641-83c3-cc03fecf876c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-clr4h" Mar 07 15:18:24 crc kubenswrapper[4943]: I0307 15:18:24.788777 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/48d6f728-3804-4641-83c3-cc03fecf876c-etc-swift\") pod \"swift-ring-rebalance-debug-clr4h\" (UID: \"48d6f728-3804-4641-83c3-cc03fecf876c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-clr4h" Mar 07 15:18:24 crc kubenswrapper[4943]: I0307 15:18:24.789248 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/48d6f728-3804-4641-83c3-cc03fecf876c-ring-data-devices\") pod \"swift-ring-rebalance-debug-clr4h\" (UID: \"48d6f728-3804-4641-83c3-cc03fecf876c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-clr4h" Mar 07 15:18:24 crc kubenswrapper[4943]: I0307 15:18:24.793366 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/48d6f728-3804-4641-83c3-cc03fecf876c-swiftconf\") pod \"swift-ring-rebalance-debug-clr4h\" (UID: \"48d6f728-3804-4641-83c3-cc03fecf876c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-clr4h" Mar 07 15:18:24 crc kubenswrapper[4943]: I0307 15:18:24.796649 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/48d6f728-3804-4641-83c3-cc03fecf876c-dispersionconf\") pod \"swift-ring-rebalance-debug-clr4h\" (UID: \"48d6f728-3804-4641-83c3-cc03fecf876c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-clr4h" Mar 07 15:18:24 crc kubenswrapper[4943]: I0307 15:18:24.822419 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkvhn\" (UniqueName: \"kubernetes.io/projected/48d6f728-3804-4641-83c3-cc03fecf876c-kube-api-access-kkvhn\") pod \"swift-ring-rebalance-debug-clr4h\" (UID: \"48d6f728-3804-4641-83c3-cc03fecf876c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-clr4h" Mar 07 15:18:24 crc kubenswrapper[4943]: I0307 15:18:24.871172 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-clr4h" Mar 07 15:18:25 crc kubenswrapper[4943]: I0307 15:18:25.360314 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-clr4h"] Mar 07 15:18:25 crc kubenswrapper[4943]: I0307 15:18:25.984654 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-clr4h" event={"ID":"48d6f728-3804-4641-83c3-cc03fecf876c","Type":"ContainerStarted","Data":"5620eaa598fdd1630ed4dc50ba057eb8a142ba1b8779e247a8255f1fe3663b21"} Mar 07 15:18:25 crc kubenswrapper[4943]: I0307 15:18:25.985173 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-clr4h" event={"ID":"48d6f728-3804-4641-83c3-cc03fecf876c","Type":"ContainerStarted","Data":"d895369fe1012ae90cdeb610bd74a13ac266b4e2362be79918fc643cff0e641a"} Mar 07 15:18:26 crc kubenswrapper[4943]: I0307 15:18:26.011738 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-clr4h" podStartSLOduration=2.011718437 podStartE2EDuration="2.011718437s" podCreationTimestamp="2026-03-07 15:18:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:18:26.000538103 +0000 UTC m=+2347.952674632" watchObservedRunningTime="2026-03-07 15:18:26.011718437 +0000 UTC m=+2347.963854945" Mar 07 15:18:26 crc kubenswrapper[4943]: I0307 15:18:26.999108 4943 generic.go:334] "Generic (PLEG): container finished" podID="48d6f728-3804-4641-83c3-cc03fecf876c" containerID="5620eaa598fdd1630ed4dc50ba057eb8a142ba1b8779e247a8255f1fe3663b21" exitCode=0 Mar 07 15:18:26 crc kubenswrapper[4943]: I0307 15:18:26.999246 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-clr4h" event={"ID":"48d6f728-3804-4641-83c3-cc03fecf876c","Type":"ContainerDied","Data":"5620eaa598fdd1630ed4dc50ba057eb8a142ba1b8779e247a8255f1fe3663b21"} Mar 07 15:18:28 crc kubenswrapper[4943]: I0307 15:18:28.408738 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-clr4h" Mar 07 15:18:28 crc kubenswrapper[4943]: I0307 15:18:28.455425 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-clr4h"] Mar 07 15:18:28 crc kubenswrapper[4943]: I0307 15:18:28.463145 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-clr4h"] Mar 07 15:18:28 crc kubenswrapper[4943]: I0307 15:18:28.553282 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/48d6f728-3804-4641-83c3-cc03fecf876c-scripts\") pod \"48d6f728-3804-4641-83c3-cc03fecf876c\" (UID: \"48d6f728-3804-4641-83c3-cc03fecf876c\") " Mar 07 15:18:28 crc kubenswrapper[4943]: I0307 15:18:28.553408 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kkvhn\" (UniqueName: \"kubernetes.io/projected/48d6f728-3804-4641-83c3-cc03fecf876c-kube-api-access-kkvhn\") pod \"48d6f728-3804-4641-83c3-cc03fecf876c\" (UID: \"48d6f728-3804-4641-83c3-cc03fecf876c\") " Mar 07 15:18:28 crc kubenswrapper[4943]: I0307 15:18:28.553449 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/48d6f728-3804-4641-83c3-cc03fecf876c-ring-data-devices\") pod \"48d6f728-3804-4641-83c3-cc03fecf876c\" (UID: \"48d6f728-3804-4641-83c3-cc03fecf876c\") " Mar 07 15:18:28 crc kubenswrapper[4943]: I0307 15:18:28.553474 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/48d6f728-3804-4641-83c3-cc03fecf876c-dispersionconf\") pod \"48d6f728-3804-4641-83c3-cc03fecf876c\" (UID: \"48d6f728-3804-4641-83c3-cc03fecf876c\") " Mar 07 15:18:28 crc kubenswrapper[4943]: I0307 15:18:28.553537 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/48d6f728-3804-4641-83c3-cc03fecf876c-swiftconf\") pod \"48d6f728-3804-4641-83c3-cc03fecf876c\" (UID: \"48d6f728-3804-4641-83c3-cc03fecf876c\") " Mar 07 15:18:28 crc kubenswrapper[4943]: I0307 15:18:28.553626 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/48d6f728-3804-4641-83c3-cc03fecf876c-etc-swift\") pod \"48d6f728-3804-4641-83c3-cc03fecf876c\" (UID: \"48d6f728-3804-4641-83c3-cc03fecf876c\") " Mar 07 15:18:28 crc kubenswrapper[4943]: I0307 15:18:28.554474 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48d6f728-3804-4641-83c3-cc03fecf876c-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "48d6f728-3804-4641-83c3-cc03fecf876c" (UID: "48d6f728-3804-4641-83c3-cc03fecf876c"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:18:28 crc kubenswrapper[4943]: I0307 15:18:28.554832 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48d6f728-3804-4641-83c3-cc03fecf876c-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "48d6f728-3804-4641-83c3-cc03fecf876c" (UID: "48d6f728-3804-4641-83c3-cc03fecf876c"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:18:28 crc kubenswrapper[4943]: I0307 15:18:28.560975 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48d6f728-3804-4641-83c3-cc03fecf876c-kube-api-access-kkvhn" (OuterVolumeSpecName: "kube-api-access-kkvhn") pod "48d6f728-3804-4641-83c3-cc03fecf876c" (UID: "48d6f728-3804-4641-83c3-cc03fecf876c"). InnerVolumeSpecName "kube-api-access-kkvhn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:18:28 crc kubenswrapper[4943]: I0307 15:18:28.583845 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48d6f728-3804-4641-83c3-cc03fecf876c-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "48d6f728-3804-4641-83c3-cc03fecf876c" (UID: "48d6f728-3804-4641-83c3-cc03fecf876c"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:18:28 crc kubenswrapper[4943]: I0307 15:18:28.593177 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48d6f728-3804-4641-83c3-cc03fecf876c-scripts" (OuterVolumeSpecName: "scripts") pod "48d6f728-3804-4641-83c3-cc03fecf876c" (UID: "48d6f728-3804-4641-83c3-cc03fecf876c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:18:28 crc kubenswrapper[4943]: I0307 15:18:28.599035 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48d6f728-3804-4641-83c3-cc03fecf876c-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "48d6f728-3804-4641-83c3-cc03fecf876c" (UID: "48d6f728-3804-4641-83c3-cc03fecf876c"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:18:28 crc kubenswrapper[4943]: I0307 15:18:28.655840 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kkvhn\" (UniqueName: \"kubernetes.io/projected/48d6f728-3804-4641-83c3-cc03fecf876c-kube-api-access-kkvhn\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:28 crc kubenswrapper[4943]: I0307 15:18:28.656207 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/48d6f728-3804-4641-83c3-cc03fecf876c-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:28 crc kubenswrapper[4943]: I0307 15:18:28.656356 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/48d6f728-3804-4641-83c3-cc03fecf876c-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:28 crc kubenswrapper[4943]: I0307 15:18:28.656481 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/48d6f728-3804-4641-83c3-cc03fecf876c-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:28 crc kubenswrapper[4943]: I0307 15:18:28.656599 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/48d6f728-3804-4641-83c3-cc03fecf876c-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:28 crc kubenswrapper[4943]: I0307 15:18:28.656756 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/48d6f728-3804-4641-83c3-cc03fecf876c-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:28 crc kubenswrapper[4943]: I0307 15:18:28.769308 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48d6f728-3804-4641-83c3-cc03fecf876c" path="/var/lib/kubelet/pods/48d6f728-3804-4641-83c3-cc03fecf876c/volumes" Mar 07 15:18:29 crc kubenswrapper[4943]: I0307 15:18:29.025083 4943 scope.go:117] "RemoveContainer" containerID="5620eaa598fdd1630ed4dc50ba057eb8a142ba1b8779e247a8255f1fe3663b21" Mar 07 15:18:29 crc kubenswrapper[4943]: I0307 15:18:29.025176 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-clr4h" Mar 07 15:18:29 crc kubenswrapper[4943]: I0307 15:18:29.593175 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-5mts7"] Mar 07 15:18:29 crc kubenswrapper[4943]: E0307 15:18:29.593622 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48d6f728-3804-4641-83c3-cc03fecf876c" containerName="swift-ring-rebalance" Mar 07 15:18:29 crc kubenswrapper[4943]: I0307 15:18:29.593644 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="48d6f728-3804-4641-83c3-cc03fecf876c" containerName="swift-ring-rebalance" Mar 07 15:18:29 crc kubenswrapper[4943]: I0307 15:18:29.593978 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="48d6f728-3804-4641-83c3-cc03fecf876c" containerName="swift-ring-rebalance" Mar 07 15:18:29 crc kubenswrapper[4943]: I0307 15:18:29.594757 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mts7" Mar 07 15:18:29 crc kubenswrapper[4943]: I0307 15:18:29.596350 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:18:29 crc kubenswrapper[4943]: I0307 15:18:29.597313 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:18:29 crc kubenswrapper[4943]: I0307 15:18:29.609623 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-5mts7"] Mar 07 15:18:29 crc kubenswrapper[4943]: I0307 15:18:29.773077 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1a320825-b71e-41bf-a37e-d20340013873-dispersionconf\") pod \"swift-ring-rebalance-debug-5mts7\" (UID: \"1a320825-b71e-41bf-a37e-d20340013873\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mts7" Mar 07 15:18:29 crc kubenswrapper[4943]: I0307 15:18:29.773156 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1a320825-b71e-41bf-a37e-d20340013873-etc-swift\") pod \"swift-ring-rebalance-debug-5mts7\" (UID: \"1a320825-b71e-41bf-a37e-d20340013873\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mts7" Mar 07 15:18:29 crc kubenswrapper[4943]: I0307 15:18:29.773232 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1a320825-b71e-41bf-a37e-d20340013873-swiftconf\") pod \"swift-ring-rebalance-debug-5mts7\" (UID: \"1a320825-b71e-41bf-a37e-d20340013873\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mts7" Mar 07 15:18:29 crc kubenswrapper[4943]: I0307 15:18:29.773264 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1a320825-b71e-41bf-a37e-d20340013873-ring-data-devices\") pod \"swift-ring-rebalance-debug-5mts7\" (UID: \"1a320825-b71e-41bf-a37e-d20340013873\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mts7" Mar 07 15:18:29 crc kubenswrapper[4943]: I0307 15:18:29.773339 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smqvk\" (UniqueName: \"kubernetes.io/projected/1a320825-b71e-41bf-a37e-d20340013873-kube-api-access-smqvk\") pod \"swift-ring-rebalance-debug-5mts7\" (UID: \"1a320825-b71e-41bf-a37e-d20340013873\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mts7" Mar 07 15:18:29 crc kubenswrapper[4943]: I0307 15:18:29.773491 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1a320825-b71e-41bf-a37e-d20340013873-scripts\") pod \"swift-ring-rebalance-debug-5mts7\" (UID: \"1a320825-b71e-41bf-a37e-d20340013873\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mts7" Mar 07 15:18:29 crc kubenswrapper[4943]: I0307 15:18:29.875434 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1a320825-b71e-41bf-a37e-d20340013873-dispersionconf\") pod \"swift-ring-rebalance-debug-5mts7\" (UID: \"1a320825-b71e-41bf-a37e-d20340013873\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mts7" Mar 07 15:18:29 crc kubenswrapper[4943]: I0307 15:18:29.875515 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1a320825-b71e-41bf-a37e-d20340013873-etc-swift\") pod \"swift-ring-rebalance-debug-5mts7\" (UID: \"1a320825-b71e-41bf-a37e-d20340013873\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mts7" Mar 07 15:18:29 crc kubenswrapper[4943]: I0307 15:18:29.876283 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1a320825-b71e-41bf-a37e-d20340013873-etc-swift\") pod \"swift-ring-rebalance-debug-5mts7\" (UID: \"1a320825-b71e-41bf-a37e-d20340013873\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mts7" Mar 07 15:18:29 crc kubenswrapper[4943]: I0307 15:18:29.876396 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1a320825-b71e-41bf-a37e-d20340013873-ring-data-devices\") pod \"swift-ring-rebalance-debug-5mts7\" (UID: \"1a320825-b71e-41bf-a37e-d20340013873\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mts7" Mar 07 15:18:29 crc kubenswrapper[4943]: I0307 15:18:29.877182 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1a320825-b71e-41bf-a37e-d20340013873-swiftconf\") pod \"swift-ring-rebalance-debug-5mts7\" (UID: \"1a320825-b71e-41bf-a37e-d20340013873\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mts7" Mar 07 15:18:29 crc kubenswrapper[4943]: I0307 15:18:29.877327 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smqvk\" (UniqueName: \"kubernetes.io/projected/1a320825-b71e-41bf-a37e-d20340013873-kube-api-access-smqvk\") pod \"swift-ring-rebalance-debug-5mts7\" (UID: \"1a320825-b71e-41bf-a37e-d20340013873\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mts7" Mar 07 15:18:29 crc kubenswrapper[4943]: I0307 15:18:29.877444 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1a320825-b71e-41bf-a37e-d20340013873-ring-data-devices\") pod \"swift-ring-rebalance-debug-5mts7\" (UID: \"1a320825-b71e-41bf-a37e-d20340013873\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mts7" Mar 07 15:18:29 crc kubenswrapper[4943]: I0307 15:18:29.877571 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1a320825-b71e-41bf-a37e-d20340013873-scripts\") pod \"swift-ring-rebalance-debug-5mts7\" (UID: \"1a320825-b71e-41bf-a37e-d20340013873\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mts7" Mar 07 15:18:29 crc kubenswrapper[4943]: I0307 15:18:29.878248 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1a320825-b71e-41bf-a37e-d20340013873-scripts\") pod \"swift-ring-rebalance-debug-5mts7\" (UID: \"1a320825-b71e-41bf-a37e-d20340013873\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mts7" Mar 07 15:18:29 crc kubenswrapper[4943]: I0307 15:18:29.882633 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1a320825-b71e-41bf-a37e-d20340013873-swiftconf\") pod \"swift-ring-rebalance-debug-5mts7\" (UID: \"1a320825-b71e-41bf-a37e-d20340013873\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mts7" Mar 07 15:18:29 crc kubenswrapper[4943]: I0307 15:18:29.885112 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1a320825-b71e-41bf-a37e-d20340013873-dispersionconf\") pod \"swift-ring-rebalance-debug-5mts7\" (UID: \"1a320825-b71e-41bf-a37e-d20340013873\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mts7" Mar 07 15:18:29 crc kubenswrapper[4943]: I0307 15:18:29.911835 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smqvk\" (UniqueName: \"kubernetes.io/projected/1a320825-b71e-41bf-a37e-d20340013873-kube-api-access-smqvk\") pod \"swift-ring-rebalance-debug-5mts7\" (UID: \"1a320825-b71e-41bf-a37e-d20340013873\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mts7" Mar 07 15:18:30 crc kubenswrapper[4943]: I0307 15:18:30.210723 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mts7" Mar 07 15:18:30 crc kubenswrapper[4943]: I0307 15:18:30.507252 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-5mts7"] Mar 07 15:18:30 crc kubenswrapper[4943]: I0307 15:18:30.755597 4943 scope.go:117] "RemoveContainer" containerID="448308ad6e34ac922d66b3c90e66d0927407a94dd28633ec350502d69e4ca8aa" Mar 07 15:18:30 crc kubenswrapper[4943]: E0307 15:18:30.756194 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cpdmm_openshift-machine-config-operator(4c1f18ea-e557-41fc-95dd-ce4e1722a61e)\"" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" Mar 07 15:18:31 crc kubenswrapper[4943]: I0307 15:18:31.050273 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mts7" event={"ID":"1a320825-b71e-41bf-a37e-d20340013873","Type":"ContainerStarted","Data":"05d36441cf236aeb49863eea9d387746e7ed71f540dd8f004d6b7dadb5fc62aa"} Mar 07 15:18:31 crc kubenswrapper[4943]: I0307 15:18:31.050337 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mts7" event={"ID":"1a320825-b71e-41bf-a37e-d20340013873","Type":"ContainerStarted","Data":"94d61b8721a42268de6e6b0ca0b7b19e915d6bc90d03d5cb7c3e0a0465cca066"} Mar 07 15:18:31 crc kubenswrapper[4943]: I0307 15:18:31.074614 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mts7" podStartSLOduration=2.074587632 podStartE2EDuration="2.074587632s" podCreationTimestamp="2026-03-07 15:18:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:18:31.06878653 +0000 UTC m=+2353.020923058" watchObservedRunningTime="2026-03-07 15:18:31.074587632 +0000 UTC m=+2353.026724160" Mar 07 15:18:32 crc kubenswrapper[4943]: I0307 15:18:32.062585 4943 generic.go:334] "Generic (PLEG): container finished" podID="1a320825-b71e-41bf-a37e-d20340013873" containerID="05d36441cf236aeb49863eea9d387746e7ed71f540dd8f004d6b7dadb5fc62aa" exitCode=0 Mar 07 15:18:32 crc kubenswrapper[4943]: I0307 15:18:32.062679 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mts7" event={"ID":"1a320825-b71e-41bf-a37e-d20340013873","Type":"ContainerDied","Data":"05d36441cf236aeb49863eea9d387746e7ed71f540dd8f004d6b7dadb5fc62aa"} Mar 07 15:18:33 crc kubenswrapper[4943]: I0307 15:18:33.405800 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mts7" Mar 07 15:18:33 crc kubenswrapper[4943]: I0307 15:18:33.453983 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-5mts7"] Mar 07 15:18:33 crc kubenswrapper[4943]: I0307 15:18:33.458655 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-5mts7"] Mar 07 15:18:33 crc kubenswrapper[4943]: I0307 15:18:33.600451 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1a320825-b71e-41bf-a37e-d20340013873-etc-swift\") pod \"1a320825-b71e-41bf-a37e-d20340013873\" (UID: \"1a320825-b71e-41bf-a37e-d20340013873\") " Mar 07 15:18:33 crc kubenswrapper[4943]: I0307 15:18:33.600547 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-smqvk\" (UniqueName: \"kubernetes.io/projected/1a320825-b71e-41bf-a37e-d20340013873-kube-api-access-smqvk\") pod \"1a320825-b71e-41bf-a37e-d20340013873\" (UID: \"1a320825-b71e-41bf-a37e-d20340013873\") " Mar 07 15:18:33 crc kubenswrapper[4943]: I0307 15:18:33.600591 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1a320825-b71e-41bf-a37e-d20340013873-swiftconf\") pod \"1a320825-b71e-41bf-a37e-d20340013873\" (UID: \"1a320825-b71e-41bf-a37e-d20340013873\") " Mar 07 15:18:33 crc kubenswrapper[4943]: I0307 15:18:33.600695 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1a320825-b71e-41bf-a37e-d20340013873-ring-data-devices\") pod \"1a320825-b71e-41bf-a37e-d20340013873\" (UID: \"1a320825-b71e-41bf-a37e-d20340013873\") " Mar 07 15:18:33 crc kubenswrapper[4943]: I0307 15:18:33.600726 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1a320825-b71e-41bf-a37e-d20340013873-dispersionconf\") pod \"1a320825-b71e-41bf-a37e-d20340013873\" (UID: \"1a320825-b71e-41bf-a37e-d20340013873\") " Mar 07 15:18:33 crc kubenswrapper[4943]: I0307 15:18:33.600767 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1a320825-b71e-41bf-a37e-d20340013873-scripts\") pod \"1a320825-b71e-41bf-a37e-d20340013873\" (UID: \"1a320825-b71e-41bf-a37e-d20340013873\") " Mar 07 15:18:33 crc kubenswrapper[4943]: I0307 15:18:33.601387 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a320825-b71e-41bf-a37e-d20340013873-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "1a320825-b71e-41bf-a37e-d20340013873" (UID: "1a320825-b71e-41bf-a37e-d20340013873"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:18:33 crc kubenswrapper[4943]: I0307 15:18:33.602034 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a320825-b71e-41bf-a37e-d20340013873-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "1a320825-b71e-41bf-a37e-d20340013873" (UID: "1a320825-b71e-41bf-a37e-d20340013873"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:18:33 crc kubenswrapper[4943]: I0307 15:18:33.607079 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a320825-b71e-41bf-a37e-d20340013873-kube-api-access-smqvk" (OuterVolumeSpecName: "kube-api-access-smqvk") pod "1a320825-b71e-41bf-a37e-d20340013873" (UID: "1a320825-b71e-41bf-a37e-d20340013873"). InnerVolumeSpecName "kube-api-access-smqvk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:18:33 crc kubenswrapper[4943]: I0307 15:18:33.633541 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a320825-b71e-41bf-a37e-d20340013873-scripts" (OuterVolumeSpecName: "scripts") pod "1a320825-b71e-41bf-a37e-d20340013873" (UID: "1a320825-b71e-41bf-a37e-d20340013873"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:18:33 crc kubenswrapper[4943]: I0307 15:18:33.638212 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a320825-b71e-41bf-a37e-d20340013873-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "1a320825-b71e-41bf-a37e-d20340013873" (UID: "1a320825-b71e-41bf-a37e-d20340013873"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:18:33 crc kubenswrapper[4943]: I0307 15:18:33.641895 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a320825-b71e-41bf-a37e-d20340013873-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "1a320825-b71e-41bf-a37e-d20340013873" (UID: "1a320825-b71e-41bf-a37e-d20340013873"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:18:33 crc kubenswrapper[4943]: I0307 15:18:33.702755 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1a320825-b71e-41bf-a37e-d20340013873-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:33 crc kubenswrapper[4943]: I0307 15:18:33.702815 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-smqvk\" (UniqueName: \"kubernetes.io/projected/1a320825-b71e-41bf-a37e-d20340013873-kube-api-access-smqvk\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:33 crc kubenswrapper[4943]: I0307 15:18:33.702841 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1a320825-b71e-41bf-a37e-d20340013873-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:33 crc kubenswrapper[4943]: I0307 15:18:33.702859 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1a320825-b71e-41bf-a37e-d20340013873-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:33 crc kubenswrapper[4943]: I0307 15:18:33.702881 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1a320825-b71e-41bf-a37e-d20340013873-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:33 crc kubenswrapper[4943]: I0307 15:18:33.702903 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1a320825-b71e-41bf-a37e-d20340013873-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:34 crc kubenswrapper[4943]: I0307 15:18:34.083663 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="94d61b8721a42268de6e6b0ca0b7b19e915d6bc90d03d5cb7c3e0a0465cca066" Mar 07 15:18:34 crc kubenswrapper[4943]: I0307 15:18:34.083775 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5mts7" Mar 07 15:18:34 crc kubenswrapper[4943]: I0307 15:18:34.620013 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-pl45z"] Mar 07 15:18:34 crc kubenswrapper[4943]: E0307 15:18:34.620669 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a320825-b71e-41bf-a37e-d20340013873" containerName="swift-ring-rebalance" Mar 07 15:18:34 crc kubenswrapper[4943]: I0307 15:18:34.620697 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a320825-b71e-41bf-a37e-d20340013873" containerName="swift-ring-rebalance" Mar 07 15:18:34 crc kubenswrapper[4943]: I0307 15:18:34.621019 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a320825-b71e-41bf-a37e-d20340013873" containerName="swift-ring-rebalance" Mar 07 15:18:34 crc kubenswrapper[4943]: I0307 15:18:34.621867 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pl45z" Mar 07 15:18:34 crc kubenswrapper[4943]: I0307 15:18:34.625839 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:18:34 crc kubenswrapper[4943]: I0307 15:18:34.630030 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-pl45z"] Mar 07 15:18:34 crc kubenswrapper[4943]: I0307 15:18:34.631101 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:18:34 crc kubenswrapper[4943]: I0307 15:18:34.784618 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a320825-b71e-41bf-a37e-d20340013873" path="/var/lib/kubelet/pods/1a320825-b71e-41bf-a37e-d20340013873/volumes" Mar 07 15:18:34 crc kubenswrapper[4943]: I0307 15:18:34.821334 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bd3f1480-73d4-4b51-a42e-e116d8169352-swiftconf\") pod \"swift-ring-rebalance-debug-pl45z\" (UID: \"bd3f1480-73d4-4b51-a42e-e116d8169352\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pl45z" Mar 07 15:18:34 crc kubenswrapper[4943]: I0307 15:18:34.821584 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6tbk\" (UniqueName: \"kubernetes.io/projected/bd3f1480-73d4-4b51-a42e-e116d8169352-kube-api-access-h6tbk\") pod \"swift-ring-rebalance-debug-pl45z\" (UID: \"bd3f1480-73d4-4b51-a42e-e116d8169352\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pl45z" Mar 07 15:18:34 crc kubenswrapper[4943]: I0307 15:18:34.821634 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bd3f1480-73d4-4b51-a42e-e116d8169352-ring-data-devices\") pod \"swift-ring-rebalance-debug-pl45z\" (UID: \"bd3f1480-73d4-4b51-a42e-e116d8169352\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pl45z" Mar 07 15:18:34 crc kubenswrapper[4943]: I0307 15:18:34.821719 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bd3f1480-73d4-4b51-a42e-e116d8169352-dispersionconf\") pod \"swift-ring-rebalance-debug-pl45z\" (UID: \"bd3f1480-73d4-4b51-a42e-e116d8169352\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pl45z" Mar 07 15:18:34 crc kubenswrapper[4943]: I0307 15:18:34.821778 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bd3f1480-73d4-4b51-a42e-e116d8169352-etc-swift\") pod \"swift-ring-rebalance-debug-pl45z\" (UID: \"bd3f1480-73d4-4b51-a42e-e116d8169352\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pl45z" Mar 07 15:18:34 crc kubenswrapper[4943]: I0307 15:18:34.821890 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bd3f1480-73d4-4b51-a42e-e116d8169352-scripts\") pod \"swift-ring-rebalance-debug-pl45z\" (UID: \"bd3f1480-73d4-4b51-a42e-e116d8169352\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pl45z" Mar 07 15:18:34 crc kubenswrapper[4943]: I0307 15:18:34.923266 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bd3f1480-73d4-4b51-a42e-e116d8169352-scripts\") pod \"swift-ring-rebalance-debug-pl45z\" (UID: \"bd3f1480-73d4-4b51-a42e-e116d8169352\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pl45z" Mar 07 15:18:34 crc kubenswrapper[4943]: I0307 15:18:34.923451 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bd3f1480-73d4-4b51-a42e-e116d8169352-swiftconf\") pod \"swift-ring-rebalance-debug-pl45z\" (UID: \"bd3f1480-73d4-4b51-a42e-e116d8169352\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pl45z" Mar 07 15:18:34 crc kubenswrapper[4943]: I0307 15:18:34.923550 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6tbk\" (UniqueName: \"kubernetes.io/projected/bd3f1480-73d4-4b51-a42e-e116d8169352-kube-api-access-h6tbk\") pod \"swift-ring-rebalance-debug-pl45z\" (UID: \"bd3f1480-73d4-4b51-a42e-e116d8169352\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pl45z" Mar 07 15:18:34 crc kubenswrapper[4943]: I0307 15:18:34.923587 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bd3f1480-73d4-4b51-a42e-e116d8169352-ring-data-devices\") pod \"swift-ring-rebalance-debug-pl45z\" (UID: \"bd3f1480-73d4-4b51-a42e-e116d8169352\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pl45z" Mar 07 15:18:34 crc kubenswrapper[4943]: I0307 15:18:34.923640 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bd3f1480-73d4-4b51-a42e-e116d8169352-dispersionconf\") pod \"swift-ring-rebalance-debug-pl45z\" (UID: \"bd3f1480-73d4-4b51-a42e-e116d8169352\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pl45z" Mar 07 15:18:34 crc kubenswrapper[4943]: I0307 15:18:34.923672 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bd3f1480-73d4-4b51-a42e-e116d8169352-etc-swift\") pod \"swift-ring-rebalance-debug-pl45z\" (UID: \"bd3f1480-73d4-4b51-a42e-e116d8169352\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pl45z" Mar 07 15:18:34 crc kubenswrapper[4943]: I0307 15:18:34.924185 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bd3f1480-73d4-4b51-a42e-e116d8169352-scripts\") pod \"swift-ring-rebalance-debug-pl45z\" (UID: \"bd3f1480-73d4-4b51-a42e-e116d8169352\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pl45z" Mar 07 15:18:34 crc kubenswrapper[4943]: I0307 15:18:34.924295 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bd3f1480-73d4-4b51-a42e-e116d8169352-etc-swift\") pod \"swift-ring-rebalance-debug-pl45z\" (UID: \"bd3f1480-73d4-4b51-a42e-e116d8169352\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pl45z" Mar 07 15:18:34 crc kubenswrapper[4943]: I0307 15:18:34.925228 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bd3f1480-73d4-4b51-a42e-e116d8169352-ring-data-devices\") pod \"swift-ring-rebalance-debug-pl45z\" (UID: \"bd3f1480-73d4-4b51-a42e-e116d8169352\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pl45z" Mar 07 15:18:34 crc kubenswrapper[4943]: I0307 15:18:34.933989 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bd3f1480-73d4-4b51-a42e-e116d8169352-dispersionconf\") pod \"swift-ring-rebalance-debug-pl45z\" (UID: \"bd3f1480-73d4-4b51-a42e-e116d8169352\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pl45z" Mar 07 15:18:34 crc kubenswrapper[4943]: I0307 15:18:34.935370 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bd3f1480-73d4-4b51-a42e-e116d8169352-swiftconf\") pod \"swift-ring-rebalance-debug-pl45z\" (UID: \"bd3f1480-73d4-4b51-a42e-e116d8169352\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pl45z" Mar 07 15:18:34 crc kubenswrapper[4943]: I0307 15:18:34.952464 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6tbk\" (UniqueName: \"kubernetes.io/projected/bd3f1480-73d4-4b51-a42e-e116d8169352-kube-api-access-h6tbk\") pod \"swift-ring-rebalance-debug-pl45z\" (UID: \"bd3f1480-73d4-4b51-a42e-e116d8169352\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pl45z" Mar 07 15:18:34 crc kubenswrapper[4943]: I0307 15:18:34.954130 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pl45z" Mar 07 15:18:35 crc kubenswrapper[4943]: I0307 15:18:35.204580 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-pl45z"] Mar 07 15:18:35 crc kubenswrapper[4943]: W0307 15:18:35.213214 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbd3f1480_73d4_4b51_a42e_e116d8169352.slice/crio-3371e3f5e4c93c68bf959d9a8814837aa6e473a65aa63bd67950075c2b7ba3b8 WatchSource:0}: Error finding container 3371e3f5e4c93c68bf959d9a8814837aa6e473a65aa63bd67950075c2b7ba3b8: Status 404 returned error can't find the container with id 3371e3f5e4c93c68bf959d9a8814837aa6e473a65aa63bd67950075c2b7ba3b8 Mar 07 15:18:36 crc kubenswrapper[4943]: I0307 15:18:36.116443 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pl45z" event={"ID":"bd3f1480-73d4-4b51-a42e-e116d8169352","Type":"ContainerStarted","Data":"6b8315a5d9fc7805d4d6da03cf5f8c8fb3b54ec78398796bb109755c3951f9a4"} Mar 07 15:18:36 crc kubenswrapper[4943]: I0307 15:18:36.116526 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pl45z" event={"ID":"bd3f1480-73d4-4b51-a42e-e116d8169352","Type":"ContainerStarted","Data":"3371e3f5e4c93c68bf959d9a8814837aa6e473a65aa63bd67950075c2b7ba3b8"} Mar 07 15:18:36 crc kubenswrapper[4943]: I0307 15:18:36.140528 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pl45z" podStartSLOduration=2.140510762 podStartE2EDuration="2.140510762s" podCreationTimestamp="2026-03-07 15:18:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:18:36.138111214 +0000 UTC m=+2358.090247712" watchObservedRunningTime="2026-03-07 15:18:36.140510762 +0000 UTC m=+2358.092647260" Mar 07 15:18:37 crc kubenswrapper[4943]: I0307 15:18:37.137591 4943 generic.go:334] "Generic (PLEG): container finished" podID="bd3f1480-73d4-4b51-a42e-e116d8169352" containerID="6b8315a5d9fc7805d4d6da03cf5f8c8fb3b54ec78398796bb109755c3951f9a4" exitCode=0 Mar 07 15:18:37 crc kubenswrapper[4943]: I0307 15:18:37.137641 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pl45z" event={"ID":"bd3f1480-73d4-4b51-a42e-e116d8169352","Type":"ContainerDied","Data":"6b8315a5d9fc7805d4d6da03cf5f8c8fb3b54ec78398796bb109755c3951f9a4"} Mar 07 15:18:38 crc kubenswrapper[4943]: I0307 15:18:38.490405 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pl45z" Mar 07 15:18:38 crc kubenswrapper[4943]: I0307 15:18:38.526452 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-pl45z"] Mar 07 15:18:38 crc kubenswrapper[4943]: I0307 15:18:38.538365 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-pl45z"] Mar 07 15:18:38 crc kubenswrapper[4943]: I0307 15:18:38.683596 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bd3f1480-73d4-4b51-a42e-e116d8169352-dispersionconf\") pod \"bd3f1480-73d4-4b51-a42e-e116d8169352\" (UID: \"bd3f1480-73d4-4b51-a42e-e116d8169352\") " Mar 07 15:18:38 crc kubenswrapper[4943]: I0307 15:18:38.683686 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h6tbk\" (UniqueName: \"kubernetes.io/projected/bd3f1480-73d4-4b51-a42e-e116d8169352-kube-api-access-h6tbk\") pod \"bd3f1480-73d4-4b51-a42e-e116d8169352\" (UID: \"bd3f1480-73d4-4b51-a42e-e116d8169352\") " Mar 07 15:18:38 crc kubenswrapper[4943]: I0307 15:18:38.684470 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bd3f1480-73d4-4b51-a42e-e116d8169352-etc-swift\") pod \"bd3f1480-73d4-4b51-a42e-e116d8169352\" (UID: \"bd3f1480-73d4-4b51-a42e-e116d8169352\") " Mar 07 15:18:38 crc kubenswrapper[4943]: I0307 15:18:38.684503 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bd3f1480-73d4-4b51-a42e-e116d8169352-scripts\") pod \"bd3f1480-73d4-4b51-a42e-e116d8169352\" (UID: \"bd3f1480-73d4-4b51-a42e-e116d8169352\") " Mar 07 15:18:38 crc kubenswrapper[4943]: I0307 15:18:38.684592 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bd3f1480-73d4-4b51-a42e-e116d8169352-ring-data-devices\") pod \"bd3f1480-73d4-4b51-a42e-e116d8169352\" (UID: \"bd3f1480-73d4-4b51-a42e-e116d8169352\") " Mar 07 15:18:38 crc kubenswrapper[4943]: I0307 15:18:38.684674 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bd3f1480-73d4-4b51-a42e-e116d8169352-swiftconf\") pod \"bd3f1480-73d4-4b51-a42e-e116d8169352\" (UID: \"bd3f1480-73d4-4b51-a42e-e116d8169352\") " Mar 07 15:18:38 crc kubenswrapper[4943]: I0307 15:18:38.685417 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd3f1480-73d4-4b51-a42e-e116d8169352-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "bd3f1480-73d4-4b51-a42e-e116d8169352" (UID: "bd3f1480-73d4-4b51-a42e-e116d8169352"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:18:38 crc kubenswrapper[4943]: I0307 15:18:38.686065 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd3f1480-73d4-4b51-a42e-e116d8169352-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "bd3f1480-73d4-4b51-a42e-e116d8169352" (UID: "bd3f1480-73d4-4b51-a42e-e116d8169352"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:18:38 crc kubenswrapper[4943]: I0307 15:18:38.689039 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd3f1480-73d4-4b51-a42e-e116d8169352-kube-api-access-h6tbk" (OuterVolumeSpecName: "kube-api-access-h6tbk") pod "bd3f1480-73d4-4b51-a42e-e116d8169352" (UID: "bd3f1480-73d4-4b51-a42e-e116d8169352"). InnerVolumeSpecName "kube-api-access-h6tbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:18:38 crc kubenswrapper[4943]: I0307 15:18:38.710097 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd3f1480-73d4-4b51-a42e-e116d8169352-scripts" (OuterVolumeSpecName: "scripts") pod "bd3f1480-73d4-4b51-a42e-e116d8169352" (UID: "bd3f1480-73d4-4b51-a42e-e116d8169352"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:18:38 crc kubenswrapper[4943]: I0307 15:18:38.715452 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd3f1480-73d4-4b51-a42e-e116d8169352-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "bd3f1480-73d4-4b51-a42e-e116d8169352" (UID: "bd3f1480-73d4-4b51-a42e-e116d8169352"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:18:38 crc kubenswrapper[4943]: I0307 15:18:38.721988 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd3f1480-73d4-4b51-a42e-e116d8169352-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "bd3f1480-73d4-4b51-a42e-e116d8169352" (UID: "bd3f1480-73d4-4b51-a42e-e116d8169352"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:18:38 crc kubenswrapper[4943]: I0307 15:18:38.766412 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd3f1480-73d4-4b51-a42e-e116d8169352" path="/var/lib/kubelet/pods/bd3f1480-73d4-4b51-a42e-e116d8169352/volumes" Mar 07 15:18:38 crc kubenswrapper[4943]: I0307 15:18:38.786127 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bd3f1480-73d4-4b51-a42e-e116d8169352-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:38 crc kubenswrapper[4943]: I0307 15:18:38.786159 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bd3f1480-73d4-4b51-a42e-e116d8169352-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:38 crc kubenswrapper[4943]: I0307 15:18:38.786174 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h6tbk\" (UniqueName: \"kubernetes.io/projected/bd3f1480-73d4-4b51-a42e-e116d8169352-kube-api-access-h6tbk\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:38 crc kubenswrapper[4943]: I0307 15:18:38.786184 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bd3f1480-73d4-4b51-a42e-e116d8169352-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:38 crc kubenswrapper[4943]: I0307 15:18:38.786195 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bd3f1480-73d4-4b51-a42e-e116d8169352-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:38 crc kubenswrapper[4943]: I0307 15:18:38.786206 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bd3f1480-73d4-4b51-a42e-e116d8169352-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:39 crc kubenswrapper[4943]: I0307 15:18:39.165770 4943 scope.go:117] "RemoveContainer" containerID="6b8315a5d9fc7805d4d6da03cf5f8c8fb3b54ec78398796bb109755c3951f9a4" Mar 07 15:18:39 crc kubenswrapper[4943]: I0307 15:18:39.165832 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pl45z" Mar 07 15:18:39 crc kubenswrapper[4943]: I0307 15:18:39.772500 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-f9x87"] Mar 07 15:18:39 crc kubenswrapper[4943]: E0307 15:18:39.773072 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd3f1480-73d4-4b51-a42e-e116d8169352" containerName="swift-ring-rebalance" Mar 07 15:18:39 crc kubenswrapper[4943]: I0307 15:18:39.773096 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd3f1480-73d4-4b51-a42e-e116d8169352" containerName="swift-ring-rebalance" Mar 07 15:18:39 crc kubenswrapper[4943]: I0307 15:18:39.773392 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd3f1480-73d4-4b51-a42e-e116d8169352" containerName="swift-ring-rebalance" Mar 07 15:18:39 crc kubenswrapper[4943]: I0307 15:18:39.774239 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-f9x87" Mar 07 15:18:39 crc kubenswrapper[4943]: I0307 15:18:39.776483 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:18:39 crc kubenswrapper[4943]: I0307 15:18:39.778254 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:18:39 crc kubenswrapper[4943]: I0307 15:18:39.783892 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-f9x87"] Mar 07 15:18:39 crc kubenswrapper[4943]: I0307 15:18:39.905687 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/009ff594-7e6d-4469-b124-4956fc3d9b67-swiftconf\") pod \"swift-ring-rebalance-debug-f9x87\" (UID: \"009ff594-7e6d-4469-b124-4956fc3d9b67\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-f9x87" Mar 07 15:18:39 crc kubenswrapper[4943]: I0307 15:18:39.905780 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/009ff594-7e6d-4469-b124-4956fc3d9b67-dispersionconf\") pod \"swift-ring-rebalance-debug-f9x87\" (UID: \"009ff594-7e6d-4469-b124-4956fc3d9b67\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-f9x87" Mar 07 15:18:39 crc kubenswrapper[4943]: I0307 15:18:39.905856 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/009ff594-7e6d-4469-b124-4956fc3d9b67-ring-data-devices\") pod \"swift-ring-rebalance-debug-f9x87\" (UID: \"009ff594-7e6d-4469-b124-4956fc3d9b67\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-f9x87" Mar 07 15:18:39 crc kubenswrapper[4943]: I0307 15:18:39.906015 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fsfrl\" (UniqueName: \"kubernetes.io/projected/009ff594-7e6d-4469-b124-4956fc3d9b67-kube-api-access-fsfrl\") pod \"swift-ring-rebalance-debug-f9x87\" (UID: \"009ff594-7e6d-4469-b124-4956fc3d9b67\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-f9x87" Mar 07 15:18:39 crc kubenswrapper[4943]: I0307 15:18:39.906106 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/009ff594-7e6d-4469-b124-4956fc3d9b67-scripts\") pod \"swift-ring-rebalance-debug-f9x87\" (UID: \"009ff594-7e6d-4469-b124-4956fc3d9b67\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-f9x87" Mar 07 15:18:39 crc kubenswrapper[4943]: I0307 15:18:39.906151 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/009ff594-7e6d-4469-b124-4956fc3d9b67-etc-swift\") pod \"swift-ring-rebalance-debug-f9x87\" (UID: \"009ff594-7e6d-4469-b124-4956fc3d9b67\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-f9x87" Mar 07 15:18:40 crc kubenswrapper[4943]: I0307 15:18:40.007632 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/009ff594-7e6d-4469-b124-4956fc3d9b67-ring-data-devices\") pod \"swift-ring-rebalance-debug-f9x87\" (UID: \"009ff594-7e6d-4469-b124-4956fc3d9b67\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-f9x87" Mar 07 15:18:40 crc kubenswrapper[4943]: I0307 15:18:40.007795 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fsfrl\" (UniqueName: \"kubernetes.io/projected/009ff594-7e6d-4469-b124-4956fc3d9b67-kube-api-access-fsfrl\") pod \"swift-ring-rebalance-debug-f9x87\" (UID: \"009ff594-7e6d-4469-b124-4956fc3d9b67\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-f9x87" Mar 07 15:18:40 crc kubenswrapper[4943]: I0307 15:18:40.007881 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/009ff594-7e6d-4469-b124-4956fc3d9b67-scripts\") pod \"swift-ring-rebalance-debug-f9x87\" (UID: \"009ff594-7e6d-4469-b124-4956fc3d9b67\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-f9x87" Mar 07 15:18:40 crc kubenswrapper[4943]: I0307 15:18:40.007975 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/009ff594-7e6d-4469-b124-4956fc3d9b67-etc-swift\") pod \"swift-ring-rebalance-debug-f9x87\" (UID: \"009ff594-7e6d-4469-b124-4956fc3d9b67\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-f9x87" Mar 07 15:18:40 crc kubenswrapper[4943]: I0307 15:18:40.008035 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/009ff594-7e6d-4469-b124-4956fc3d9b67-swiftconf\") pod \"swift-ring-rebalance-debug-f9x87\" (UID: \"009ff594-7e6d-4469-b124-4956fc3d9b67\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-f9x87" Mar 07 15:18:40 crc kubenswrapper[4943]: I0307 15:18:40.008096 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/009ff594-7e6d-4469-b124-4956fc3d9b67-dispersionconf\") pod \"swift-ring-rebalance-debug-f9x87\" (UID: \"009ff594-7e6d-4469-b124-4956fc3d9b67\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-f9x87" Mar 07 15:18:40 crc kubenswrapper[4943]: I0307 15:18:40.008564 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/009ff594-7e6d-4469-b124-4956fc3d9b67-etc-swift\") pod \"swift-ring-rebalance-debug-f9x87\" (UID: \"009ff594-7e6d-4469-b124-4956fc3d9b67\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-f9x87" Mar 07 15:18:40 crc kubenswrapper[4943]: I0307 15:18:40.009208 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/009ff594-7e6d-4469-b124-4956fc3d9b67-ring-data-devices\") pod \"swift-ring-rebalance-debug-f9x87\" (UID: \"009ff594-7e6d-4469-b124-4956fc3d9b67\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-f9x87" Mar 07 15:18:40 crc kubenswrapper[4943]: I0307 15:18:40.009692 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/009ff594-7e6d-4469-b124-4956fc3d9b67-scripts\") pod \"swift-ring-rebalance-debug-f9x87\" (UID: \"009ff594-7e6d-4469-b124-4956fc3d9b67\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-f9x87" Mar 07 15:18:40 crc kubenswrapper[4943]: I0307 15:18:40.013114 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/009ff594-7e6d-4469-b124-4956fc3d9b67-swiftconf\") pod \"swift-ring-rebalance-debug-f9x87\" (UID: \"009ff594-7e6d-4469-b124-4956fc3d9b67\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-f9x87" Mar 07 15:18:40 crc kubenswrapper[4943]: I0307 15:18:40.013175 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/009ff594-7e6d-4469-b124-4956fc3d9b67-dispersionconf\") pod \"swift-ring-rebalance-debug-f9x87\" (UID: \"009ff594-7e6d-4469-b124-4956fc3d9b67\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-f9x87" Mar 07 15:18:40 crc kubenswrapper[4943]: I0307 15:18:40.034850 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fsfrl\" (UniqueName: \"kubernetes.io/projected/009ff594-7e6d-4469-b124-4956fc3d9b67-kube-api-access-fsfrl\") pod \"swift-ring-rebalance-debug-f9x87\" (UID: \"009ff594-7e6d-4469-b124-4956fc3d9b67\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-f9x87" Mar 07 15:18:40 crc kubenswrapper[4943]: I0307 15:18:40.141182 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-f9x87" Mar 07 15:18:40 crc kubenswrapper[4943]: I0307 15:18:40.667673 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-f9x87"] Mar 07 15:18:41 crc kubenswrapper[4943]: I0307 15:18:41.214779 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-f9x87" event={"ID":"009ff594-7e6d-4469-b124-4956fc3d9b67","Type":"ContainerStarted","Data":"4308b9d76705a4bd10464643ad4935fbd249c9d3db40036c52aa167d82157339"} Mar 07 15:18:41 crc kubenswrapper[4943]: I0307 15:18:41.215060 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-f9x87" event={"ID":"009ff594-7e6d-4469-b124-4956fc3d9b67","Type":"ContainerStarted","Data":"6f2c4d98f0861b8384f8d94a2ad9cbbcb3172dea3645f404edf766bf859f73b0"} Mar 07 15:18:41 crc kubenswrapper[4943]: I0307 15:18:41.235543 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-f9x87" podStartSLOduration=2.235527712 podStartE2EDuration="2.235527712s" podCreationTimestamp="2026-03-07 15:18:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:18:41.2301356 +0000 UTC m=+2363.182272108" watchObservedRunningTime="2026-03-07 15:18:41.235527712 +0000 UTC m=+2363.187664210" Mar 07 15:18:42 crc kubenswrapper[4943]: I0307 15:18:42.235650 4943 generic.go:334] "Generic (PLEG): container finished" podID="009ff594-7e6d-4469-b124-4956fc3d9b67" containerID="4308b9d76705a4bd10464643ad4935fbd249c9d3db40036c52aa167d82157339" exitCode=0 Mar 07 15:18:42 crc kubenswrapper[4943]: I0307 15:18:42.235741 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-f9x87" event={"ID":"009ff594-7e6d-4469-b124-4956fc3d9b67","Type":"ContainerDied","Data":"4308b9d76705a4bd10464643ad4935fbd249c9d3db40036c52aa167d82157339"} Mar 07 15:18:43 crc kubenswrapper[4943]: I0307 15:18:43.642544 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-f9x87" Mar 07 15:18:43 crc kubenswrapper[4943]: I0307 15:18:43.690479 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-f9x87"] Mar 07 15:18:43 crc kubenswrapper[4943]: I0307 15:18:43.690540 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-f9x87"] Mar 07 15:18:43 crc kubenswrapper[4943]: I0307 15:18:43.776448 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/009ff594-7e6d-4469-b124-4956fc3d9b67-swiftconf\") pod \"009ff594-7e6d-4469-b124-4956fc3d9b67\" (UID: \"009ff594-7e6d-4469-b124-4956fc3d9b67\") " Mar 07 15:18:43 crc kubenswrapper[4943]: I0307 15:18:43.776650 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fsfrl\" (UniqueName: \"kubernetes.io/projected/009ff594-7e6d-4469-b124-4956fc3d9b67-kube-api-access-fsfrl\") pod \"009ff594-7e6d-4469-b124-4956fc3d9b67\" (UID: \"009ff594-7e6d-4469-b124-4956fc3d9b67\") " Mar 07 15:18:43 crc kubenswrapper[4943]: I0307 15:18:43.776732 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/009ff594-7e6d-4469-b124-4956fc3d9b67-etc-swift\") pod \"009ff594-7e6d-4469-b124-4956fc3d9b67\" (UID: \"009ff594-7e6d-4469-b124-4956fc3d9b67\") " Mar 07 15:18:43 crc kubenswrapper[4943]: I0307 15:18:43.776827 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/009ff594-7e6d-4469-b124-4956fc3d9b67-dispersionconf\") pod \"009ff594-7e6d-4469-b124-4956fc3d9b67\" (UID: \"009ff594-7e6d-4469-b124-4956fc3d9b67\") " Mar 07 15:18:43 crc kubenswrapper[4943]: I0307 15:18:43.776873 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/009ff594-7e6d-4469-b124-4956fc3d9b67-ring-data-devices\") pod \"009ff594-7e6d-4469-b124-4956fc3d9b67\" (UID: \"009ff594-7e6d-4469-b124-4956fc3d9b67\") " Mar 07 15:18:43 crc kubenswrapper[4943]: I0307 15:18:43.777228 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/009ff594-7e6d-4469-b124-4956fc3d9b67-scripts\") pod \"009ff594-7e6d-4469-b124-4956fc3d9b67\" (UID: \"009ff594-7e6d-4469-b124-4956fc3d9b67\") " Mar 07 15:18:43 crc kubenswrapper[4943]: I0307 15:18:43.777552 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/009ff594-7e6d-4469-b124-4956fc3d9b67-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "009ff594-7e6d-4469-b124-4956fc3d9b67" (UID: "009ff594-7e6d-4469-b124-4956fc3d9b67"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:18:43 crc kubenswrapper[4943]: I0307 15:18:43.777675 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/009ff594-7e6d-4469-b124-4956fc3d9b67-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "009ff594-7e6d-4469-b124-4956fc3d9b67" (UID: "009ff594-7e6d-4469-b124-4956fc3d9b67"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:18:43 crc kubenswrapper[4943]: I0307 15:18:43.778335 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/009ff594-7e6d-4469-b124-4956fc3d9b67-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:43 crc kubenswrapper[4943]: I0307 15:18:43.778366 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/009ff594-7e6d-4469-b124-4956fc3d9b67-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:43 crc kubenswrapper[4943]: I0307 15:18:43.784298 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/009ff594-7e6d-4469-b124-4956fc3d9b67-kube-api-access-fsfrl" (OuterVolumeSpecName: "kube-api-access-fsfrl") pod "009ff594-7e6d-4469-b124-4956fc3d9b67" (UID: "009ff594-7e6d-4469-b124-4956fc3d9b67"). InnerVolumeSpecName "kube-api-access-fsfrl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:18:43 crc kubenswrapper[4943]: I0307 15:18:43.798534 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/009ff594-7e6d-4469-b124-4956fc3d9b67-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "009ff594-7e6d-4469-b124-4956fc3d9b67" (UID: "009ff594-7e6d-4469-b124-4956fc3d9b67"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:18:43 crc kubenswrapper[4943]: I0307 15:18:43.801306 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/009ff594-7e6d-4469-b124-4956fc3d9b67-scripts" (OuterVolumeSpecName: "scripts") pod "009ff594-7e6d-4469-b124-4956fc3d9b67" (UID: "009ff594-7e6d-4469-b124-4956fc3d9b67"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:18:43 crc kubenswrapper[4943]: I0307 15:18:43.804462 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/009ff594-7e6d-4469-b124-4956fc3d9b67-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "009ff594-7e6d-4469-b124-4956fc3d9b67" (UID: "009ff594-7e6d-4469-b124-4956fc3d9b67"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:18:43 crc kubenswrapper[4943]: I0307 15:18:43.879957 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fsfrl\" (UniqueName: \"kubernetes.io/projected/009ff594-7e6d-4469-b124-4956fc3d9b67-kube-api-access-fsfrl\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:43 crc kubenswrapper[4943]: I0307 15:18:43.879984 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/009ff594-7e6d-4469-b124-4956fc3d9b67-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:43 crc kubenswrapper[4943]: I0307 15:18:43.880018 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/009ff594-7e6d-4469-b124-4956fc3d9b67-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:43 crc kubenswrapper[4943]: I0307 15:18:43.880029 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/009ff594-7e6d-4469-b124-4956fc3d9b67-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:44 crc kubenswrapper[4943]: I0307 15:18:44.271862 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6f2c4d98f0861b8384f8d94a2ad9cbbcb3172dea3645f404edf766bf859f73b0" Mar 07 15:18:44 crc kubenswrapper[4943]: I0307 15:18:44.272024 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-f9x87" Mar 07 15:18:44 crc kubenswrapper[4943]: I0307 15:18:44.769476 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="009ff594-7e6d-4469-b124-4956fc3d9b67" path="/var/lib/kubelet/pods/009ff594-7e6d-4469-b124-4956fc3d9b67/volumes" Mar 07 15:18:44 crc kubenswrapper[4943]: I0307 15:18:44.881132 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-jk89d"] Mar 07 15:18:44 crc kubenswrapper[4943]: E0307 15:18:44.881556 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="009ff594-7e6d-4469-b124-4956fc3d9b67" containerName="swift-ring-rebalance" Mar 07 15:18:44 crc kubenswrapper[4943]: I0307 15:18:44.881584 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="009ff594-7e6d-4469-b124-4956fc3d9b67" containerName="swift-ring-rebalance" Mar 07 15:18:44 crc kubenswrapper[4943]: I0307 15:18:44.881908 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="009ff594-7e6d-4469-b124-4956fc3d9b67" containerName="swift-ring-rebalance" Mar 07 15:18:44 crc kubenswrapper[4943]: I0307 15:18:44.882763 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-jk89d" Mar 07 15:18:44 crc kubenswrapper[4943]: I0307 15:18:44.886101 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:18:44 crc kubenswrapper[4943]: I0307 15:18:44.888236 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:18:44 crc kubenswrapper[4943]: I0307 15:18:44.940202 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-jk89d"] Mar 07 15:18:44 crc kubenswrapper[4943]: I0307 15:18:44.996771 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6eee3b09-feb6-4a12-8fda-86657b069a23-swiftconf\") pod \"swift-ring-rebalance-debug-jk89d\" (UID: \"6eee3b09-feb6-4a12-8fda-86657b069a23\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jk89d" Mar 07 15:18:44 crc kubenswrapper[4943]: I0307 15:18:44.996865 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6eee3b09-feb6-4a12-8fda-86657b069a23-ring-data-devices\") pod \"swift-ring-rebalance-debug-jk89d\" (UID: \"6eee3b09-feb6-4a12-8fda-86657b069a23\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jk89d" Mar 07 15:18:44 crc kubenswrapper[4943]: I0307 15:18:44.996924 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6eee3b09-feb6-4a12-8fda-86657b069a23-dispersionconf\") pod \"swift-ring-rebalance-debug-jk89d\" (UID: \"6eee3b09-feb6-4a12-8fda-86657b069a23\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jk89d" Mar 07 15:18:44 crc kubenswrapper[4943]: I0307 15:18:44.998061 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6eee3b09-feb6-4a12-8fda-86657b069a23-scripts\") pod \"swift-ring-rebalance-debug-jk89d\" (UID: \"6eee3b09-feb6-4a12-8fda-86657b069a23\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jk89d" Mar 07 15:18:44 crc kubenswrapper[4943]: I0307 15:18:44.998172 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6eee3b09-feb6-4a12-8fda-86657b069a23-etc-swift\") pod \"swift-ring-rebalance-debug-jk89d\" (UID: \"6eee3b09-feb6-4a12-8fda-86657b069a23\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jk89d" Mar 07 15:18:44 crc kubenswrapper[4943]: I0307 15:18:44.998195 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqn46\" (UniqueName: \"kubernetes.io/projected/6eee3b09-feb6-4a12-8fda-86657b069a23-kube-api-access-sqn46\") pod \"swift-ring-rebalance-debug-jk89d\" (UID: \"6eee3b09-feb6-4a12-8fda-86657b069a23\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jk89d" Mar 07 15:18:45 crc kubenswrapper[4943]: I0307 15:18:45.099297 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6eee3b09-feb6-4a12-8fda-86657b069a23-etc-swift\") pod \"swift-ring-rebalance-debug-jk89d\" (UID: \"6eee3b09-feb6-4a12-8fda-86657b069a23\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jk89d" Mar 07 15:18:45 crc kubenswrapper[4943]: I0307 15:18:45.099355 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqn46\" (UniqueName: \"kubernetes.io/projected/6eee3b09-feb6-4a12-8fda-86657b069a23-kube-api-access-sqn46\") pod \"swift-ring-rebalance-debug-jk89d\" (UID: \"6eee3b09-feb6-4a12-8fda-86657b069a23\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jk89d" Mar 07 15:18:45 crc kubenswrapper[4943]: I0307 15:18:45.099426 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6eee3b09-feb6-4a12-8fda-86657b069a23-swiftconf\") pod \"swift-ring-rebalance-debug-jk89d\" (UID: \"6eee3b09-feb6-4a12-8fda-86657b069a23\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jk89d" Mar 07 15:18:45 crc kubenswrapper[4943]: I0307 15:18:45.099478 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6eee3b09-feb6-4a12-8fda-86657b069a23-ring-data-devices\") pod \"swift-ring-rebalance-debug-jk89d\" (UID: \"6eee3b09-feb6-4a12-8fda-86657b069a23\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jk89d" Mar 07 15:18:45 crc kubenswrapper[4943]: I0307 15:18:45.099524 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6eee3b09-feb6-4a12-8fda-86657b069a23-dispersionconf\") pod \"swift-ring-rebalance-debug-jk89d\" (UID: \"6eee3b09-feb6-4a12-8fda-86657b069a23\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jk89d" Mar 07 15:18:45 crc kubenswrapper[4943]: I0307 15:18:45.099613 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6eee3b09-feb6-4a12-8fda-86657b069a23-scripts\") pod \"swift-ring-rebalance-debug-jk89d\" (UID: \"6eee3b09-feb6-4a12-8fda-86657b069a23\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jk89d" Mar 07 15:18:45 crc kubenswrapper[4943]: I0307 15:18:45.099663 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6eee3b09-feb6-4a12-8fda-86657b069a23-etc-swift\") pod \"swift-ring-rebalance-debug-jk89d\" (UID: \"6eee3b09-feb6-4a12-8fda-86657b069a23\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jk89d" Mar 07 15:18:45 crc kubenswrapper[4943]: I0307 15:18:45.100755 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6eee3b09-feb6-4a12-8fda-86657b069a23-ring-data-devices\") pod \"swift-ring-rebalance-debug-jk89d\" (UID: \"6eee3b09-feb6-4a12-8fda-86657b069a23\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jk89d" Mar 07 15:18:45 crc kubenswrapper[4943]: I0307 15:18:45.100794 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6eee3b09-feb6-4a12-8fda-86657b069a23-scripts\") pod \"swift-ring-rebalance-debug-jk89d\" (UID: \"6eee3b09-feb6-4a12-8fda-86657b069a23\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jk89d" Mar 07 15:18:45 crc kubenswrapper[4943]: I0307 15:18:45.106162 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6eee3b09-feb6-4a12-8fda-86657b069a23-dispersionconf\") pod \"swift-ring-rebalance-debug-jk89d\" (UID: \"6eee3b09-feb6-4a12-8fda-86657b069a23\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jk89d" Mar 07 15:18:45 crc kubenswrapper[4943]: I0307 15:18:45.108356 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6eee3b09-feb6-4a12-8fda-86657b069a23-swiftconf\") pod \"swift-ring-rebalance-debug-jk89d\" (UID: \"6eee3b09-feb6-4a12-8fda-86657b069a23\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jk89d" Mar 07 15:18:45 crc kubenswrapper[4943]: I0307 15:18:45.121283 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqn46\" (UniqueName: \"kubernetes.io/projected/6eee3b09-feb6-4a12-8fda-86657b069a23-kube-api-access-sqn46\") pod \"swift-ring-rebalance-debug-jk89d\" (UID: \"6eee3b09-feb6-4a12-8fda-86657b069a23\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jk89d" Mar 07 15:18:45 crc kubenswrapper[4943]: I0307 15:18:45.200833 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-jk89d" Mar 07 15:18:45 crc kubenswrapper[4943]: I0307 15:18:45.441167 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-jk89d"] Mar 07 15:18:45 crc kubenswrapper[4943]: W0307 15:18:45.447748 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6eee3b09_feb6_4a12_8fda_86657b069a23.slice/crio-85d1f11b45d17d9d8fdc47caf741fe64f039411c986319bbcb1eb3994706e1f6 WatchSource:0}: Error finding container 85d1f11b45d17d9d8fdc47caf741fe64f039411c986319bbcb1eb3994706e1f6: Status 404 returned error can't find the container with id 85d1f11b45d17d9d8fdc47caf741fe64f039411c986319bbcb1eb3994706e1f6 Mar 07 15:18:45 crc kubenswrapper[4943]: I0307 15:18:45.755726 4943 scope.go:117] "RemoveContainer" containerID="448308ad6e34ac922d66b3c90e66d0927407a94dd28633ec350502d69e4ca8aa" Mar 07 15:18:45 crc kubenswrapper[4943]: E0307 15:18:45.755962 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cpdmm_openshift-machine-config-operator(4c1f18ea-e557-41fc-95dd-ce4e1722a61e)\"" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" Mar 07 15:18:46 crc kubenswrapper[4943]: I0307 15:18:46.291018 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-jk89d" event={"ID":"6eee3b09-feb6-4a12-8fda-86657b069a23","Type":"ContainerStarted","Data":"ff82858ed1f603b577448411127acd4be9f0a58204f4854dbb8a8a225fc783fc"} Mar 07 15:18:46 crc kubenswrapper[4943]: I0307 15:18:46.292992 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-jk89d" event={"ID":"6eee3b09-feb6-4a12-8fda-86657b069a23","Type":"ContainerStarted","Data":"85d1f11b45d17d9d8fdc47caf741fe64f039411c986319bbcb1eb3994706e1f6"} Mar 07 15:18:46 crc kubenswrapper[4943]: I0307 15:18:46.319925 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-jk89d" podStartSLOduration=2.319898433 podStartE2EDuration="2.319898433s" podCreationTimestamp="2026-03-07 15:18:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:18:46.3115694 +0000 UTC m=+2368.263705918" watchObservedRunningTime="2026-03-07 15:18:46.319898433 +0000 UTC m=+2368.272034971" Mar 07 15:18:47 crc kubenswrapper[4943]: I0307 15:18:47.305004 4943 generic.go:334] "Generic (PLEG): container finished" podID="6eee3b09-feb6-4a12-8fda-86657b069a23" containerID="ff82858ed1f603b577448411127acd4be9f0a58204f4854dbb8a8a225fc783fc" exitCode=0 Mar 07 15:18:47 crc kubenswrapper[4943]: I0307 15:18:47.305051 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-jk89d" event={"ID":"6eee3b09-feb6-4a12-8fda-86657b069a23","Type":"ContainerDied","Data":"ff82858ed1f603b577448411127acd4be9f0a58204f4854dbb8a8a225fc783fc"} Mar 07 15:18:48 crc kubenswrapper[4943]: I0307 15:18:48.686179 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-jk89d" Mar 07 15:18:48 crc kubenswrapper[4943]: I0307 15:18:48.729379 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-jk89d"] Mar 07 15:18:48 crc kubenswrapper[4943]: I0307 15:18:48.737697 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-jk89d"] Mar 07 15:18:48 crc kubenswrapper[4943]: I0307 15:18:48.857311 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6eee3b09-feb6-4a12-8fda-86657b069a23-ring-data-devices\") pod \"6eee3b09-feb6-4a12-8fda-86657b069a23\" (UID: \"6eee3b09-feb6-4a12-8fda-86657b069a23\") " Mar 07 15:18:48 crc kubenswrapper[4943]: I0307 15:18:48.857377 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sqn46\" (UniqueName: \"kubernetes.io/projected/6eee3b09-feb6-4a12-8fda-86657b069a23-kube-api-access-sqn46\") pod \"6eee3b09-feb6-4a12-8fda-86657b069a23\" (UID: \"6eee3b09-feb6-4a12-8fda-86657b069a23\") " Mar 07 15:18:48 crc kubenswrapper[4943]: I0307 15:18:48.857418 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6eee3b09-feb6-4a12-8fda-86657b069a23-swiftconf\") pod \"6eee3b09-feb6-4a12-8fda-86657b069a23\" (UID: \"6eee3b09-feb6-4a12-8fda-86657b069a23\") " Mar 07 15:18:48 crc kubenswrapper[4943]: I0307 15:18:48.857456 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6eee3b09-feb6-4a12-8fda-86657b069a23-scripts\") pod \"6eee3b09-feb6-4a12-8fda-86657b069a23\" (UID: \"6eee3b09-feb6-4a12-8fda-86657b069a23\") " Mar 07 15:18:48 crc kubenswrapper[4943]: I0307 15:18:48.857534 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6eee3b09-feb6-4a12-8fda-86657b069a23-dispersionconf\") pod \"6eee3b09-feb6-4a12-8fda-86657b069a23\" (UID: \"6eee3b09-feb6-4a12-8fda-86657b069a23\") " Mar 07 15:18:48 crc kubenswrapper[4943]: I0307 15:18:48.857716 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6eee3b09-feb6-4a12-8fda-86657b069a23-etc-swift\") pod \"6eee3b09-feb6-4a12-8fda-86657b069a23\" (UID: \"6eee3b09-feb6-4a12-8fda-86657b069a23\") " Mar 07 15:18:48 crc kubenswrapper[4943]: I0307 15:18:48.858527 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6eee3b09-feb6-4a12-8fda-86657b069a23-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "6eee3b09-feb6-4a12-8fda-86657b069a23" (UID: "6eee3b09-feb6-4a12-8fda-86657b069a23"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:18:48 crc kubenswrapper[4943]: I0307 15:18:48.858986 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6eee3b09-feb6-4a12-8fda-86657b069a23-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "6eee3b09-feb6-4a12-8fda-86657b069a23" (UID: "6eee3b09-feb6-4a12-8fda-86657b069a23"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:18:48 crc kubenswrapper[4943]: I0307 15:18:48.869535 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6eee3b09-feb6-4a12-8fda-86657b069a23-kube-api-access-sqn46" (OuterVolumeSpecName: "kube-api-access-sqn46") pod "6eee3b09-feb6-4a12-8fda-86657b069a23" (UID: "6eee3b09-feb6-4a12-8fda-86657b069a23"). InnerVolumeSpecName "kube-api-access-sqn46". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:18:48 crc kubenswrapper[4943]: I0307 15:18:48.880706 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6eee3b09-feb6-4a12-8fda-86657b069a23-scripts" (OuterVolumeSpecName: "scripts") pod "6eee3b09-feb6-4a12-8fda-86657b069a23" (UID: "6eee3b09-feb6-4a12-8fda-86657b069a23"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:18:48 crc kubenswrapper[4943]: I0307 15:18:48.888035 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6eee3b09-feb6-4a12-8fda-86657b069a23-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "6eee3b09-feb6-4a12-8fda-86657b069a23" (UID: "6eee3b09-feb6-4a12-8fda-86657b069a23"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:18:48 crc kubenswrapper[4943]: I0307 15:18:48.897549 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6eee3b09-feb6-4a12-8fda-86657b069a23-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "6eee3b09-feb6-4a12-8fda-86657b069a23" (UID: "6eee3b09-feb6-4a12-8fda-86657b069a23"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:18:48 crc kubenswrapper[4943]: I0307 15:18:48.959709 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6eee3b09-feb6-4a12-8fda-86657b069a23-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:48 crc kubenswrapper[4943]: I0307 15:18:48.959765 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6eee3b09-feb6-4a12-8fda-86657b069a23-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:48 crc kubenswrapper[4943]: I0307 15:18:48.959782 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sqn46\" (UniqueName: \"kubernetes.io/projected/6eee3b09-feb6-4a12-8fda-86657b069a23-kube-api-access-sqn46\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:48 crc kubenswrapper[4943]: I0307 15:18:48.959794 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6eee3b09-feb6-4a12-8fda-86657b069a23-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:48 crc kubenswrapper[4943]: I0307 15:18:48.959806 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6eee3b09-feb6-4a12-8fda-86657b069a23-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:48 crc kubenswrapper[4943]: I0307 15:18:48.959817 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6eee3b09-feb6-4a12-8fda-86657b069a23-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:49 crc kubenswrapper[4943]: I0307 15:18:49.329043 4943 scope.go:117] "RemoveContainer" containerID="ff82858ed1f603b577448411127acd4be9f0a58204f4854dbb8a8a225fc783fc" Mar 07 15:18:49 crc kubenswrapper[4943]: I0307 15:18:49.329644 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-jk89d" Mar 07 15:18:49 crc kubenswrapper[4943]: I0307 15:18:49.911372 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-rkbcj"] Mar 07 15:18:49 crc kubenswrapper[4943]: E0307 15:18:49.912491 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6eee3b09-feb6-4a12-8fda-86657b069a23" containerName="swift-ring-rebalance" Mar 07 15:18:49 crc kubenswrapper[4943]: I0307 15:18:49.912522 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="6eee3b09-feb6-4a12-8fda-86657b069a23" containerName="swift-ring-rebalance" Mar 07 15:18:49 crc kubenswrapper[4943]: I0307 15:18:49.912916 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="6eee3b09-feb6-4a12-8fda-86657b069a23" containerName="swift-ring-rebalance" Mar 07 15:18:49 crc kubenswrapper[4943]: I0307 15:18:49.913982 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-rkbcj" Mar 07 15:18:49 crc kubenswrapper[4943]: I0307 15:18:49.917590 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:18:49 crc kubenswrapper[4943]: I0307 15:18:49.917883 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:18:49 crc kubenswrapper[4943]: I0307 15:18:49.923082 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-rkbcj"] Mar 07 15:18:50 crc kubenswrapper[4943]: I0307 15:18:50.079672 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/14813b62-6230-4532-ae3e-1a5a686d5156-dispersionconf\") pod \"swift-ring-rebalance-debug-rkbcj\" (UID: \"14813b62-6230-4532-ae3e-1a5a686d5156\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rkbcj" Mar 07 15:18:50 crc kubenswrapper[4943]: I0307 15:18:50.079761 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/14813b62-6230-4532-ae3e-1a5a686d5156-etc-swift\") pod \"swift-ring-rebalance-debug-rkbcj\" (UID: \"14813b62-6230-4532-ae3e-1a5a686d5156\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rkbcj" Mar 07 15:18:50 crc kubenswrapper[4943]: I0307 15:18:50.079803 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/14813b62-6230-4532-ae3e-1a5a686d5156-swiftconf\") pod \"swift-ring-rebalance-debug-rkbcj\" (UID: \"14813b62-6230-4532-ae3e-1a5a686d5156\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rkbcj" Mar 07 15:18:50 crc kubenswrapper[4943]: I0307 15:18:50.080106 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/14813b62-6230-4532-ae3e-1a5a686d5156-ring-data-devices\") pod \"swift-ring-rebalance-debug-rkbcj\" (UID: \"14813b62-6230-4532-ae3e-1a5a686d5156\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rkbcj" Mar 07 15:18:50 crc kubenswrapper[4943]: I0307 15:18:50.080193 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/14813b62-6230-4532-ae3e-1a5a686d5156-scripts\") pod \"swift-ring-rebalance-debug-rkbcj\" (UID: \"14813b62-6230-4532-ae3e-1a5a686d5156\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rkbcj" Mar 07 15:18:50 crc kubenswrapper[4943]: I0307 15:18:50.080316 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v56tv\" (UniqueName: \"kubernetes.io/projected/14813b62-6230-4532-ae3e-1a5a686d5156-kube-api-access-v56tv\") pod \"swift-ring-rebalance-debug-rkbcj\" (UID: \"14813b62-6230-4532-ae3e-1a5a686d5156\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rkbcj" Mar 07 15:18:50 crc kubenswrapper[4943]: I0307 15:18:50.181299 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v56tv\" (UniqueName: \"kubernetes.io/projected/14813b62-6230-4532-ae3e-1a5a686d5156-kube-api-access-v56tv\") pod \"swift-ring-rebalance-debug-rkbcj\" (UID: \"14813b62-6230-4532-ae3e-1a5a686d5156\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rkbcj" Mar 07 15:18:50 crc kubenswrapper[4943]: I0307 15:18:50.181377 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/14813b62-6230-4532-ae3e-1a5a686d5156-dispersionconf\") pod \"swift-ring-rebalance-debug-rkbcj\" (UID: \"14813b62-6230-4532-ae3e-1a5a686d5156\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rkbcj" Mar 07 15:18:50 crc kubenswrapper[4943]: I0307 15:18:50.181401 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/14813b62-6230-4532-ae3e-1a5a686d5156-etc-swift\") pod \"swift-ring-rebalance-debug-rkbcj\" (UID: \"14813b62-6230-4532-ae3e-1a5a686d5156\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rkbcj" Mar 07 15:18:50 crc kubenswrapper[4943]: I0307 15:18:50.181424 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/14813b62-6230-4532-ae3e-1a5a686d5156-swiftconf\") pod \"swift-ring-rebalance-debug-rkbcj\" (UID: \"14813b62-6230-4532-ae3e-1a5a686d5156\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rkbcj" Mar 07 15:18:50 crc kubenswrapper[4943]: I0307 15:18:50.181506 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/14813b62-6230-4532-ae3e-1a5a686d5156-ring-data-devices\") pod \"swift-ring-rebalance-debug-rkbcj\" (UID: \"14813b62-6230-4532-ae3e-1a5a686d5156\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rkbcj" Mar 07 15:18:50 crc kubenswrapper[4943]: I0307 15:18:50.181532 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/14813b62-6230-4532-ae3e-1a5a686d5156-scripts\") pod \"swift-ring-rebalance-debug-rkbcj\" (UID: \"14813b62-6230-4532-ae3e-1a5a686d5156\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rkbcj" Mar 07 15:18:50 crc kubenswrapper[4943]: I0307 15:18:50.182005 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/14813b62-6230-4532-ae3e-1a5a686d5156-etc-swift\") pod \"swift-ring-rebalance-debug-rkbcj\" (UID: \"14813b62-6230-4532-ae3e-1a5a686d5156\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rkbcj" Mar 07 15:18:50 crc kubenswrapper[4943]: I0307 15:18:50.182561 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/14813b62-6230-4532-ae3e-1a5a686d5156-scripts\") pod \"swift-ring-rebalance-debug-rkbcj\" (UID: \"14813b62-6230-4532-ae3e-1a5a686d5156\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rkbcj" Mar 07 15:18:50 crc kubenswrapper[4943]: I0307 15:18:50.182787 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/14813b62-6230-4532-ae3e-1a5a686d5156-ring-data-devices\") pod \"swift-ring-rebalance-debug-rkbcj\" (UID: \"14813b62-6230-4532-ae3e-1a5a686d5156\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rkbcj" Mar 07 15:18:50 crc kubenswrapper[4943]: I0307 15:18:50.188810 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/14813b62-6230-4532-ae3e-1a5a686d5156-dispersionconf\") pod \"swift-ring-rebalance-debug-rkbcj\" (UID: \"14813b62-6230-4532-ae3e-1a5a686d5156\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rkbcj" Mar 07 15:18:50 crc kubenswrapper[4943]: I0307 15:18:50.188912 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/14813b62-6230-4532-ae3e-1a5a686d5156-swiftconf\") pod \"swift-ring-rebalance-debug-rkbcj\" (UID: \"14813b62-6230-4532-ae3e-1a5a686d5156\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rkbcj" Mar 07 15:18:50 crc kubenswrapper[4943]: I0307 15:18:50.215706 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v56tv\" (UniqueName: \"kubernetes.io/projected/14813b62-6230-4532-ae3e-1a5a686d5156-kube-api-access-v56tv\") pod \"swift-ring-rebalance-debug-rkbcj\" (UID: \"14813b62-6230-4532-ae3e-1a5a686d5156\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rkbcj" Mar 07 15:18:50 crc kubenswrapper[4943]: I0307 15:18:50.235114 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-rkbcj" Mar 07 15:18:50 crc kubenswrapper[4943]: I0307 15:18:50.475898 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-rkbcj"] Mar 07 15:18:50 crc kubenswrapper[4943]: W0307 15:18:50.478075 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod14813b62_6230_4532_ae3e_1a5a686d5156.slice/crio-2eae6204031e3822990691e5d61d58c7d2537d6dbb575a479fc4aef389215236 WatchSource:0}: Error finding container 2eae6204031e3822990691e5d61d58c7d2537d6dbb575a479fc4aef389215236: Status 404 returned error can't find the container with id 2eae6204031e3822990691e5d61d58c7d2537d6dbb575a479fc4aef389215236 Mar 07 15:18:50 crc kubenswrapper[4943]: I0307 15:18:50.767892 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6eee3b09-feb6-4a12-8fda-86657b069a23" path="/var/lib/kubelet/pods/6eee3b09-feb6-4a12-8fda-86657b069a23/volumes" Mar 07 15:18:51 crc kubenswrapper[4943]: I0307 15:18:51.360483 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-rkbcj" event={"ID":"14813b62-6230-4532-ae3e-1a5a686d5156","Type":"ContainerStarted","Data":"368159a0e84aef45760627cc1b2cc7fd358c90db1a223bd795d6690286a4f335"} Mar 07 15:18:51 crc kubenswrapper[4943]: I0307 15:18:51.360844 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-rkbcj" event={"ID":"14813b62-6230-4532-ae3e-1a5a686d5156","Type":"ContainerStarted","Data":"2eae6204031e3822990691e5d61d58c7d2537d6dbb575a479fc4aef389215236"} Mar 07 15:18:51 crc kubenswrapper[4943]: I0307 15:18:51.393767 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-rkbcj" podStartSLOduration=2.393743546 podStartE2EDuration="2.393743546s" podCreationTimestamp="2026-03-07 15:18:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:18:51.387059512 +0000 UTC m=+2373.339196020" watchObservedRunningTime="2026-03-07 15:18:51.393743546 +0000 UTC m=+2373.345880054" Mar 07 15:18:52 crc kubenswrapper[4943]: I0307 15:18:52.372897 4943 generic.go:334] "Generic (PLEG): container finished" podID="14813b62-6230-4532-ae3e-1a5a686d5156" containerID="368159a0e84aef45760627cc1b2cc7fd358c90db1a223bd795d6690286a4f335" exitCode=0 Mar 07 15:18:52 crc kubenswrapper[4943]: I0307 15:18:52.372963 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-rkbcj" event={"ID":"14813b62-6230-4532-ae3e-1a5a686d5156","Type":"ContainerDied","Data":"368159a0e84aef45760627cc1b2cc7fd358c90db1a223bd795d6690286a4f335"} Mar 07 15:18:53 crc kubenswrapper[4943]: I0307 15:18:53.788004 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-rkbcj" Mar 07 15:18:53 crc kubenswrapper[4943]: I0307 15:18:53.830061 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-rkbcj"] Mar 07 15:18:53 crc kubenswrapper[4943]: I0307 15:18:53.843454 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-rkbcj"] Mar 07 15:18:53 crc kubenswrapper[4943]: I0307 15:18:53.938140 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v56tv\" (UniqueName: \"kubernetes.io/projected/14813b62-6230-4532-ae3e-1a5a686d5156-kube-api-access-v56tv\") pod \"14813b62-6230-4532-ae3e-1a5a686d5156\" (UID: \"14813b62-6230-4532-ae3e-1a5a686d5156\") " Mar 07 15:18:53 crc kubenswrapper[4943]: I0307 15:18:53.938316 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/14813b62-6230-4532-ae3e-1a5a686d5156-ring-data-devices\") pod \"14813b62-6230-4532-ae3e-1a5a686d5156\" (UID: \"14813b62-6230-4532-ae3e-1a5a686d5156\") " Mar 07 15:18:53 crc kubenswrapper[4943]: I0307 15:18:53.939435 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14813b62-6230-4532-ae3e-1a5a686d5156-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "14813b62-6230-4532-ae3e-1a5a686d5156" (UID: "14813b62-6230-4532-ae3e-1a5a686d5156"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:18:53 crc kubenswrapper[4943]: I0307 15:18:53.939626 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/14813b62-6230-4532-ae3e-1a5a686d5156-etc-swift\") pod \"14813b62-6230-4532-ae3e-1a5a686d5156\" (UID: \"14813b62-6230-4532-ae3e-1a5a686d5156\") " Mar 07 15:18:53 crc kubenswrapper[4943]: I0307 15:18:53.940859 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14813b62-6230-4532-ae3e-1a5a686d5156-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "14813b62-6230-4532-ae3e-1a5a686d5156" (UID: "14813b62-6230-4532-ae3e-1a5a686d5156"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:18:53 crc kubenswrapper[4943]: I0307 15:18:53.941067 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/14813b62-6230-4532-ae3e-1a5a686d5156-swiftconf\") pod \"14813b62-6230-4532-ae3e-1a5a686d5156\" (UID: \"14813b62-6230-4532-ae3e-1a5a686d5156\") " Mar 07 15:18:53 crc kubenswrapper[4943]: I0307 15:18:53.941601 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/14813b62-6230-4532-ae3e-1a5a686d5156-dispersionconf\") pod \"14813b62-6230-4532-ae3e-1a5a686d5156\" (UID: \"14813b62-6230-4532-ae3e-1a5a686d5156\") " Mar 07 15:18:53 crc kubenswrapper[4943]: I0307 15:18:53.941716 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/14813b62-6230-4532-ae3e-1a5a686d5156-scripts\") pod \"14813b62-6230-4532-ae3e-1a5a686d5156\" (UID: \"14813b62-6230-4532-ae3e-1a5a686d5156\") " Mar 07 15:18:53 crc kubenswrapper[4943]: I0307 15:18:53.942983 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/14813b62-6230-4532-ae3e-1a5a686d5156-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:53 crc kubenswrapper[4943]: I0307 15:18:53.943019 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/14813b62-6230-4532-ae3e-1a5a686d5156-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:53 crc kubenswrapper[4943]: I0307 15:18:53.947216 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14813b62-6230-4532-ae3e-1a5a686d5156-kube-api-access-v56tv" (OuterVolumeSpecName: "kube-api-access-v56tv") pod "14813b62-6230-4532-ae3e-1a5a686d5156" (UID: "14813b62-6230-4532-ae3e-1a5a686d5156"). InnerVolumeSpecName "kube-api-access-v56tv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:18:53 crc kubenswrapper[4943]: I0307 15:18:53.971831 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14813b62-6230-4532-ae3e-1a5a686d5156-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "14813b62-6230-4532-ae3e-1a5a686d5156" (UID: "14813b62-6230-4532-ae3e-1a5a686d5156"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:18:53 crc kubenswrapper[4943]: I0307 15:18:53.989177 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14813b62-6230-4532-ae3e-1a5a686d5156-scripts" (OuterVolumeSpecName: "scripts") pod "14813b62-6230-4532-ae3e-1a5a686d5156" (UID: "14813b62-6230-4532-ae3e-1a5a686d5156"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:18:53 crc kubenswrapper[4943]: I0307 15:18:53.989736 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14813b62-6230-4532-ae3e-1a5a686d5156-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "14813b62-6230-4532-ae3e-1a5a686d5156" (UID: "14813b62-6230-4532-ae3e-1a5a686d5156"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:18:54 crc kubenswrapper[4943]: I0307 15:18:54.044789 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v56tv\" (UniqueName: \"kubernetes.io/projected/14813b62-6230-4532-ae3e-1a5a686d5156-kube-api-access-v56tv\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:54 crc kubenswrapper[4943]: I0307 15:18:54.045059 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/14813b62-6230-4532-ae3e-1a5a686d5156-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:54 crc kubenswrapper[4943]: I0307 15:18:54.045184 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/14813b62-6230-4532-ae3e-1a5a686d5156-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:54 crc kubenswrapper[4943]: I0307 15:18:54.045320 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/14813b62-6230-4532-ae3e-1a5a686d5156-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:54 crc kubenswrapper[4943]: I0307 15:18:54.401051 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2eae6204031e3822990691e5d61d58c7d2537d6dbb575a479fc4aef389215236" Mar 07 15:18:54 crc kubenswrapper[4943]: I0307 15:18:54.401151 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-rkbcj" Mar 07 15:18:54 crc kubenswrapper[4943]: I0307 15:18:54.774282 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14813b62-6230-4532-ae3e-1a5a686d5156" path="/var/lib/kubelet/pods/14813b62-6230-4532-ae3e-1a5a686d5156/volumes" Mar 07 15:18:54 crc kubenswrapper[4943]: I0307 15:18:54.966509 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-8f4vs"] Mar 07 15:18:54 crc kubenswrapper[4943]: E0307 15:18:54.967709 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14813b62-6230-4532-ae3e-1a5a686d5156" containerName="swift-ring-rebalance" Mar 07 15:18:54 crc kubenswrapper[4943]: I0307 15:18:54.967779 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="14813b62-6230-4532-ae3e-1a5a686d5156" containerName="swift-ring-rebalance" Mar 07 15:18:54 crc kubenswrapper[4943]: I0307 15:18:54.968007 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="14813b62-6230-4532-ae3e-1a5a686d5156" containerName="swift-ring-rebalance" Mar 07 15:18:54 crc kubenswrapper[4943]: I0307 15:18:54.968552 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8f4vs" Mar 07 15:18:54 crc kubenswrapper[4943]: I0307 15:18:54.970820 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:18:54 crc kubenswrapper[4943]: I0307 15:18:54.970855 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:18:54 crc kubenswrapper[4943]: I0307 15:18:54.985807 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-8f4vs"] Mar 07 15:18:55 crc kubenswrapper[4943]: I0307 15:18:55.060377 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a599aee5-0766-4f14-a961-c1c743b71b65-swiftconf\") pod \"swift-ring-rebalance-debug-8f4vs\" (UID: \"a599aee5-0766-4f14-a961-c1c743b71b65\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8f4vs" Mar 07 15:18:55 crc kubenswrapper[4943]: I0307 15:18:55.060430 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a599aee5-0766-4f14-a961-c1c743b71b65-etc-swift\") pod \"swift-ring-rebalance-debug-8f4vs\" (UID: \"a599aee5-0766-4f14-a961-c1c743b71b65\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8f4vs" Mar 07 15:18:55 crc kubenswrapper[4943]: I0307 15:18:55.060464 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnhld\" (UniqueName: \"kubernetes.io/projected/a599aee5-0766-4f14-a961-c1c743b71b65-kube-api-access-qnhld\") pod \"swift-ring-rebalance-debug-8f4vs\" (UID: \"a599aee5-0766-4f14-a961-c1c743b71b65\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8f4vs" Mar 07 15:18:55 crc kubenswrapper[4943]: I0307 15:18:55.060524 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a599aee5-0766-4f14-a961-c1c743b71b65-dispersionconf\") pod \"swift-ring-rebalance-debug-8f4vs\" (UID: \"a599aee5-0766-4f14-a961-c1c743b71b65\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8f4vs" Mar 07 15:18:55 crc kubenswrapper[4943]: I0307 15:18:55.060545 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a599aee5-0766-4f14-a961-c1c743b71b65-ring-data-devices\") pod \"swift-ring-rebalance-debug-8f4vs\" (UID: \"a599aee5-0766-4f14-a961-c1c743b71b65\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8f4vs" Mar 07 15:18:55 crc kubenswrapper[4943]: I0307 15:18:55.060567 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a599aee5-0766-4f14-a961-c1c743b71b65-scripts\") pod \"swift-ring-rebalance-debug-8f4vs\" (UID: \"a599aee5-0766-4f14-a961-c1c743b71b65\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8f4vs" Mar 07 15:18:55 crc kubenswrapper[4943]: I0307 15:18:55.161889 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnhld\" (UniqueName: \"kubernetes.io/projected/a599aee5-0766-4f14-a961-c1c743b71b65-kube-api-access-qnhld\") pod \"swift-ring-rebalance-debug-8f4vs\" (UID: \"a599aee5-0766-4f14-a961-c1c743b71b65\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8f4vs" Mar 07 15:18:55 crc kubenswrapper[4943]: I0307 15:18:55.162120 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a599aee5-0766-4f14-a961-c1c743b71b65-dispersionconf\") pod \"swift-ring-rebalance-debug-8f4vs\" (UID: \"a599aee5-0766-4f14-a961-c1c743b71b65\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8f4vs" Mar 07 15:18:55 crc kubenswrapper[4943]: I0307 15:18:55.162182 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a599aee5-0766-4f14-a961-c1c743b71b65-ring-data-devices\") pod \"swift-ring-rebalance-debug-8f4vs\" (UID: \"a599aee5-0766-4f14-a961-c1c743b71b65\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8f4vs" Mar 07 15:18:55 crc kubenswrapper[4943]: I0307 15:18:55.162251 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a599aee5-0766-4f14-a961-c1c743b71b65-scripts\") pod \"swift-ring-rebalance-debug-8f4vs\" (UID: \"a599aee5-0766-4f14-a961-c1c743b71b65\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8f4vs" Mar 07 15:18:55 crc kubenswrapper[4943]: I0307 15:18:55.162467 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a599aee5-0766-4f14-a961-c1c743b71b65-swiftconf\") pod \"swift-ring-rebalance-debug-8f4vs\" (UID: \"a599aee5-0766-4f14-a961-c1c743b71b65\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8f4vs" Mar 07 15:18:55 crc kubenswrapper[4943]: I0307 15:18:55.162537 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a599aee5-0766-4f14-a961-c1c743b71b65-etc-swift\") pod \"swift-ring-rebalance-debug-8f4vs\" (UID: \"a599aee5-0766-4f14-a961-c1c743b71b65\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8f4vs" Mar 07 15:18:55 crc kubenswrapper[4943]: I0307 15:18:55.163335 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a599aee5-0766-4f14-a961-c1c743b71b65-etc-swift\") pod \"swift-ring-rebalance-debug-8f4vs\" (UID: \"a599aee5-0766-4f14-a961-c1c743b71b65\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8f4vs" Mar 07 15:18:55 crc kubenswrapper[4943]: I0307 15:18:55.163653 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a599aee5-0766-4f14-a961-c1c743b71b65-ring-data-devices\") pod \"swift-ring-rebalance-debug-8f4vs\" (UID: \"a599aee5-0766-4f14-a961-c1c743b71b65\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8f4vs" Mar 07 15:18:55 crc kubenswrapper[4943]: I0307 15:18:55.163812 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a599aee5-0766-4f14-a961-c1c743b71b65-scripts\") pod \"swift-ring-rebalance-debug-8f4vs\" (UID: \"a599aee5-0766-4f14-a961-c1c743b71b65\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8f4vs" Mar 07 15:18:55 crc kubenswrapper[4943]: I0307 15:18:55.168865 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a599aee5-0766-4f14-a961-c1c743b71b65-dispersionconf\") pod \"swift-ring-rebalance-debug-8f4vs\" (UID: \"a599aee5-0766-4f14-a961-c1c743b71b65\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8f4vs" Mar 07 15:18:55 crc kubenswrapper[4943]: I0307 15:18:55.171698 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a599aee5-0766-4f14-a961-c1c743b71b65-swiftconf\") pod \"swift-ring-rebalance-debug-8f4vs\" (UID: \"a599aee5-0766-4f14-a961-c1c743b71b65\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8f4vs" Mar 07 15:18:55 crc kubenswrapper[4943]: I0307 15:18:55.197226 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnhld\" (UniqueName: \"kubernetes.io/projected/a599aee5-0766-4f14-a961-c1c743b71b65-kube-api-access-qnhld\") pod \"swift-ring-rebalance-debug-8f4vs\" (UID: \"a599aee5-0766-4f14-a961-c1c743b71b65\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8f4vs" Mar 07 15:18:55 crc kubenswrapper[4943]: I0307 15:18:55.304320 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8f4vs" Mar 07 15:18:55 crc kubenswrapper[4943]: I0307 15:18:55.550515 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-8f4vs"] Mar 07 15:18:55 crc kubenswrapper[4943]: W0307 15:18:55.568359 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda599aee5_0766_4f14_a961_c1c743b71b65.slice/crio-4443d32defce2bd5e1bd5dcd3da62fb1ba0dd27bca9911fa44dd70f87a299dca WatchSource:0}: Error finding container 4443d32defce2bd5e1bd5dcd3da62fb1ba0dd27bca9911fa44dd70f87a299dca: Status 404 returned error can't find the container with id 4443d32defce2bd5e1bd5dcd3da62fb1ba0dd27bca9911fa44dd70f87a299dca Mar 07 15:18:56 crc kubenswrapper[4943]: I0307 15:18:56.444864 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8f4vs" event={"ID":"a599aee5-0766-4f14-a961-c1c743b71b65","Type":"ContainerStarted","Data":"e955e20dafce7619e536a75178762f613fe0332cc33685e03107bd475fa4d6c0"} Mar 07 15:18:56 crc kubenswrapper[4943]: I0307 15:18:56.445326 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8f4vs" event={"ID":"a599aee5-0766-4f14-a961-c1c743b71b65","Type":"ContainerStarted","Data":"4443d32defce2bd5e1bd5dcd3da62fb1ba0dd27bca9911fa44dd70f87a299dca"} Mar 07 15:18:56 crc kubenswrapper[4943]: I0307 15:18:56.482108 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8f4vs" podStartSLOduration=2.482075413 podStartE2EDuration="2.482075413s" podCreationTimestamp="2026-03-07 15:18:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:18:56.467741733 +0000 UTC m=+2378.419878271" watchObservedRunningTime="2026-03-07 15:18:56.482075413 +0000 UTC m=+2378.434211961" Mar 07 15:18:56 crc kubenswrapper[4943]: I0307 15:18:56.756631 4943 scope.go:117] "RemoveContainer" containerID="448308ad6e34ac922d66b3c90e66d0927407a94dd28633ec350502d69e4ca8aa" Mar 07 15:18:56 crc kubenswrapper[4943]: E0307 15:18:56.757106 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cpdmm_openshift-machine-config-operator(4c1f18ea-e557-41fc-95dd-ce4e1722a61e)\"" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" Mar 07 15:18:57 crc kubenswrapper[4943]: I0307 15:18:57.460922 4943 generic.go:334] "Generic (PLEG): container finished" podID="a599aee5-0766-4f14-a961-c1c743b71b65" containerID="e955e20dafce7619e536a75178762f613fe0332cc33685e03107bd475fa4d6c0" exitCode=0 Mar 07 15:18:57 crc kubenswrapper[4943]: I0307 15:18:57.461030 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8f4vs" event={"ID":"a599aee5-0766-4f14-a961-c1c743b71b65","Type":"ContainerDied","Data":"e955e20dafce7619e536a75178762f613fe0332cc33685e03107bd475fa4d6c0"} Mar 07 15:18:58 crc kubenswrapper[4943]: I0307 15:18:58.848875 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8f4vs" Mar 07 15:18:58 crc kubenswrapper[4943]: I0307 15:18:58.897454 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-8f4vs"] Mar 07 15:18:58 crc kubenswrapper[4943]: I0307 15:18:58.905371 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-8f4vs"] Mar 07 15:18:59 crc kubenswrapper[4943]: I0307 15:18:59.027560 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qnhld\" (UniqueName: \"kubernetes.io/projected/a599aee5-0766-4f14-a961-c1c743b71b65-kube-api-access-qnhld\") pod \"a599aee5-0766-4f14-a961-c1c743b71b65\" (UID: \"a599aee5-0766-4f14-a961-c1c743b71b65\") " Mar 07 15:18:59 crc kubenswrapper[4943]: I0307 15:18:59.027713 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a599aee5-0766-4f14-a961-c1c743b71b65-ring-data-devices\") pod \"a599aee5-0766-4f14-a961-c1c743b71b65\" (UID: \"a599aee5-0766-4f14-a961-c1c743b71b65\") " Mar 07 15:18:59 crc kubenswrapper[4943]: I0307 15:18:59.027758 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a599aee5-0766-4f14-a961-c1c743b71b65-dispersionconf\") pod \"a599aee5-0766-4f14-a961-c1c743b71b65\" (UID: \"a599aee5-0766-4f14-a961-c1c743b71b65\") " Mar 07 15:18:59 crc kubenswrapper[4943]: I0307 15:18:59.027856 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a599aee5-0766-4f14-a961-c1c743b71b65-etc-swift\") pod \"a599aee5-0766-4f14-a961-c1c743b71b65\" (UID: \"a599aee5-0766-4f14-a961-c1c743b71b65\") " Mar 07 15:18:59 crc kubenswrapper[4943]: I0307 15:18:59.027954 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a599aee5-0766-4f14-a961-c1c743b71b65-swiftconf\") pod \"a599aee5-0766-4f14-a961-c1c743b71b65\" (UID: \"a599aee5-0766-4f14-a961-c1c743b71b65\") " Mar 07 15:18:59 crc kubenswrapper[4943]: I0307 15:18:59.027990 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a599aee5-0766-4f14-a961-c1c743b71b65-scripts\") pod \"a599aee5-0766-4f14-a961-c1c743b71b65\" (UID: \"a599aee5-0766-4f14-a961-c1c743b71b65\") " Mar 07 15:18:59 crc kubenswrapper[4943]: I0307 15:18:59.028688 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a599aee5-0766-4f14-a961-c1c743b71b65-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "a599aee5-0766-4f14-a961-c1c743b71b65" (UID: "a599aee5-0766-4f14-a961-c1c743b71b65"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:18:59 crc kubenswrapper[4943]: I0307 15:18:59.029585 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a599aee5-0766-4f14-a961-c1c743b71b65-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "a599aee5-0766-4f14-a961-c1c743b71b65" (UID: "a599aee5-0766-4f14-a961-c1c743b71b65"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:18:59 crc kubenswrapper[4943]: I0307 15:18:59.034802 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a599aee5-0766-4f14-a961-c1c743b71b65-kube-api-access-qnhld" (OuterVolumeSpecName: "kube-api-access-qnhld") pod "a599aee5-0766-4f14-a961-c1c743b71b65" (UID: "a599aee5-0766-4f14-a961-c1c743b71b65"). InnerVolumeSpecName "kube-api-access-qnhld". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:18:59 crc kubenswrapper[4943]: I0307 15:18:59.054819 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a599aee5-0766-4f14-a961-c1c743b71b65-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "a599aee5-0766-4f14-a961-c1c743b71b65" (UID: "a599aee5-0766-4f14-a961-c1c743b71b65"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:18:59 crc kubenswrapper[4943]: I0307 15:18:59.064883 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a599aee5-0766-4f14-a961-c1c743b71b65-scripts" (OuterVolumeSpecName: "scripts") pod "a599aee5-0766-4f14-a961-c1c743b71b65" (UID: "a599aee5-0766-4f14-a961-c1c743b71b65"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:18:59 crc kubenswrapper[4943]: I0307 15:18:59.073799 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a599aee5-0766-4f14-a961-c1c743b71b65-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "a599aee5-0766-4f14-a961-c1c743b71b65" (UID: "a599aee5-0766-4f14-a961-c1c743b71b65"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:18:59 crc kubenswrapper[4943]: I0307 15:18:59.130169 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a599aee5-0766-4f14-a961-c1c743b71b65-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:59 crc kubenswrapper[4943]: I0307 15:18:59.130212 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a599aee5-0766-4f14-a961-c1c743b71b65-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:59 crc kubenswrapper[4943]: I0307 15:18:59.130223 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a599aee5-0766-4f14-a961-c1c743b71b65-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:59 crc kubenswrapper[4943]: I0307 15:18:59.130234 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a599aee5-0766-4f14-a961-c1c743b71b65-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:59 crc kubenswrapper[4943]: I0307 15:18:59.130245 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a599aee5-0766-4f14-a961-c1c743b71b65-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:59 crc kubenswrapper[4943]: I0307 15:18:59.130256 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qnhld\" (UniqueName: \"kubernetes.io/projected/a599aee5-0766-4f14-a961-c1c743b71b65-kube-api-access-qnhld\") on node \"crc\" DevicePath \"\"" Mar 07 15:18:59 crc kubenswrapper[4943]: I0307 15:18:59.483738 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4443d32defce2bd5e1bd5dcd3da62fb1ba0dd27bca9911fa44dd70f87a299dca" Mar 07 15:18:59 crc kubenswrapper[4943]: I0307 15:18:59.483791 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8f4vs" Mar 07 15:19:00 crc kubenswrapper[4943]: I0307 15:19:00.059636 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-6xvld"] Mar 07 15:19:00 crc kubenswrapper[4943]: E0307 15:19:00.060176 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a599aee5-0766-4f14-a961-c1c743b71b65" containerName="swift-ring-rebalance" Mar 07 15:19:00 crc kubenswrapper[4943]: I0307 15:19:00.060201 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="a599aee5-0766-4f14-a961-c1c743b71b65" containerName="swift-ring-rebalance" Mar 07 15:19:00 crc kubenswrapper[4943]: I0307 15:19:00.060466 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="a599aee5-0766-4f14-a961-c1c743b71b65" containerName="swift-ring-rebalance" Mar 07 15:19:00 crc kubenswrapper[4943]: I0307 15:19:00.061269 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6xvld" Mar 07 15:19:00 crc kubenswrapper[4943]: I0307 15:19:00.064690 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:19:00 crc kubenswrapper[4943]: I0307 15:19:00.065018 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:19:00 crc kubenswrapper[4943]: I0307 15:19:00.087834 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-6xvld"] Mar 07 15:19:00 crc kubenswrapper[4943]: I0307 15:19:00.144622 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/0c5f5211-2520-45c1-acae-68e88d702031-dispersionconf\") pod \"swift-ring-rebalance-debug-6xvld\" (UID: \"0c5f5211-2520-45c1-acae-68e88d702031\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6xvld" Mar 07 15:19:00 crc kubenswrapper[4943]: I0307 15:19:00.144708 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0c5f5211-2520-45c1-acae-68e88d702031-scripts\") pod \"swift-ring-rebalance-debug-6xvld\" (UID: \"0c5f5211-2520-45c1-acae-68e88d702031\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6xvld" Mar 07 15:19:00 crc kubenswrapper[4943]: I0307 15:19:00.144791 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/0c5f5211-2520-45c1-acae-68e88d702031-ring-data-devices\") pod \"swift-ring-rebalance-debug-6xvld\" (UID: \"0c5f5211-2520-45c1-acae-68e88d702031\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6xvld" Mar 07 15:19:00 crc kubenswrapper[4943]: I0307 15:19:00.144828 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/0c5f5211-2520-45c1-acae-68e88d702031-swiftconf\") pod \"swift-ring-rebalance-debug-6xvld\" (UID: \"0c5f5211-2520-45c1-acae-68e88d702031\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6xvld" Mar 07 15:19:00 crc kubenswrapper[4943]: I0307 15:19:00.145001 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/0c5f5211-2520-45c1-acae-68e88d702031-etc-swift\") pod \"swift-ring-rebalance-debug-6xvld\" (UID: \"0c5f5211-2520-45c1-acae-68e88d702031\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6xvld" Mar 07 15:19:00 crc kubenswrapper[4943]: I0307 15:19:00.145091 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6vvd\" (UniqueName: \"kubernetes.io/projected/0c5f5211-2520-45c1-acae-68e88d702031-kube-api-access-d6vvd\") pod \"swift-ring-rebalance-debug-6xvld\" (UID: \"0c5f5211-2520-45c1-acae-68e88d702031\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6xvld" Mar 07 15:19:00 crc kubenswrapper[4943]: I0307 15:19:00.245990 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/0c5f5211-2520-45c1-acae-68e88d702031-etc-swift\") pod \"swift-ring-rebalance-debug-6xvld\" (UID: \"0c5f5211-2520-45c1-acae-68e88d702031\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6xvld" Mar 07 15:19:00 crc kubenswrapper[4943]: I0307 15:19:00.246510 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6vvd\" (UniqueName: \"kubernetes.io/projected/0c5f5211-2520-45c1-acae-68e88d702031-kube-api-access-d6vvd\") pod \"swift-ring-rebalance-debug-6xvld\" (UID: \"0c5f5211-2520-45c1-acae-68e88d702031\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6xvld" Mar 07 15:19:00 crc kubenswrapper[4943]: I0307 15:19:00.246652 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/0c5f5211-2520-45c1-acae-68e88d702031-dispersionconf\") pod \"swift-ring-rebalance-debug-6xvld\" (UID: \"0c5f5211-2520-45c1-acae-68e88d702031\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6xvld" Mar 07 15:19:00 crc kubenswrapper[4943]: I0307 15:19:00.246773 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0c5f5211-2520-45c1-acae-68e88d702031-scripts\") pod \"swift-ring-rebalance-debug-6xvld\" (UID: \"0c5f5211-2520-45c1-acae-68e88d702031\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6xvld" Mar 07 15:19:00 crc kubenswrapper[4943]: I0307 15:19:00.246894 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/0c5f5211-2520-45c1-acae-68e88d702031-ring-data-devices\") pod \"swift-ring-rebalance-debug-6xvld\" (UID: \"0c5f5211-2520-45c1-acae-68e88d702031\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6xvld" Mar 07 15:19:00 crc kubenswrapper[4943]: I0307 15:19:00.247026 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/0c5f5211-2520-45c1-acae-68e88d702031-swiftconf\") pod \"swift-ring-rebalance-debug-6xvld\" (UID: \"0c5f5211-2520-45c1-acae-68e88d702031\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6xvld" Mar 07 15:19:00 crc kubenswrapper[4943]: I0307 15:19:00.246720 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/0c5f5211-2520-45c1-acae-68e88d702031-etc-swift\") pod \"swift-ring-rebalance-debug-6xvld\" (UID: \"0c5f5211-2520-45c1-acae-68e88d702031\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6xvld" Mar 07 15:19:00 crc kubenswrapper[4943]: I0307 15:19:00.248714 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0c5f5211-2520-45c1-acae-68e88d702031-scripts\") pod \"swift-ring-rebalance-debug-6xvld\" (UID: \"0c5f5211-2520-45c1-acae-68e88d702031\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6xvld" Mar 07 15:19:00 crc kubenswrapper[4943]: I0307 15:19:00.248764 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/0c5f5211-2520-45c1-acae-68e88d702031-ring-data-devices\") pod \"swift-ring-rebalance-debug-6xvld\" (UID: \"0c5f5211-2520-45c1-acae-68e88d702031\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6xvld" Mar 07 15:19:00 crc kubenswrapper[4943]: I0307 15:19:00.254153 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/0c5f5211-2520-45c1-acae-68e88d702031-dispersionconf\") pod \"swift-ring-rebalance-debug-6xvld\" (UID: \"0c5f5211-2520-45c1-acae-68e88d702031\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6xvld" Mar 07 15:19:00 crc kubenswrapper[4943]: I0307 15:19:00.261304 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/0c5f5211-2520-45c1-acae-68e88d702031-swiftconf\") pod \"swift-ring-rebalance-debug-6xvld\" (UID: \"0c5f5211-2520-45c1-acae-68e88d702031\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6xvld" Mar 07 15:19:00 crc kubenswrapper[4943]: I0307 15:19:00.277973 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6vvd\" (UniqueName: \"kubernetes.io/projected/0c5f5211-2520-45c1-acae-68e88d702031-kube-api-access-d6vvd\") pod \"swift-ring-rebalance-debug-6xvld\" (UID: \"0c5f5211-2520-45c1-acae-68e88d702031\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6xvld" Mar 07 15:19:00 crc kubenswrapper[4943]: I0307 15:19:00.391841 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6xvld" Mar 07 15:19:00 crc kubenswrapper[4943]: I0307 15:19:00.598805 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-6xvld"] Mar 07 15:19:00 crc kubenswrapper[4943]: I0307 15:19:00.766223 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a599aee5-0766-4f14-a961-c1c743b71b65" path="/var/lib/kubelet/pods/a599aee5-0766-4f14-a961-c1c743b71b65/volumes" Mar 07 15:19:01 crc kubenswrapper[4943]: I0307 15:19:01.510591 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6xvld" event={"ID":"0c5f5211-2520-45c1-acae-68e88d702031","Type":"ContainerStarted","Data":"9ef46553800be268c4887197ec8125dc8e3308f4f5928d408f0b7b257fc240f9"} Mar 07 15:19:01 crc kubenswrapper[4943]: I0307 15:19:01.511056 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6xvld" event={"ID":"0c5f5211-2520-45c1-acae-68e88d702031","Type":"ContainerStarted","Data":"40342041ea06610208f01bba1b305aeb5fe2fc8838984730a56d7f08f990d629"} Mar 07 15:19:01 crc kubenswrapper[4943]: I0307 15:19:01.539169 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6xvld" podStartSLOduration=1.539144936 podStartE2EDuration="1.539144936s" podCreationTimestamp="2026-03-07 15:19:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:19:01.53233936 +0000 UTC m=+2383.484475888" watchObservedRunningTime="2026-03-07 15:19:01.539144936 +0000 UTC m=+2383.491281474" Mar 07 15:19:02 crc kubenswrapper[4943]: I0307 15:19:02.523998 4943 generic.go:334] "Generic (PLEG): container finished" podID="0c5f5211-2520-45c1-acae-68e88d702031" containerID="9ef46553800be268c4887197ec8125dc8e3308f4f5928d408f0b7b257fc240f9" exitCode=0 Mar 07 15:19:02 crc kubenswrapper[4943]: I0307 15:19:02.524122 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6xvld" event={"ID":"0c5f5211-2520-45c1-acae-68e88d702031","Type":"ContainerDied","Data":"9ef46553800be268c4887197ec8125dc8e3308f4f5928d408f0b7b257fc240f9"} Mar 07 15:19:03 crc kubenswrapper[4943]: I0307 15:19:03.924500 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6xvld" Mar 07 15:19:03 crc kubenswrapper[4943]: I0307 15:19:03.975066 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-6xvld"] Mar 07 15:19:03 crc kubenswrapper[4943]: I0307 15:19:03.984378 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-6xvld"] Mar 07 15:19:04 crc kubenswrapper[4943]: I0307 15:19:04.008854 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/0c5f5211-2520-45c1-acae-68e88d702031-ring-data-devices\") pod \"0c5f5211-2520-45c1-acae-68e88d702031\" (UID: \"0c5f5211-2520-45c1-acae-68e88d702031\") " Mar 07 15:19:04 crc kubenswrapper[4943]: I0307 15:19:04.008967 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6vvd\" (UniqueName: \"kubernetes.io/projected/0c5f5211-2520-45c1-acae-68e88d702031-kube-api-access-d6vvd\") pod \"0c5f5211-2520-45c1-acae-68e88d702031\" (UID: \"0c5f5211-2520-45c1-acae-68e88d702031\") " Mar 07 15:19:04 crc kubenswrapper[4943]: I0307 15:19:04.009036 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0c5f5211-2520-45c1-acae-68e88d702031-scripts\") pod \"0c5f5211-2520-45c1-acae-68e88d702031\" (UID: \"0c5f5211-2520-45c1-acae-68e88d702031\") " Mar 07 15:19:04 crc kubenswrapper[4943]: I0307 15:19:04.009126 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/0c5f5211-2520-45c1-acae-68e88d702031-etc-swift\") pod \"0c5f5211-2520-45c1-acae-68e88d702031\" (UID: \"0c5f5211-2520-45c1-acae-68e88d702031\") " Mar 07 15:19:04 crc kubenswrapper[4943]: I0307 15:19:04.009172 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/0c5f5211-2520-45c1-acae-68e88d702031-swiftconf\") pod \"0c5f5211-2520-45c1-acae-68e88d702031\" (UID: \"0c5f5211-2520-45c1-acae-68e88d702031\") " Mar 07 15:19:04 crc kubenswrapper[4943]: I0307 15:19:04.009209 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/0c5f5211-2520-45c1-acae-68e88d702031-dispersionconf\") pod \"0c5f5211-2520-45c1-acae-68e88d702031\" (UID: \"0c5f5211-2520-45c1-acae-68e88d702031\") " Mar 07 15:19:04 crc kubenswrapper[4943]: I0307 15:19:04.009959 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c5f5211-2520-45c1-acae-68e88d702031-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "0c5f5211-2520-45c1-acae-68e88d702031" (UID: "0c5f5211-2520-45c1-acae-68e88d702031"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:19:04 crc kubenswrapper[4943]: I0307 15:19:04.010279 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c5f5211-2520-45c1-acae-68e88d702031-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "0c5f5211-2520-45c1-acae-68e88d702031" (UID: "0c5f5211-2520-45c1-acae-68e88d702031"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:19:04 crc kubenswrapper[4943]: I0307 15:19:04.018308 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c5f5211-2520-45c1-acae-68e88d702031-kube-api-access-d6vvd" (OuterVolumeSpecName: "kube-api-access-d6vvd") pod "0c5f5211-2520-45c1-acae-68e88d702031" (UID: "0c5f5211-2520-45c1-acae-68e88d702031"). InnerVolumeSpecName "kube-api-access-d6vvd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:19:04 crc kubenswrapper[4943]: I0307 15:19:04.034795 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c5f5211-2520-45c1-acae-68e88d702031-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "0c5f5211-2520-45c1-acae-68e88d702031" (UID: "0c5f5211-2520-45c1-acae-68e88d702031"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:19:04 crc kubenswrapper[4943]: I0307 15:19:04.035104 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c5f5211-2520-45c1-acae-68e88d702031-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "0c5f5211-2520-45c1-acae-68e88d702031" (UID: "0c5f5211-2520-45c1-acae-68e88d702031"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:19:04 crc kubenswrapper[4943]: I0307 15:19:04.052594 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c5f5211-2520-45c1-acae-68e88d702031-scripts" (OuterVolumeSpecName: "scripts") pod "0c5f5211-2520-45c1-acae-68e88d702031" (UID: "0c5f5211-2520-45c1-acae-68e88d702031"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:19:04 crc kubenswrapper[4943]: I0307 15:19:04.111434 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/0c5f5211-2520-45c1-acae-68e88d702031-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:04 crc kubenswrapper[4943]: I0307 15:19:04.111477 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/0c5f5211-2520-45c1-acae-68e88d702031-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:04 crc kubenswrapper[4943]: I0307 15:19:04.111492 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/0c5f5211-2520-45c1-acae-68e88d702031-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:04 crc kubenswrapper[4943]: I0307 15:19:04.111508 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6vvd\" (UniqueName: \"kubernetes.io/projected/0c5f5211-2520-45c1-acae-68e88d702031-kube-api-access-d6vvd\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:04 crc kubenswrapper[4943]: I0307 15:19:04.111521 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0c5f5211-2520-45c1-acae-68e88d702031-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:04 crc kubenswrapper[4943]: I0307 15:19:04.111531 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/0c5f5211-2520-45c1-acae-68e88d702031-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:04 crc kubenswrapper[4943]: I0307 15:19:04.554339 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="40342041ea06610208f01bba1b305aeb5fe2fc8838984730a56d7f08f990d629" Mar 07 15:19:04 crc kubenswrapper[4943]: I0307 15:19:04.554413 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6xvld" Mar 07 15:19:04 crc kubenswrapper[4943]: I0307 15:19:04.770120 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c5f5211-2520-45c1-acae-68e88d702031" path="/var/lib/kubelet/pods/0c5f5211-2520-45c1-acae-68e88d702031/volumes" Mar 07 15:19:05 crc kubenswrapper[4943]: I0307 15:19:05.140645 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-hq8ct"] Mar 07 15:19:05 crc kubenswrapper[4943]: E0307 15:19:05.141531 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c5f5211-2520-45c1-acae-68e88d702031" containerName="swift-ring-rebalance" Mar 07 15:19:05 crc kubenswrapper[4943]: I0307 15:19:05.141553 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c5f5211-2520-45c1-acae-68e88d702031" containerName="swift-ring-rebalance" Mar 07 15:19:05 crc kubenswrapper[4943]: I0307 15:19:05.141827 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c5f5211-2520-45c1-acae-68e88d702031" containerName="swift-ring-rebalance" Mar 07 15:19:05 crc kubenswrapper[4943]: I0307 15:19:05.142622 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-hq8ct" Mar 07 15:19:05 crc kubenswrapper[4943]: I0307 15:19:05.144865 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:19:05 crc kubenswrapper[4943]: I0307 15:19:05.145166 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:19:05 crc kubenswrapper[4943]: I0307 15:19:05.149300 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-hq8ct"] Mar 07 15:19:05 crc kubenswrapper[4943]: I0307 15:19:05.230204 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/24c70aa5-1e2e-4c8f-95d5-2cb9961021eb-etc-swift\") pod \"swift-ring-rebalance-debug-hq8ct\" (UID: \"24c70aa5-1e2e-4c8f-95d5-2cb9961021eb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hq8ct" Mar 07 15:19:05 crc kubenswrapper[4943]: I0307 15:19:05.230389 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/24c70aa5-1e2e-4c8f-95d5-2cb9961021eb-dispersionconf\") pod \"swift-ring-rebalance-debug-hq8ct\" (UID: \"24c70aa5-1e2e-4c8f-95d5-2cb9961021eb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hq8ct" Mar 07 15:19:05 crc kubenswrapper[4943]: I0307 15:19:05.230461 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/24c70aa5-1e2e-4c8f-95d5-2cb9961021eb-ring-data-devices\") pod \"swift-ring-rebalance-debug-hq8ct\" (UID: \"24c70aa5-1e2e-4c8f-95d5-2cb9961021eb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hq8ct" Mar 07 15:19:05 crc kubenswrapper[4943]: I0307 15:19:05.230613 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/24c70aa5-1e2e-4c8f-95d5-2cb9961021eb-scripts\") pod \"swift-ring-rebalance-debug-hq8ct\" (UID: \"24c70aa5-1e2e-4c8f-95d5-2cb9961021eb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hq8ct" Mar 07 15:19:05 crc kubenswrapper[4943]: I0307 15:19:05.230701 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/24c70aa5-1e2e-4c8f-95d5-2cb9961021eb-swiftconf\") pod \"swift-ring-rebalance-debug-hq8ct\" (UID: \"24c70aa5-1e2e-4c8f-95d5-2cb9961021eb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hq8ct" Mar 07 15:19:05 crc kubenswrapper[4943]: I0307 15:19:05.230860 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5f7b4\" (UniqueName: \"kubernetes.io/projected/24c70aa5-1e2e-4c8f-95d5-2cb9961021eb-kube-api-access-5f7b4\") pod \"swift-ring-rebalance-debug-hq8ct\" (UID: \"24c70aa5-1e2e-4c8f-95d5-2cb9961021eb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hq8ct" Mar 07 15:19:05 crc kubenswrapper[4943]: I0307 15:19:05.331986 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/24c70aa5-1e2e-4c8f-95d5-2cb9961021eb-ring-data-devices\") pod \"swift-ring-rebalance-debug-hq8ct\" (UID: \"24c70aa5-1e2e-4c8f-95d5-2cb9961021eb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hq8ct" Mar 07 15:19:05 crc kubenswrapper[4943]: I0307 15:19:05.332072 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/24c70aa5-1e2e-4c8f-95d5-2cb9961021eb-scripts\") pod \"swift-ring-rebalance-debug-hq8ct\" (UID: \"24c70aa5-1e2e-4c8f-95d5-2cb9961021eb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hq8ct" Mar 07 15:19:05 crc kubenswrapper[4943]: I0307 15:19:05.332110 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/24c70aa5-1e2e-4c8f-95d5-2cb9961021eb-swiftconf\") pod \"swift-ring-rebalance-debug-hq8ct\" (UID: \"24c70aa5-1e2e-4c8f-95d5-2cb9961021eb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hq8ct" Mar 07 15:19:05 crc kubenswrapper[4943]: I0307 15:19:05.332173 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5f7b4\" (UniqueName: \"kubernetes.io/projected/24c70aa5-1e2e-4c8f-95d5-2cb9961021eb-kube-api-access-5f7b4\") pod \"swift-ring-rebalance-debug-hq8ct\" (UID: \"24c70aa5-1e2e-4c8f-95d5-2cb9961021eb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hq8ct" Mar 07 15:19:05 crc kubenswrapper[4943]: I0307 15:19:05.332199 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/24c70aa5-1e2e-4c8f-95d5-2cb9961021eb-etc-swift\") pod \"swift-ring-rebalance-debug-hq8ct\" (UID: \"24c70aa5-1e2e-4c8f-95d5-2cb9961021eb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hq8ct" Mar 07 15:19:05 crc kubenswrapper[4943]: I0307 15:19:05.332255 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/24c70aa5-1e2e-4c8f-95d5-2cb9961021eb-dispersionconf\") pod \"swift-ring-rebalance-debug-hq8ct\" (UID: \"24c70aa5-1e2e-4c8f-95d5-2cb9961021eb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hq8ct" Mar 07 15:19:05 crc kubenswrapper[4943]: I0307 15:19:05.332899 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/24c70aa5-1e2e-4c8f-95d5-2cb9961021eb-etc-swift\") pod \"swift-ring-rebalance-debug-hq8ct\" (UID: \"24c70aa5-1e2e-4c8f-95d5-2cb9961021eb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hq8ct" Mar 07 15:19:05 crc kubenswrapper[4943]: I0307 15:19:05.333578 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/24c70aa5-1e2e-4c8f-95d5-2cb9961021eb-ring-data-devices\") pod \"swift-ring-rebalance-debug-hq8ct\" (UID: \"24c70aa5-1e2e-4c8f-95d5-2cb9961021eb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hq8ct" Mar 07 15:19:05 crc kubenswrapper[4943]: I0307 15:19:05.334042 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/24c70aa5-1e2e-4c8f-95d5-2cb9961021eb-scripts\") pod \"swift-ring-rebalance-debug-hq8ct\" (UID: \"24c70aa5-1e2e-4c8f-95d5-2cb9961021eb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hq8ct" Mar 07 15:19:05 crc kubenswrapper[4943]: I0307 15:19:05.339020 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/24c70aa5-1e2e-4c8f-95d5-2cb9961021eb-swiftconf\") pod \"swift-ring-rebalance-debug-hq8ct\" (UID: \"24c70aa5-1e2e-4c8f-95d5-2cb9961021eb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hq8ct" Mar 07 15:19:05 crc kubenswrapper[4943]: I0307 15:19:05.339160 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/24c70aa5-1e2e-4c8f-95d5-2cb9961021eb-dispersionconf\") pod \"swift-ring-rebalance-debug-hq8ct\" (UID: \"24c70aa5-1e2e-4c8f-95d5-2cb9961021eb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hq8ct" Mar 07 15:19:05 crc kubenswrapper[4943]: I0307 15:19:05.364137 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5f7b4\" (UniqueName: \"kubernetes.io/projected/24c70aa5-1e2e-4c8f-95d5-2cb9961021eb-kube-api-access-5f7b4\") pod \"swift-ring-rebalance-debug-hq8ct\" (UID: \"24c70aa5-1e2e-4c8f-95d5-2cb9961021eb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hq8ct" Mar 07 15:19:05 crc kubenswrapper[4943]: I0307 15:19:05.470555 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-hq8ct" Mar 07 15:19:05 crc kubenswrapper[4943]: I0307 15:19:05.771586 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-hq8ct"] Mar 07 15:19:06 crc kubenswrapper[4943]: I0307 15:19:06.576645 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-hq8ct" event={"ID":"24c70aa5-1e2e-4c8f-95d5-2cb9961021eb","Type":"ContainerStarted","Data":"056e6ef8ca10962682d5005ded8199f65bab3faf1b66a8bfb7135c94cbd0cb1a"} Mar 07 15:19:06 crc kubenswrapper[4943]: I0307 15:19:06.576994 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-hq8ct" event={"ID":"24c70aa5-1e2e-4c8f-95d5-2cb9961021eb","Type":"ContainerStarted","Data":"6062be4f6caa4fe393c972fe357460286c8657ad7c700b95ed2ccba125c2b76d"} Mar 07 15:19:06 crc kubenswrapper[4943]: I0307 15:19:06.598539 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-hq8ct" podStartSLOduration=1.5985188460000002 podStartE2EDuration="1.598518846s" podCreationTimestamp="2026-03-07 15:19:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:19:06.590712675 +0000 UTC m=+2388.542849183" watchObservedRunningTime="2026-03-07 15:19:06.598518846 +0000 UTC m=+2388.550655344" Mar 07 15:19:07 crc kubenswrapper[4943]: I0307 15:19:07.587363 4943 generic.go:334] "Generic (PLEG): container finished" podID="24c70aa5-1e2e-4c8f-95d5-2cb9961021eb" containerID="056e6ef8ca10962682d5005ded8199f65bab3faf1b66a8bfb7135c94cbd0cb1a" exitCode=0 Mar 07 15:19:07 crc kubenswrapper[4943]: I0307 15:19:07.587424 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-hq8ct" event={"ID":"24c70aa5-1e2e-4c8f-95d5-2cb9961021eb","Type":"ContainerDied","Data":"056e6ef8ca10962682d5005ded8199f65bab3faf1b66a8bfb7135c94cbd0cb1a"} Mar 07 15:19:08 crc kubenswrapper[4943]: I0307 15:19:08.972535 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-hq8ct" Mar 07 15:19:09 crc kubenswrapper[4943]: I0307 15:19:09.005243 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-hq8ct"] Mar 07 15:19:09 crc kubenswrapper[4943]: I0307 15:19:09.016398 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-hq8ct"] Mar 07 15:19:09 crc kubenswrapper[4943]: I0307 15:19:09.100273 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/24c70aa5-1e2e-4c8f-95d5-2cb9961021eb-etc-swift\") pod \"24c70aa5-1e2e-4c8f-95d5-2cb9961021eb\" (UID: \"24c70aa5-1e2e-4c8f-95d5-2cb9961021eb\") " Mar 07 15:19:09 crc kubenswrapper[4943]: I0307 15:19:09.100463 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/24c70aa5-1e2e-4c8f-95d5-2cb9961021eb-dispersionconf\") pod \"24c70aa5-1e2e-4c8f-95d5-2cb9961021eb\" (UID: \"24c70aa5-1e2e-4c8f-95d5-2cb9961021eb\") " Mar 07 15:19:09 crc kubenswrapper[4943]: I0307 15:19:09.100534 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/24c70aa5-1e2e-4c8f-95d5-2cb9961021eb-swiftconf\") pod \"24c70aa5-1e2e-4c8f-95d5-2cb9961021eb\" (UID: \"24c70aa5-1e2e-4c8f-95d5-2cb9961021eb\") " Mar 07 15:19:09 crc kubenswrapper[4943]: I0307 15:19:09.100581 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/24c70aa5-1e2e-4c8f-95d5-2cb9961021eb-scripts\") pod \"24c70aa5-1e2e-4c8f-95d5-2cb9961021eb\" (UID: \"24c70aa5-1e2e-4c8f-95d5-2cb9961021eb\") " Mar 07 15:19:09 crc kubenswrapper[4943]: I0307 15:19:09.100684 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5f7b4\" (UniqueName: \"kubernetes.io/projected/24c70aa5-1e2e-4c8f-95d5-2cb9961021eb-kube-api-access-5f7b4\") pod \"24c70aa5-1e2e-4c8f-95d5-2cb9961021eb\" (UID: \"24c70aa5-1e2e-4c8f-95d5-2cb9961021eb\") " Mar 07 15:19:09 crc kubenswrapper[4943]: I0307 15:19:09.100737 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/24c70aa5-1e2e-4c8f-95d5-2cb9961021eb-ring-data-devices\") pod \"24c70aa5-1e2e-4c8f-95d5-2cb9961021eb\" (UID: \"24c70aa5-1e2e-4c8f-95d5-2cb9961021eb\") " Mar 07 15:19:09 crc kubenswrapper[4943]: I0307 15:19:09.102333 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/24c70aa5-1e2e-4c8f-95d5-2cb9961021eb-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "24c70aa5-1e2e-4c8f-95d5-2cb9961021eb" (UID: "24c70aa5-1e2e-4c8f-95d5-2cb9961021eb"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:19:09 crc kubenswrapper[4943]: I0307 15:19:09.102394 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/24c70aa5-1e2e-4c8f-95d5-2cb9961021eb-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "24c70aa5-1e2e-4c8f-95d5-2cb9961021eb" (UID: "24c70aa5-1e2e-4c8f-95d5-2cb9961021eb"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:19:09 crc kubenswrapper[4943]: I0307 15:19:09.108218 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24c70aa5-1e2e-4c8f-95d5-2cb9961021eb-kube-api-access-5f7b4" (OuterVolumeSpecName: "kube-api-access-5f7b4") pod "24c70aa5-1e2e-4c8f-95d5-2cb9961021eb" (UID: "24c70aa5-1e2e-4c8f-95d5-2cb9961021eb"). InnerVolumeSpecName "kube-api-access-5f7b4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:19:09 crc kubenswrapper[4943]: I0307 15:19:09.141252 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/24c70aa5-1e2e-4c8f-95d5-2cb9961021eb-scripts" (OuterVolumeSpecName: "scripts") pod "24c70aa5-1e2e-4c8f-95d5-2cb9961021eb" (UID: "24c70aa5-1e2e-4c8f-95d5-2cb9961021eb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:19:09 crc kubenswrapper[4943]: I0307 15:19:09.141477 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24c70aa5-1e2e-4c8f-95d5-2cb9961021eb-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "24c70aa5-1e2e-4c8f-95d5-2cb9961021eb" (UID: "24c70aa5-1e2e-4c8f-95d5-2cb9961021eb"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:19:09 crc kubenswrapper[4943]: I0307 15:19:09.148186 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24c70aa5-1e2e-4c8f-95d5-2cb9961021eb-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "24c70aa5-1e2e-4c8f-95d5-2cb9961021eb" (UID: "24c70aa5-1e2e-4c8f-95d5-2cb9961021eb"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:19:09 crc kubenswrapper[4943]: I0307 15:19:09.203776 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/24c70aa5-1e2e-4c8f-95d5-2cb9961021eb-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:09 crc kubenswrapper[4943]: I0307 15:19:09.203827 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/24c70aa5-1e2e-4c8f-95d5-2cb9961021eb-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:09 crc kubenswrapper[4943]: I0307 15:19:09.203846 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/24c70aa5-1e2e-4c8f-95d5-2cb9961021eb-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:09 crc kubenswrapper[4943]: I0307 15:19:09.203864 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5f7b4\" (UniqueName: \"kubernetes.io/projected/24c70aa5-1e2e-4c8f-95d5-2cb9961021eb-kube-api-access-5f7b4\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:09 crc kubenswrapper[4943]: I0307 15:19:09.203884 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/24c70aa5-1e2e-4c8f-95d5-2cb9961021eb-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:09 crc kubenswrapper[4943]: I0307 15:19:09.203902 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/24c70aa5-1e2e-4c8f-95d5-2cb9961021eb-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:09 crc kubenswrapper[4943]: I0307 15:19:09.612287 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6062be4f6caa4fe393c972fe357460286c8657ad7c700b95ed2ccba125c2b76d" Mar 07 15:19:09 crc kubenswrapper[4943]: I0307 15:19:09.612375 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-hq8ct" Mar 07 15:19:09 crc kubenswrapper[4943]: I0307 15:19:09.755850 4943 scope.go:117] "RemoveContainer" containerID="448308ad6e34ac922d66b3c90e66d0927407a94dd28633ec350502d69e4ca8aa" Mar 07 15:19:09 crc kubenswrapper[4943]: E0307 15:19:09.756347 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cpdmm_openshift-machine-config-operator(4c1f18ea-e557-41fc-95dd-ce4e1722a61e)\"" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" Mar 07 15:19:10 crc kubenswrapper[4943]: I0307 15:19:10.242122 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-2mzkr"] Mar 07 15:19:10 crc kubenswrapper[4943]: E0307 15:19:10.242907 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24c70aa5-1e2e-4c8f-95d5-2cb9961021eb" containerName="swift-ring-rebalance" Mar 07 15:19:10 crc kubenswrapper[4943]: I0307 15:19:10.242962 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="24c70aa5-1e2e-4c8f-95d5-2cb9961021eb" containerName="swift-ring-rebalance" Mar 07 15:19:10 crc kubenswrapper[4943]: I0307 15:19:10.245291 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="24c70aa5-1e2e-4c8f-95d5-2cb9961021eb" containerName="swift-ring-rebalance" Mar 07 15:19:10 crc kubenswrapper[4943]: I0307 15:19:10.246527 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-2mzkr" Mar 07 15:19:10 crc kubenswrapper[4943]: I0307 15:19:10.250350 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:19:10 crc kubenswrapper[4943]: I0307 15:19:10.252347 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:19:10 crc kubenswrapper[4943]: I0307 15:19:10.267477 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-2mzkr"] Mar 07 15:19:10 crc kubenswrapper[4943]: I0307 15:19:10.425375 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/229acf6a-87a7-45b2-a6fa-d54ece8d8019-scripts\") pod \"swift-ring-rebalance-debug-2mzkr\" (UID: \"229acf6a-87a7-45b2-a6fa-d54ece8d8019\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2mzkr" Mar 07 15:19:10 crc kubenswrapper[4943]: I0307 15:19:10.425454 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/229acf6a-87a7-45b2-a6fa-d54ece8d8019-ring-data-devices\") pod \"swift-ring-rebalance-debug-2mzkr\" (UID: \"229acf6a-87a7-45b2-a6fa-d54ece8d8019\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2mzkr" Mar 07 15:19:10 crc kubenswrapper[4943]: I0307 15:19:10.425492 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/229acf6a-87a7-45b2-a6fa-d54ece8d8019-swiftconf\") pod \"swift-ring-rebalance-debug-2mzkr\" (UID: \"229acf6a-87a7-45b2-a6fa-d54ece8d8019\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2mzkr" Mar 07 15:19:10 crc kubenswrapper[4943]: I0307 15:19:10.425782 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/229acf6a-87a7-45b2-a6fa-d54ece8d8019-dispersionconf\") pod \"swift-ring-rebalance-debug-2mzkr\" (UID: \"229acf6a-87a7-45b2-a6fa-d54ece8d8019\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2mzkr" Mar 07 15:19:10 crc kubenswrapper[4943]: I0307 15:19:10.425905 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/229acf6a-87a7-45b2-a6fa-d54ece8d8019-etc-swift\") pod \"swift-ring-rebalance-debug-2mzkr\" (UID: \"229acf6a-87a7-45b2-a6fa-d54ece8d8019\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2mzkr" Mar 07 15:19:10 crc kubenswrapper[4943]: I0307 15:19:10.426053 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdnpm\" (UniqueName: \"kubernetes.io/projected/229acf6a-87a7-45b2-a6fa-d54ece8d8019-kube-api-access-fdnpm\") pod \"swift-ring-rebalance-debug-2mzkr\" (UID: \"229acf6a-87a7-45b2-a6fa-d54ece8d8019\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2mzkr" Mar 07 15:19:10 crc kubenswrapper[4943]: I0307 15:19:10.527480 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/229acf6a-87a7-45b2-a6fa-d54ece8d8019-scripts\") pod \"swift-ring-rebalance-debug-2mzkr\" (UID: \"229acf6a-87a7-45b2-a6fa-d54ece8d8019\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2mzkr" Mar 07 15:19:10 crc kubenswrapper[4943]: I0307 15:19:10.527599 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/229acf6a-87a7-45b2-a6fa-d54ece8d8019-ring-data-devices\") pod \"swift-ring-rebalance-debug-2mzkr\" (UID: \"229acf6a-87a7-45b2-a6fa-d54ece8d8019\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2mzkr" Mar 07 15:19:10 crc kubenswrapper[4943]: I0307 15:19:10.527816 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/229acf6a-87a7-45b2-a6fa-d54ece8d8019-swiftconf\") pod \"swift-ring-rebalance-debug-2mzkr\" (UID: \"229acf6a-87a7-45b2-a6fa-d54ece8d8019\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2mzkr" Mar 07 15:19:10 crc kubenswrapper[4943]: I0307 15:19:10.528171 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/229acf6a-87a7-45b2-a6fa-d54ece8d8019-dispersionconf\") pod \"swift-ring-rebalance-debug-2mzkr\" (UID: \"229acf6a-87a7-45b2-a6fa-d54ece8d8019\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2mzkr" Mar 07 15:19:10 crc kubenswrapper[4943]: I0307 15:19:10.528358 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/229acf6a-87a7-45b2-a6fa-d54ece8d8019-etc-swift\") pod \"swift-ring-rebalance-debug-2mzkr\" (UID: \"229acf6a-87a7-45b2-a6fa-d54ece8d8019\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2mzkr" Mar 07 15:19:10 crc kubenswrapper[4943]: I0307 15:19:10.528444 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdnpm\" (UniqueName: \"kubernetes.io/projected/229acf6a-87a7-45b2-a6fa-d54ece8d8019-kube-api-access-fdnpm\") pod \"swift-ring-rebalance-debug-2mzkr\" (UID: \"229acf6a-87a7-45b2-a6fa-d54ece8d8019\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2mzkr" Mar 07 15:19:10 crc kubenswrapper[4943]: I0307 15:19:10.528845 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/229acf6a-87a7-45b2-a6fa-d54ece8d8019-ring-data-devices\") pod \"swift-ring-rebalance-debug-2mzkr\" (UID: \"229acf6a-87a7-45b2-a6fa-d54ece8d8019\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2mzkr" Mar 07 15:19:10 crc kubenswrapper[4943]: I0307 15:19:10.529109 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/229acf6a-87a7-45b2-a6fa-d54ece8d8019-etc-swift\") pod \"swift-ring-rebalance-debug-2mzkr\" (UID: \"229acf6a-87a7-45b2-a6fa-d54ece8d8019\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2mzkr" Mar 07 15:19:10 crc kubenswrapper[4943]: I0307 15:19:10.530200 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/229acf6a-87a7-45b2-a6fa-d54ece8d8019-scripts\") pod \"swift-ring-rebalance-debug-2mzkr\" (UID: \"229acf6a-87a7-45b2-a6fa-d54ece8d8019\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2mzkr" Mar 07 15:19:10 crc kubenswrapper[4943]: I0307 15:19:10.535438 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/229acf6a-87a7-45b2-a6fa-d54ece8d8019-dispersionconf\") pod \"swift-ring-rebalance-debug-2mzkr\" (UID: \"229acf6a-87a7-45b2-a6fa-d54ece8d8019\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2mzkr" Mar 07 15:19:10 crc kubenswrapper[4943]: I0307 15:19:10.541611 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/229acf6a-87a7-45b2-a6fa-d54ece8d8019-swiftconf\") pod \"swift-ring-rebalance-debug-2mzkr\" (UID: \"229acf6a-87a7-45b2-a6fa-d54ece8d8019\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2mzkr" Mar 07 15:19:10 crc kubenswrapper[4943]: I0307 15:19:10.555581 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdnpm\" (UniqueName: \"kubernetes.io/projected/229acf6a-87a7-45b2-a6fa-d54ece8d8019-kube-api-access-fdnpm\") pod \"swift-ring-rebalance-debug-2mzkr\" (UID: \"229acf6a-87a7-45b2-a6fa-d54ece8d8019\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2mzkr" Mar 07 15:19:10 crc kubenswrapper[4943]: I0307 15:19:10.579026 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-2mzkr" Mar 07 15:19:10 crc kubenswrapper[4943]: I0307 15:19:10.773215 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24c70aa5-1e2e-4c8f-95d5-2cb9961021eb" path="/var/lib/kubelet/pods/24c70aa5-1e2e-4c8f-95d5-2cb9961021eb/volumes" Mar 07 15:19:11 crc kubenswrapper[4943]: I0307 15:19:11.081563 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-2mzkr"] Mar 07 15:19:11 crc kubenswrapper[4943]: I0307 15:19:11.634085 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-2mzkr" event={"ID":"229acf6a-87a7-45b2-a6fa-d54ece8d8019","Type":"ContainerStarted","Data":"dfd77d95691d128ef6b7a43d462d7203fe5ff87140cdd17cb672078fd27f5293"} Mar 07 15:19:11 crc kubenswrapper[4943]: I0307 15:19:11.634151 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-2mzkr" event={"ID":"229acf6a-87a7-45b2-a6fa-d54ece8d8019","Type":"ContainerStarted","Data":"0fa1fa6dae9425ae8762ac96a1bd42deb5cbf0a51e3ac331db2cb2d1df3d2df9"} Mar 07 15:19:11 crc kubenswrapper[4943]: I0307 15:19:11.674999 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-2mzkr" podStartSLOduration=1.674964803 podStartE2EDuration="1.674964803s" podCreationTimestamp="2026-03-07 15:19:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:19:11.662648312 +0000 UTC m=+2393.614784850" watchObservedRunningTime="2026-03-07 15:19:11.674964803 +0000 UTC m=+2393.627101341" Mar 07 15:19:12 crc kubenswrapper[4943]: I0307 15:19:12.654264 4943 generic.go:334] "Generic (PLEG): container finished" podID="229acf6a-87a7-45b2-a6fa-d54ece8d8019" containerID="dfd77d95691d128ef6b7a43d462d7203fe5ff87140cdd17cb672078fd27f5293" exitCode=0 Mar 07 15:19:12 crc kubenswrapper[4943]: I0307 15:19:12.654496 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-2mzkr" event={"ID":"229acf6a-87a7-45b2-a6fa-d54ece8d8019","Type":"ContainerDied","Data":"dfd77d95691d128ef6b7a43d462d7203fe5ff87140cdd17cb672078fd27f5293"} Mar 07 15:19:14 crc kubenswrapper[4943]: I0307 15:19:14.082596 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-2mzkr" Mar 07 15:19:14 crc kubenswrapper[4943]: I0307 15:19:14.145598 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-2mzkr"] Mar 07 15:19:14 crc kubenswrapper[4943]: I0307 15:19:14.150049 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-2mzkr"] Mar 07 15:19:14 crc kubenswrapper[4943]: I0307 15:19:14.195077 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/229acf6a-87a7-45b2-a6fa-d54ece8d8019-swiftconf\") pod \"229acf6a-87a7-45b2-a6fa-d54ece8d8019\" (UID: \"229acf6a-87a7-45b2-a6fa-d54ece8d8019\") " Mar 07 15:19:14 crc kubenswrapper[4943]: I0307 15:19:14.195164 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fdnpm\" (UniqueName: \"kubernetes.io/projected/229acf6a-87a7-45b2-a6fa-d54ece8d8019-kube-api-access-fdnpm\") pod \"229acf6a-87a7-45b2-a6fa-d54ece8d8019\" (UID: \"229acf6a-87a7-45b2-a6fa-d54ece8d8019\") " Mar 07 15:19:14 crc kubenswrapper[4943]: I0307 15:19:14.195201 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/229acf6a-87a7-45b2-a6fa-d54ece8d8019-ring-data-devices\") pod \"229acf6a-87a7-45b2-a6fa-d54ece8d8019\" (UID: \"229acf6a-87a7-45b2-a6fa-d54ece8d8019\") " Mar 07 15:19:14 crc kubenswrapper[4943]: I0307 15:19:14.195238 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/229acf6a-87a7-45b2-a6fa-d54ece8d8019-dispersionconf\") pod \"229acf6a-87a7-45b2-a6fa-d54ece8d8019\" (UID: \"229acf6a-87a7-45b2-a6fa-d54ece8d8019\") " Mar 07 15:19:14 crc kubenswrapper[4943]: I0307 15:19:14.195334 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/229acf6a-87a7-45b2-a6fa-d54ece8d8019-scripts\") pod \"229acf6a-87a7-45b2-a6fa-d54ece8d8019\" (UID: \"229acf6a-87a7-45b2-a6fa-d54ece8d8019\") " Mar 07 15:19:14 crc kubenswrapper[4943]: I0307 15:19:14.195372 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/229acf6a-87a7-45b2-a6fa-d54ece8d8019-etc-swift\") pod \"229acf6a-87a7-45b2-a6fa-d54ece8d8019\" (UID: \"229acf6a-87a7-45b2-a6fa-d54ece8d8019\") " Mar 07 15:19:14 crc kubenswrapper[4943]: I0307 15:19:14.196139 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/229acf6a-87a7-45b2-a6fa-d54ece8d8019-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "229acf6a-87a7-45b2-a6fa-d54ece8d8019" (UID: "229acf6a-87a7-45b2-a6fa-d54ece8d8019"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:19:14 crc kubenswrapper[4943]: I0307 15:19:14.196370 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/229acf6a-87a7-45b2-a6fa-d54ece8d8019-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "229acf6a-87a7-45b2-a6fa-d54ece8d8019" (UID: "229acf6a-87a7-45b2-a6fa-d54ece8d8019"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:19:14 crc kubenswrapper[4943]: I0307 15:19:14.202160 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/229acf6a-87a7-45b2-a6fa-d54ece8d8019-kube-api-access-fdnpm" (OuterVolumeSpecName: "kube-api-access-fdnpm") pod "229acf6a-87a7-45b2-a6fa-d54ece8d8019" (UID: "229acf6a-87a7-45b2-a6fa-d54ece8d8019"). InnerVolumeSpecName "kube-api-access-fdnpm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:19:14 crc kubenswrapper[4943]: I0307 15:19:14.219832 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/229acf6a-87a7-45b2-a6fa-d54ece8d8019-scripts" (OuterVolumeSpecName: "scripts") pod "229acf6a-87a7-45b2-a6fa-d54ece8d8019" (UID: "229acf6a-87a7-45b2-a6fa-d54ece8d8019"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:19:14 crc kubenswrapper[4943]: I0307 15:19:14.234434 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/229acf6a-87a7-45b2-a6fa-d54ece8d8019-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "229acf6a-87a7-45b2-a6fa-d54ece8d8019" (UID: "229acf6a-87a7-45b2-a6fa-d54ece8d8019"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:19:14 crc kubenswrapper[4943]: I0307 15:19:14.234647 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/229acf6a-87a7-45b2-a6fa-d54ece8d8019-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "229acf6a-87a7-45b2-a6fa-d54ece8d8019" (UID: "229acf6a-87a7-45b2-a6fa-d54ece8d8019"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:19:14 crc kubenswrapper[4943]: I0307 15:19:14.298461 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fdnpm\" (UniqueName: \"kubernetes.io/projected/229acf6a-87a7-45b2-a6fa-d54ece8d8019-kube-api-access-fdnpm\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:14 crc kubenswrapper[4943]: I0307 15:19:14.298517 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/229acf6a-87a7-45b2-a6fa-d54ece8d8019-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:14 crc kubenswrapper[4943]: I0307 15:19:14.298531 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/229acf6a-87a7-45b2-a6fa-d54ece8d8019-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:14 crc kubenswrapper[4943]: I0307 15:19:14.298543 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/229acf6a-87a7-45b2-a6fa-d54ece8d8019-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:14 crc kubenswrapper[4943]: I0307 15:19:14.298558 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/229acf6a-87a7-45b2-a6fa-d54ece8d8019-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:14 crc kubenswrapper[4943]: I0307 15:19:14.298576 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/229acf6a-87a7-45b2-a6fa-d54ece8d8019-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:14 crc kubenswrapper[4943]: I0307 15:19:14.689215 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0fa1fa6dae9425ae8762ac96a1bd42deb5cbf0a51e3ac331db2cb2d1df3d2df9" Mar 07 15:19:14 crc kubenswrapper[4943]: I0307 15:19:14.689364 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-2mzkr" Mar 07 15:19:14 crc kubenswrapper[4943]: I0307 15:19:14.770881 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="229acf6a-87a7-45b2-a6fa-d54ece8d8019" path="/var/lib/kubelet/pods/229acf6a-87a7-45b2-a6fa-d54ece8d8019/volumes" Mar 07 15:19:15 crc kubenswrapper[4943]: I0307 15:19:15.389611 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-8r9sv"] Mar 07 15:19:15 crc kubenswrapper[4943]: E0307 15:19:15.390595 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="229acf6a-87a7-45b2-a6fa-d54ece8d8019" containerName="swift-ring-rebalance" Mar 07 15:19:15 crc kubenswrapper[4943]: I0307 15:19:15.390623 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="229acf6a-87a7-45b2-a6fa-d54ece8d8019" containerName="swift-ring-rebalance" Mar 07 15:19:15 crc kubenswrapper[4943]: I0307 15:19:15.390909 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="229acf6a-87a7-45b2-a6fa-d54ece8d8019" containerName="swift-ring-rebalance" Mar 07 15:19:15 crc kubenswrapper[4943]: I0307 15:19:15.391799 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8r9sv" Mar 07 15:19:15 crc kubenswrapper[4943]: I0307 15:19:15.394729 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:19:15 crc kubenswrapper[4943]: I0307 15:19:15.395084 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:19:15 crc kubenswrapper[4943]: I0307 15:19:15.409681 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-8r9sv"] Mar 07 15:19:15 crc kubenswrapper[4943]: I0307 15:19:15.517698 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/4585133d-e3c5-4530-a96d-d7d4f28a4d20-dispersionconf\") pod \"swift-ring-rebalance-debug-8r9sv\" (UID: \"4585133d-e3c5-4530-a96d-d7d4f28a4d20\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8r9sv" Mar 07 15:19:15 crc kubenswrapper[4943]: I0307 15:19:15.517790 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/4585133d-e3c5-4530-a96d-d7d4f28a4d20-swiftconf\") pod \"swift-ring-rebalance-debug-8r9sv\" (UID: \"4585133d-e3c5-4530-a96d-d7d4f28a4d20\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8r9sv" Mar 07 15:19:15 crc kubenswrapper[4943]: I0307 15:19:15.517868 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/4585133d-e3c5-4530-a96d-d7d4f28a4d20-ring-data-devices\") pod \"swift-ring-rebalance-debug-8r9sv\" (UID: \"4585133d-e3c5-4530-a96d-d7d4f28a4d20\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8r9sv" Mar 07 15:19:15 crc kubenswrapper[4943]: I0307 15:19:15.518321 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8br4\" (UniqueName: \"kubernetes.io/projected/4585133d-e3c5-4530-a96d-d7d4f28a4d20-kube-api-access-l8br4\") pod \"swift-ring-rebalance-debug-8r9sv\" (UID: \"4585133d-e3c5-4530-a96d-d7d4f28a4d20\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8r9sv" Mar 07 15:19:15 crc kubenswrapper[4943]: I0307 15:19:15.518623 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/4585133d-e3c5-4530-a96d-d7d4f28a4d20-etc-swift\") pod \"swift-ring-rebalance-debug-8r9sv\" (UID: \"4585133d-e3c5-4530-a96d-d7d4f28a4d20\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8r9sv" Mar 07 15:19:15 crc kubenswrapper[4943]: I0307 15:19:15.518788 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4585133d-e3c5-4530-a96d-d7d4f28a4d20-scripts\") pod \"swift-ring-rebalance-debug-8r9sv\" (UID: \"4585133d-e3c5-4530-a96d-d7d4f28a4d20\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8r9sv" Mar 07 15:19:15 crc kubenswrapper[4943]: I0307 15:19:15.619834 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8br4\" (UniqueName: \"kubernetes.io/projected/4585133d-e3c5-4530-a96d-d7d4f28a4d20-kube-api-access-l8br4\") pod \"swift-ring-rebalance-debug-8r9sv\" (UID: \"4585133d-e3c5-4530-a96d-d7d4f28a4d20\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8r9sv" Mar 07 15:19:15 crc kubenswrapper[4943]: I0307 15:19:15.620004 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/4585133d-e3c5-4530-a96d-d7d4f28a4d20-etc-swift\") pod \"swift-ring-rebalance-debug-8r9sv\" (UID: \"4585133d-e3c5-4530-a96d-d7d4f28a4d20\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8r9sv" Mar 07 15:19:15 crc kubenswrapper[4943]: I0307 15:19:15.620075 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4585133d-e3c5-4530-a96d-d7d4f28a4d20-scripts\") pod \"swift-ring-rebalance-debug-8r9sv\" (UID: \"4585133d-e3c5-4530-a96d-d7d4f28a4d20\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8r9sv" Mar 07 15:19:15 crc kubenswrapper[4943]: I0307 15:19:15.620135 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/4585133d-e3c5-4530-a96d-d7d4f28a4d20-dispersionconf\") pod \"swift-ring-rebalance-debug-8r9sv\" (UID: \"4585133d-e3c5-4530-a96d-d7d4f28a4d20\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8r9sv" Mar 07 15:19:15 crc kubenswrapper[4943]: I0307 15:19:15.620183 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/4585133d-e3c5-4530-a96d-d7d4f28a4d20-swiftconf\") pod \"swift-ring-rebalance-debug-8r9sv\" (UID: \"4585133d-e3c5-4530-a96d-d7d4f28a4d20\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8r9sv" Mar 07 15:19:15 crc kubenswrapper[4943]: I0307 15:19:15.620245 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/4585133d-e3c5-4530-a96d-d7d4f28a4d20-ring-data-devices\") pod \"swift-ring-rebalance-debug-8r9sv\" (UID: \"4585133d-e3c5-4530-a96d-d7d4f28a4d20\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8r9sv" Mar 07 15:19:15 crc kubenswrapper[4943]: I0307 15:19:15.620780 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/4585133d-e3c5-4530-a96d-d7d4f28a4d20-etc-swift\") pod \"swift-ring-rebalance-debug-8r9sv\" (UID: \"4585133d-e3c5-4530-a96d-d7d4f28a4d20\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8r9sv" Mar 07 15:19:15 crc kubenswrapper[4943]: I0307 15:19:15.621697 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/4585133d-e3c5-4530-a96d-d7d4f28a4d20-ring-data-devices\") pod \"swift-ring-rebalance-debug-8r9sv\" (UID: \"4585133d-e3c5-4530-a96d-d7d4f28a4d20\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8r9sv" Mar 07 15:19:15 crc kubenswrapper[4943]: I0307 15:19:15.622405 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4585133d-e3c5-4530-a96d-d7d4f28a4d20-scripts\") pod \"swift-ring-rebalance-debug-8r9sv\" (UID: \"4585133d-e3c5-4530-a96d-d7d4f28a4d20\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8r9sv" Mar 07 15:19:15 crc kubenswrapper[4943]: I0307 15:19:15.629023 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/4585133d-e3c5-4530-a96d-d7d4f28a4d20-dispersionconf\") pod \"swift-ring-rebalance-debug-8r9sv\" (UID: \"4585133d-e3c5-4530-a96d-d7d4f28a4d20\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8r9sv" Mar 07 15:19:15 crc kubenswrapper[4943]: I0307 15:19:15.629105 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/4585133d-e3c5-4530-a96d-d7d4f28a4d20-swiftconf\") pod \"swift-ring-rebalance-debug-8r9sv\" (UID: \"4585133d-e3c5-4530-a96d-d7d4f28a4d20\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8r9sv" Mar 07 15:19:15 crc kubenswrapper[4943]: I0307 15:19:15.650704 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8br4\" (UniqueName: \"kubernetes.io/projected/4585133d-e3c5-4530-a96d-d7d4f28a4d20-kube-api-access-l8br4\") pod \"swift-ring-rebalance-debug-8r9sv\" (UID: \"4585133d-e3c5-4530-a96d-d7d4f28a4d20\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8r9sv" Mar 07 15:19:15 crc kubenswrapper[4943]: I0307 15:19:15.720857 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8r9sv" Mar 07 15:19:16 crc kubenswrapper[4943]: I0307 15:19:16.192175 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-8r9sv"] Mar 07 15:19:16 crc kubenswrapper[4943]: W0307 15:19:16.198972 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4585133d_e3c5_4530_a96d_d7d4f28a4d20.slice/crio-62efe15d6ad0b98a4497048a33e29efeb5e7be033e92e5d5b793db90a890b2c8 WatchSource:0}: Error finding container 62efe15d6ad0b98a4497048a33e29efeb5e7be033e92e5d5b793db90a890b2c8: Status 404 returned error can't find the container with id 62efe15d6ad0b98a4497048a33e29efeb5e7be033e92e5d5b793db90a890b2c8 Mar 07 15:19:16 crc kubenswrapper[4943]: I0307 15:19:16.730434 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8r9sv" event={"ID":"4585133d-e3c5-4530-a96d-d7d4f28a4d20","Type":"ContainerStarted","Data":"0623eb03d59763f4f564529b3dcc4f43445fc960a2f8ef5de86205aa98d6fad1"} Mar 07 15:19:16 crc kubenswrapper[4943]: I0307 15:19:16.732063 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8r9sv" event={"ID":"4585133d-e3c5-4530-a96d-d7d4f28a4d20","Type":"ContainerStarted","Data":"62efe15d6ad0b98a4497048a33e29efeb5e7be033e92e5d5b793db90a890b2c8"} Mar 07 15:19:16 crc kubenswrapper[4943]: I0307 15:19:16.762232 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8r9sv" podStartSLOduration=1.762214033 podStartE2EDuration="1.762214033s" podCreationTimestamp="2026-03-07 15:19:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:19:16.759330153 +0000 UTC m=+2398.711466661" watchObservedRunningTime="2026-03-07 15:19:16.762214033 +0000 UTC m=+2398.714350541" Mar 07 15:19:18 crc kubenswrapper[4943]: I0307 15:19:18.764692 4943 generic.go:334] "Generic (PLEG): container finished" podID="4585133d-e3c5-4530-a96d-d7d4f28a4d20" containerID="0623eb03d59763f4f564529b3dcc4f43445fc960a2f8ef5de86205aa98d6fad1" exitCode=0 Mar 07 15:19:18 crc kubenswrapper[4943]: I0307 15:19:18.776245 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8r9sv" event={"ID":"4585133d-e3c5-4530-a96d-d7d4f28a4d20","Type":"ContainerDied","Data":"0623eb03d59763f4f564529b3dcc4f43445fc960a2f8ef5de86205aa98d6fad1"} Mar 07 15:19:20 crc kubenswrapper[4943]: I0307 15:19:20.145898 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8r9sv" Mar 07 15:19:20 crc kubenswrapper[4943]: I0307 15:19:20.187299 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-8r9sv"] Mar 07 15:19:20 crc kubenswrapper[4943]: I0307 15:19:20.194700 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-8r9sv"] Mar 07 15:19:20 crc kubenswrapper[4943]: I0307 15:19:20.303465 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/4585133d-e3c5-4530-a96d-d7d4f28a4d20-swiftconf\") pod \"4585133d-e3c5-4530-a96d-d7d4f28a4d20\" (UID: \"4585133d-e3c5-4530-a96d-d7d4f28a4d20\") " Mar 07 15:19:20 crc kubenswrapper[4943]: I0307 15:19:20.303580 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/4585133d-e3c5-4530-a96d-d7d4f28a4d20-etc-swift\") pod \"4585133d-e3c5-4530-a96d-d7d4f28a4d20\" (UID: \"4585133d-e3c5-4530-a96d-d7d4f28a4d20\") " Mar 07 15:19:20 crc kubenswrapper[4943]: I0307 15:19:20.303643 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/4585133d-e3c5-4530-a96d-d7d4f28a4d20-ring-data-devices\") pod \"4585133d-e3c5-4530-a96d-d7d4f28a4d20\" (UID: \"4585133d-e3c5-4530-a96d-d7d4f28a4d20\") " Mar 07 15:19:20 crc kubenswrapper[4943]: I0307 15:19:20.303687 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/4585133d-e3c5-4530-a96d-d7d4f28a4d20-dispersionconf\") pod \"4585133d-e3c5-4530-a96d-d7d4f28a4d20\" (UID: \"4585133d-e3c5-4530-a96d-d7d4f28a4d20\") " Mar 07 15:19:20 crc kubenswrapper[4943]: I0307 15:19:20.303727 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l8br4\" (UniqueName: \"kubernetes.io/projected/4585133d-e3c5-4530-a96d-d7d4f28a4d20-kube-api-access-l8br4\") pod \"4585133d-e3c5-4530-a96d-d7d4f28a4d20\" (UID: \"4585133d-e3c5-4530-a96d-d7d4f28a4d20\") " Mar 07 15:19:20 crc kubenswrapper[4943]: I0307 15:19:20.303749 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4585133d-e3c5-4530-a96d-d7d4f28a4d20-scripts\") pod \"4585133d-e3c5-4530-a96d-d7d4f28a4d20\" (UID: \"4585133d-e3c5-4530-a96d-d7d4f28a4d20\") " Mar 07 15:19:20 crc kubenswrapper[4943]: I0307 15:19:20.304685 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4585133d-e3c5-4530-a96d-d7d4f28a4d20-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "4585133d-e3c5-4530-a96d-d7d4f28a4d20" (UID: "4585133d-e3c5-4530-a96d-d7d4f28a4d20"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:19:20 crc kubenswrapper[4943]: I0307 15:19:20.304739 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4585133d-e3c5-4530-a96d-d7d4f28a4d20-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "4585133d-e3c5-4530-a96d-d7d4f28a4d20" (UID: "4585133d-e3c5-4530-a96d-d7d4f28a4d20"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:19:20 crc kubenswrapper[4943]: I0307 15:19:20.308979 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4585133d-e3c5-4530-a96d-d7d4f28a4d20-kube-api-access-l8br4" (OuterVolumeSpecName: "kube-api-access-l8br4") pod "4585133d-e3c5-4530-a96d-d7d4f28a4d20" (UID: "4585133d-e3c5-4530-a96d-d7d4f28a4d20"). InnerVolumeSpecName "kube-api-access-l8br4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:19:20 crc kubenswrapper[4943]: I0307 15:19:20.326165 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4585133d-e3c5-4530-a96d-d7d4f28a4d20-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "4585133d-e3c5-4530-a96d-d7d4f28a4d20" (UID: "4585133d-e3c5-4530-a96d-d7d4f28a4d20"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:19:20 crc kubenswrapper[4943]: I0307 15:19:20.339727 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4585133d-e3c5-4530-a96d-d7d4f28a4d20-scripts" (OuterVolumeSpecName: "scripts") pod "4585133d-e3c5-4530-a96d-d7d4f28a4d20" (UID: "4585133d-e3c5-4530-a96d-d7d4f28a4d20"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:19:20 crc kubenswrapper[4943]: I0307 15:19:20.342326 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4585133d-e3c5-4530-a96d-d7d4f28a4d20-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "4585133d-e3c5-4530-a96d-d7d4f28a4d20" (UID: "4585133d-e3c5-4530-a96d-d7d4f28a4d20"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:19:20 crc kubenswrapper[4943]: I0307 15:19:20.405094 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l8br4\" (UniqueName: \"kubernetes.io/projected/4585133d-e3c5-4530-a96d-d7d4f28a4d20-kube-api-access-l8br4\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:20 crc kubenswrapper[4943]: I0307 15:19:20.405130 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4585133d-e3c5-4530-a96d-d7d4f28a4d20-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:20 crc kubenswrapper[4943]: I0307 15:19:20.405138 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/4585133d-e3c5-4530-a96d-d7d4f28a4d20-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:20 crc kubenswrapper[4943]: I0307 15:19:20.405146 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/4585133d-e3c5-4530-a96d-d7d4f28a4d20-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:20 crc kubenswrapper[4943]: I0307 15:19:20.405154 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/4585133d-e3c5-4530-a96d-d7d4f28a4d20-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:20 crc kubenswrapper[4943]: I0307 15:19:20.405162 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/4585133d-e3c5-4530-a96d-d7d4f28a4d20-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:20 crc kubenswrapper[4943]: I0307 15:19:20.772754 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4585133d-e3c5-4530-a96d-d7d4f28a4d20" path="/var/lib/kubelet/pods/4585133d-e3c5-4530-a96d-d7d4f28a4d20/volumes" Mar 07 15:19:20 crc kubenswrapper[4943]: I0307 15:19:20.787245 4943 scope.go:117] "RemoveContainer" containerID="0623eb03d59763f4f564529b3dcc4f43445fc960a2f8ef5de86205aa98d6fad1" Mar 07 15:19:20 crc kubenswrapper[4943]: I0307 15:19:20.787286 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8r9sv" Mar 07 15:19:21 crc kubenswrapper[4943]: I0307 15:19:21.330885 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-n7xkh"] Mar 07 15:19:21 crc kubenswrapper[4943]: E0307 15:19:21.331537 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4585133d-e3c5-4530-a96d-d7d4f28a4d20" containerName="swift-ring-rebalance" Mar 07 15:19:21 crc kubenswrapper[4943]: I0307 15:19:21.331552 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="4585133d-e3c5-4530-a96d-d7d4f28a4d20" containerName="swift-ring-rebalance" Mar 07 15:19:21 crc kubenswrapper[4943]: I0307 15:19:21.331711 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="4585133d-e3c5-4530-a96d-d7d4f28a4d20" containerName="swift-ring-rebalance" Mar 07 15:19:21 crc kubenswrapper[4943]: I0307 15:19:21.332200 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-n7xkh" Mar 07 15:19:21 crc kubenswrapper[4943]: I0307 15:19:21.335491 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:19:21 crc kubenswrapper[4943]: I0307 15:19:21.336227 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:19:21 crc kubenswrapper[4943]: I0307 15:19:21.343356 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-n7xkh"] Mar 07 15:19:21 crc kubenswrapper[4943]: I0307 15:19:21.421838 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb-dispersionconf\") pod \"swift-ring-rebalance-debug-n7xkh\" (UID: \"b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-n7xkh" Mar 07 15:19:21 crc kubenswrapper[4943]: I0307 15:19:21.421903 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4wtz\" (UniqueName: \"kubernetes.io/projected/b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb-kube-api-access-z4wtz\") pod \"swift-ring-rebalance-debug-n7xkh\" (UID: \"b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-n7xkh" Mar 07 15:19:21 crc kubenswrapper[4943]: I0307 15:19:21.421967 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb-scripts\") pod \"swift-ring-rebalance-debug-n7xkh\" (UID: \"b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-n7xkh" Mar 07 15:19:21 crc kubenswrapper[4943]: I0307 15:19:21.422004 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb-swiftconf\") pod \"swift-ring-rebalance-debug-n7xkh\" (UID: \"b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-n7xkh" Mar 07 15:19:21 crc kubenswrapper[4943]: I0307 15:19:21.422026 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb-etc-swift\") pod \"swift-ring-rebalance-debug-n7xkh\" (UID: \"b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-n7xkh" Mar 07 15:19:21 crc kubenswrapper[4943]: I0307 15:19:21.422051 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb-ring-data-devices\") pod \"swift-ring-rebalance-debug-n7xkh\" (UID: \"b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-n7xkh" Mar 07 15:19:21 crc kubenswrapper[4943]: I0307 15:19:21.523599 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb-scripts\") pod \"swift-ring-rebalance-debug-n7xkh\" (UID: \"b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-n7xkh" Mar 07 15:19:21 crc kubenswrapper[4943]: I0307 15:19:21.523708 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb-swiftconf\") pod \"swift-ring-rebalance-debug-n7xkh\" (UID: \"b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-n7xkh" Mar 07 15:19:21 crc kubenswrapper[4943]: I0307 15:19:21.523752 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb-etc-swift\") pod \"swift-ring-rebalance-debug-n7xkh\" (UID: \"b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-n7xkh" Mar 07 15:19:21 crc kubenswrapper[4943]: I0307 15:19:21.523797 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb-ring-data-devices\") pod \"swift-ring-rebalance-debug-n7xkh\" (UID: \"b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-n7xkh" Mar 07 15:19:21 crc kubenswrapper[4943]: I0307 15:19:21.523842 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb-dispersionconf\") pod \"swift-ring-rebalance-debug-n7xkh\" (UID: \"b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-n7xkh" Mar 07 15:19:21 crc kubenswrapper[4943]: I0307 15:19:21.523896 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4wtz\" (UniqueName: \"kubernetes.io/projected/b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb-kube-api-access-z4wtz\") pod \"swift-ring-rebalance-debug-n7xkh\" (UID: \"b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-n7xkh" Mar 07 15:19:21 crc kubenswrapper[4943]: I0307 15:19:21.524452 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb-scripts\") pod \"swift-ring-rebalance-debug-n7xkh\" (UID: \"b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-n7xkh" Mar 07 15:19:21 crc kubenswrapper[4943]: I0307 15:19:21.524545 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb-etc-swift\") pod \"swift-ring-rebalance-debug-n7xkh\" (UID: \"b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-n7xkh" Mar 07 15:19:21 crc kubenswrapper[4943]: I0307 15:19:21.524641 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb-ring-data-devices\") pod \"swift-ring-rebalance-debug-n7xkh\" (UID: \"b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-n7xkh" Mar 07 15:19:21 crc kubenswrapper[4943]: I0307 15:19:21.529864 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb-dispersionconf\") pod \"swift-ring-rebalance-debug-n7xkh\" (UID: \"b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-n7xkh" Mar 07 15:19:21 crc kubenswrapper[4943]: I0307 15:19:21.529915 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb-swiftconf\") pod \"swift-ring-rebalance-debug-n7xkh\" (UID: \"b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-n7xkh" Mar 07 15:19:21 crc kubenswrapper[4943]: I0307 15:19:21.548863 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4wtz\" (UniqueName: \"kubernetes.io/projected/b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb-kube-api-access-z4wtz\") pod \"swift-ring-rebalance-debug-n7xkh\" (UID: \"b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-n7xkh" Mar 07 15:19:21 crc kubenswrapper[4943]: I0307 15:19:21.691714 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-n7xkh" Mar 07 15:19:22 crc kubenswrapper[4943]: I0307 15:19:22.152575 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-n7xkh"] Mar 07 15:19:22 crc kubenswrapper[4943]: W0307 15:19:22.165244 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb2d676e7_82d6_4fc3_adf2_39f0bdb2fdbb.slice/crio-1cc9d7e3fc90688c430034cc838efbc641686b971b9124a07ceba274754a8e9c WatchSource:0}: Error finding container 1cc9d7e3fc90688c430034cc838efbc641686b971b9124a07ceba274754a8e9c: Status 404 returned error can't find the container with id 1cc9d7e3fc90688c430034cc838efbc641686b971b9124a07ceba274754a8e9c Mar 07 15:19:22 crc kubenswrapper[4943]: I0307 15:19:22.814938 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-n7xkh" event={"ID":"b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb","Type":"ContainerStarted","Data":"dba7383a12bbbc41263368a88607ad7cdea69d0d7123bff82132321a508f4ffc"} Mar 07 15:19:22 crc kubenswrapper[4943]: I0307 15:19:22.815343 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-n7xkh" event={"ID":"b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb","Type":"ContainerStarted","Data":"1cc9d7e3fc90688c430034cc838efbc641686b971b9124a07ceba274754a8e9c"} Mar 07 15:19:22 crc kubenswrapper[4943]: I0307 15:19:22.850752 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-n7xkh" podStartSLOduration=1.85073345 podStartE2EDuration="1.85073345s" podCreationTimestamp="2026-03-07 15:19:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:19:22.840649302 +0000 UTC m=+2404.792785810" watchObservedRunningTime="2026-03-07 15:19:22.85073345 +0000 UTC m=+2404.802869958" Mar 07 15:19:23 crc kubenswrapper[4943]: I0307 15:19:23.569968 4943 scope.go:117] "RemoveContainer" containerID="7ab86a755eb914c5752287227d41b72269aced522ad634d946657faf7f98e060" Mar 07 15:19:23 crc kubenswrapper[4943]: I0307 15:19:23.604921 4943 scope.go:117] "RemoveContainer" containerID="8dba4aa6f09b0fe5fd0bfa8f4abad1b0555ab0d6d2cf190e7b2a9f200fbddb08" Mar 07 15:19:23 crc kubenswrapper[4943]: I0307 15:19:23.650990 4943 scope.go:117] "RemoveContainer" containerID="20ffa016619f0915dfcc66196b10b19a28e3969cb9c044c1763d00640683bcba" Mar 07 15:19:23 crc kubenswrapper[4943]: I0307 15:19:23.699900 4943 scope.go:117] "RemoveContainer" containerID="bd8ffc2cc6457280f506fef1ee2317694d39880dd2efc16a67fc94cb69e6f936" Mar 07 15:19:23 crc kubenswrapper[4943]: I0307 15:19:23.740883 4943 scope.go:117] "RemoveContainer" containerID="724090a96661631af14d84e2e4a041922bff6bdf27022c51d53e1d415be01555" Mar 07 15:19:23 crc kubenswrapper[4943]: I0307 15:19:23.755690 4943 scope.go:117] "RemoveContainer" containerID="448308ad6e34ac922d66b3c90e66d0927407a94dd28633ec350502d69e4ca8aa" Mar 07 15:19:23 crc kubenswrapper[4943]: E0307 15:19:23.755988 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cpdmm_openshift-machine-config-operator(4c1f18ea-e557-41fc-95dd-ce4e1722a61e)\"" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" Mar 07 15:19:23 crc kubenswrapper[4943]: I0307 15:19:23.781511 4943 scope.go:117] "RemoveContainer" containerID="d6a31cdb138264f6507f5de9000244ab3aeca8f599d047e044379d49425395da" Mar 07 15:19:23 crc kubenswrapper[4943]: I0307 15:19:23.827262 4943 generic.go:334] "Generic (PLEG): container finished" podID="b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb" containerID="dba7383a12bbbc41263368a88607ad7cdea69d0d7123bff82132321a508f4ffc" exitCode=0 Mar 07 15:19:23 crc kubenswrapper[4943]: I0307 15:19:23.827348 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-n7xkh" event={"ID":"b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb","Type":"ContainerDied","Data":"dba7383a12bbbc41263368a88607ad7cdea69d0d7123bff82132321a508f4ffc"} Mar 07 15:19:23 crc kubenswrapper[4943]: I0307 15:19:23.888839 4943 scope.go:117] "RemoveContainer" containerID="2d04fcdab766b2b930785d656361aae896447910660dbeb62ee7e4215298cf07" Mar 07 15:19:23 crc kubenswrapper[4943]: I0307 15:19:23.946206 4943 scope.go:117] "RemoveContainer" containerID="5ef4437771ebe7f583d725cefda637736677ea6c6a02e7b62dac24c36a88ebf1" Mar 07 15:19:25 crc kubenswrapper[4943]: I0307 15:19:25.161855 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-n7xkh" Mar 07 15:19:25 crc kubenswrapper[4943]: I0307 15:19:25.223205 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-n7xkh"] Mar 07 15:19:25 crc kubenswrapper[4943]: I0307 15:19:25.234231 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-n7xkh"] Mar 07 15:19:25 crc kubenswrapper[4943]: I0307 15:19:25.283568 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb-ring-data-devices\") pod \"b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb\" (UID: \"b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb\") " Mar 07 15:19:25 crc kubenswrapper[4943]: I0307 15:19:25.284344 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb-swiftconf\") pod \"b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb\" (UID: \"b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb\") " Mar 07 15:19:25 crc kubenswrapper[4943]: I0307 15:19:25.284580 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb" (UID: "b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:19:25 crc kubenswrapper[4943]: I0307 15:19:25.284769 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb-dispersionconf\") pod \"b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb\" (UID: \"b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb\") " Mar 07 15:19:25 crc kubenswrapper[4943]: I0307 15:19:25.285085 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb-etc-swift\") pod \"b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb\" (UID: \"b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb\") " Mar 07 15:19:25 crc kubenswrapper[4943]: I0307 15:19:25.286356 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb" (UID: "b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:19:25 crc kubenswrapper[4943]: I0307 15:19:25.286663 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb-scripts\") pod \"b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb\" (UID: \"b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb\") " Mar 07 15:19:25 crc kubenswrapper[4943]: I0307 15:19:25.286819 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z4wtz\" (UniqueName: \"kubernetes.io/projected/b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb-kube-api-access-z4wtz\") pod \"b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb\" (UID: \"b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb\") " Mar 07 15:19:25 crc kubenswrapper[4943]: I0307 15:19:25.287665 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:25 crc kubenswrapper[4943]: I0307 15:19:25.287710 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:25 crc kubenswrapper[4943]: I0307 15:19:25.291224 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb-kube-api-access-z4wtz" (OuterVolumeSpecName: "kube-api-access-z4wtz") pod "b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb" (UID: "b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb"). InnerVolumeSpecName "kube-api-access-z4wtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:19:25 crc kubenswrapper[4943]: I0307 15:19:25.319520 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb-scripts" (OuterVolumeSpecName: "scripts") pod "b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb" (UID: "b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:19:25 crc kubenswrapper[4943]: I0307 15:19:25.330643 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb" (UID: "b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:19:25 crc kubenswrapper[4943]: I0307 15:19:25.334782 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb" (UID: "b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:19:25 crc kubenswrapper[4943]: I0307 15:19:25.389910 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:25 crc kubenswrapper[4943]: I0307 15:19:25.390012 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:25 crc kubenswrapper[4943]: I0307 15:19:25.390041 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z4wtz\" (UniqueName: \"kubernetes.io/projected/b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb-kube-api-access-z4wtz\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:25 crc kubenswrapper[4943]: I0307 15:19:25.390074 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:25 crc kubenswrapper[4943]: I0307 15:19:25.865265 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1cc9d7e3fc90688c430034cc838efbc641686b971b9124a07ceba274754a8e9c" Mar 07 15:19:25 crc kubenswrapper[4943]: I0307 15:19:25.865371 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-n7xkh" Mar 07 15:19:26 crc kubenswrapper[4943]: I0307 15:19:26.455839 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-b6cwb"] Mar 07 15:19:26 crc kubenswrapper[4943]: E0307 15:19:26.456271 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb" containerName="swift-ring-rebalance" Mar 07 15:19:26 crc kubenswrapper[4943]: I0307 15:19:26.456287 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb" containerName="swift-ring-rebalance" Mar 07 15:19:26 crc kubenswrapper[4943]: I0307 15:19:26.456447 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb" containerName="swift-ring-rebalance" Mar 07 15:19:26 crc kubenswrapper[4943]: I0307 15:19:26.457079 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-b6cwb" Mar 07 15:19:26 crc kubenswrapper[4943]: I0307 15:19:26.466324 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:19:26 crc kubenswrapper[4943]: I0307 15:19:26.466354 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:19:26 crc kubenswrapper[4943]: I0307 15:19:26.483293 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-b6cwb"] Mar 07 15:19:26 crc kubenswrapper[4943]: I0307 15:19:26.610877 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd-dispersionconf\") pod \"swift-ring-rebalance-debug-b6cwb\" (UID: \"f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-b6cwb" Mar 07 15:19:26 crc kubenswrapper[4943]: I0307 15:19:26.611001 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd-swiftconf\") pod \"swift-ring-rebalance-debug-b6cwb\" (UID: \"f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-b6cwb" Mar 07 15:19:26 crc kubenswrapper[4943]: I0307 15:19:26.611051 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcddl\" (UniqueName: \"kubernetes.io/projected/f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd-kube-api-access-mcddl\") pod \"swift-ring-rebalance-debug-b6cwb\" (UID: \"f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-b6cwb" Mar 07 15:19:26 crc kubenswrapper[4943]: I0307 15:19:26.611093 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd-ring-data-devices\") pod \"swift-ring-rebalance-debug-b6cwb\" (UID: \"f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-b6cwb" Mar 07 15:19:26 crc kubenswrapper[4943]: I0307 15:19:26.611201 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd-scripts\") pod \"swift-ring-rebalance-debug-b6cwb\" (UID: \"f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-b6cwb" Mar 07 15:19:26 crc kubenswrapper[4943]: I0307 15:19:26.611293 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd-etc-swift\") pod \"swift-ring-rebalance-debug-b6cwb\" (UID: \"f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-b6cwb" Mar 07 15:19:26 crc kubenswrapper[4943]: I0307 15:19:26.713030 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd-scripts\") pod \"swift-ring-rebalance-debug-b6cwb\" (UID: \"f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-b6cwb" Mar 07 15:19:26 crc kubenswrapper[4943]: I0307 15:19:26.713263 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd-etc-swift\") pod \"swift-ring-rebalance-debug-b6cwb\" (UID: \"f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-b6cwb" Mar 07 15:19:26 crc kubenswrapper[4943]: I0307 15:19:26.713339 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd-dispersionconf\") pod \"swift-ring-rebalance-debug-b6cwb\" (UID: \"f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-b6cwb" Mar 07 15:19:26 crc kubenswrapper[4943]: I0307 15:19:26.713420 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd-swiftconf\") pod \"swift-ring-rebalance-debug-b6cwb\" (UID: \"f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-b6cwb" Mar 07 15:19:26 crc kubenswrapper[4943]: I0307 15:19:26.713488 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcddl\" (UniqueName: \"kubernetes.io/projected/f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd-kube-api-access-mcddl\") pod \"swift-ring-rebalance-debug-b6cwb\" (UID: \"f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-b6cwb" Mar 07 15:19:26 crc kubenswrapper[4943]: I0307 15:19:26.713546 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd-ring-data-devices\") pod \"swift-ring-rebalance-debug-b6cwb\" (UID: \"f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-b6cwb" Mar 07 15:19:26 crc kubenswrapper[4943]: I0307 15:19:26.714477 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd-scripts\") pod \"swift-ring-rebalance-debug-b6cwb\" (UID: \"f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-b6cwb" Mar 07 15:19:26 crc kubenswrapper[4943]: I0307 15:19:26.714806 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd-etc-swift\") pod \"swift-ring-rebalance-debug-b6cwb\" (UID: \"f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-b6cwb" Mar 07 15:19:26 crc kubenswrapper[4943]: I0307 15:19:26.715310 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd-ring-data-devices\") pod \"swift-ring-rebalance-debug-b6cwb\" (UID: \"f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-b6cwb" Mar 07 15:19:26 crc kubenswrapper[4943]: I0307 15:19:26.719583 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd-dispersionconf\") pod \"swift-ring-rebalance-debug-b6cwb\" (UID: \"f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-b6cwb" Mar 07 15:19:26 crc kubenswrapper[4943]: I0307 15:19:26.722471 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd-swiftconf\") pod \"swift-ring-rebalance-debug-b6cwb\" (UID: \"f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-b6cwb" Mar 07 15:19:26 crc kubenswrapper[4943]: I0307 15:19:26.740121 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcddl\" (UniqueName: \"kubernetes.io/projected/f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd-kube-api-access-mcddl\") pod \"swift-ring-rebalance-debug-b6cwb\" (UID: \"f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-b6cwb" Mar 07 15:19:26 crc kubenswrapper[4943]: I0307 15:19:26.773168 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb" path="/var/lib/kubelet/pods/b2d676e7-82d6-4fc3-adf2-39f0bdb2fdbb/volumes" Mar 07 15:19:26 crc kubenswrapper[4943]: I0307 15:19:26.788687 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-b6cwb" Mar 07 15:19:27 crc kubenswrapper[4943]: I0307 15:19:27.052148 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-b6cwb"] Mar 07 15:19:27 crc kubenswrapper[4943]: I0307 15:19:27.893536 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-b6cwb" event={"ID":"f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd","Type":"ContainerStarted","Data":"b8d0009e5cc84db560d54a6c0980a4cba727f318d70b9f8e0b4378e77218a550"} Mar 07 15:19:27 crc kubenswrapper[4943]: I0307 15:19:27.894191 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-b6cwb" event={"ID":"f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd","Type":"ContainerStarted","Data":"a255e9a6f481abfeb44b87c1c60578341471a60c4f5d814cc1b8d9bff63cbb88"} Mar 07 15:19:27 crc kubenswrapper[4943]: I0307 15:19:27.920145 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-b6cwb" podStartSLOduration=1.9201120459999999 podStartE2EDuration="1.920112046s" podCreationTimestamp="2026-03-07 15:19:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:19:27.917625585 +0000 UTC m=+2409.869762123" watchObservedRunningTime="2026-03-07 15:19:27.920112046 +0000 UTC m=+2409.872248584" Mar 07 15:19:28 crc kubenswrapper[4943]: I0307 15:19:28.908031 4943 generic.go:334] "Generic (PLEG): container finished" podID="f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd" containerID="b8d0009e5cc84db560d54a6c0980a4cba727f318d70b9f8e0b4378e77218a550" exitCode=0 Mar 07 15:19:28 crc kubenswrapper[4943]: I0307 15:19:28.908109 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-b6cwb" event={"ID":"f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd","Type":"ContainerDied","Data":"b8d0009e5cc84db560d54a6c0980a4cba727f318d70b9f8e0b4378e77218a550"} Mar 07 15:19:30 crc kubenswrapper[4943]: I0307 15:19:30.225392 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-b6cwb" Mar 07 15:19:30 crc kubenswrapper[4943]: I0307 15:19:30.259207 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-b6cwb"] Mar 07 15:19:30 crc kubenswrapper[4943]: I0307 15:19:30.264192 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-b6cwb"] Mar 07 15:19:30 crc kubenswrapper[4943]: I0307 15:19:30.386547 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mcddl\" (UniqueName: \"kubernetes.io/projected/f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd-kube-api-access-mcddl\") pod \"f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd\" (UID: \"f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd\") " Mar 07 15:19:30 crc kubenswrapper[4943]: I0307 15:19:30.386622 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd-scripts\") pod \"f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd\" (UID: \"f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd\") " Mar 07 15:19:30 crc kubenswrapper[4943]: I0307 15:19:30.386775 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd-etc-swift\") pod \"f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd\" (UID: \"f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd\") " Mar 07 15:19:30 crc kubenswrapper[4943]: I0307 15:19:30.386815 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd-ring-data-devices\") pod \"f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd\" (UID: \"f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd\") " Mar 07 15:19:30 crc kubenswrapper[4943]: I0307 15:19:30.386884 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd-swiftconf\") pod \"f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd\" (UID: \"f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd\") " Mar 07 15:19:30 crc kubenswrapper[4943]: I0307 15:19:30.386953 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd-dispersionconf\") pod \"f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd\" (UID: \"f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd\") " Mar 07 15:19:30 crc kubenswrapper[4943]: I0307 15:19:30.387552 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd" (UID: "f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:19:30 crc kubenswrapper[4943]: I0307 15:19:30.388259 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd" (UID: "f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:19:30 crc kubenswrapper[4943]: I0307 15:19:30.399172 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd-kube-api-access-mcddl" (OuterVolumeSpecName: "kube-api-access-mcddl") pod "f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd" (UID: "f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd"). InnerVolumeSpecName "kube-api-access-mcddl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:19:30 crc kubenswrapper[4943]: I0307 15:19:30.417516 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd" (UID: "f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:19:30 crc kubenswrapper[4943]: I0307 15:19:30.429908 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd-scripts" (OuterVolumeSpecName: "scripts") pod "f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd" (UID: "f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:19:30 crc kubenswrapper[4943]: I0307 15:19:30.436244 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd" (UID: "f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:19:30 crc kubenswrapper[4943]: I0307 15:19:30.489284 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:30 crc kubenswrapper[4943]: I0307 15:19:30.489347 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:30 crc kubenswrapper[4943]: I0307 15:19:30.489377 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mcddl\" (UniqueName: \"kubernetes.io/projected/f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd-kube-api-access-mcddl\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:30 crc kubenswrapper[4943]: I0307 15:19:30.489404 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:30 crc kubenswrapper[4943]: I0307 15:19:30.489430 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:30 crc kubenswrapper[4943]: I0307 15:19:30.489452 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:30 crc kubenswrapper[4943]: I0307 15:19:30.767079 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd" path="/var/lib/kubelet/pods/f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd/volumes" Mar 07 15:19:30 crc kubenswrapper[4943]: I0307 15:19:30.929503 4943 scope.go:117] "RemoveContainer" containerID="b8d0009e5cc84db560d54a6c0980a4cba727f318d70b9f8e0b4378e77218a550" Mar 07 15:19:30 crc kubenswrapper[4943]: I0307 15:19:30.929999 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-b6cwb" Mar 07 15:19:31 crc kubenswrapper[4943]: I0307 15:19:31.461507 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-sh5w2"] Mar 07 15:19:31 crc kubenswrapper[4943]: E0307 15:19:31.461952 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd" containerName="swift-ring-rebalance" Mar 07 15:19:31 crc kubenswrapper[4943]: I0307 15:19:31.461972 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd" containerName="swift-ring-rebalance" Mar 07 15:19:31 crc kubenswrapper[4943]: I0307 15:19:31.462297 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="f84c18ba-35b4-43c2-9ed8-4b931f9ab8fd" containerName="swift-ring-rebalance" Mar 07 15:19:31 crc kubenswrapper[4943]: I0307 15:19:31.463109 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-sh5w2" Mar 07 15:19:31 crc kubenswrapper[4943]: I0307 15:19:31.466008 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:19:31 crc kubenswrapper[4943]: I0307 15:19:31.470427 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:19:31 crc kubenswrapper[4943]: I0307 15:19:31.477115 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-sh5w2"] Mar 07 15:19:31 crc kubenswrapper[4943]: I0307 15:19:31.610457 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fe69851d-4ce1-468f-8e74-91b935103c43-scripts\") pod \"swift-ring-rebalance-debug-sh5w2\" (UID: \"fe69851d-4ce1-468f-8e74-91b935103c43\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sh5w2" Mar 07 15:19:31 crc kubenswrapper[4943]: I0307 15:19:31.610566 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/fe69851d-4ce1-468f-8e74-91b935103c43-etc-swift\") pod \"swift-ring-rebalance-debug-sh5w2\" (UID: \"fe69851d-4ce1-468f-8e74-91b935103c43\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sh5w2" Mar 07 15:19:31 crc kubenswrapper[4943]: I0307 15:19:31.610619 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/fe69851d-4ce1-468f-8e74-91b935103c43-swiftconf\") pod \"swift-ring-rebalance-debug-sh5w2\" (UID: \"fe69851d-4ce1-468f-8e74-91b935103c43\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sh5w2" Mar 07 15:19:31 crc kubenswrapper[4943]: I0307 15:19:31.610668 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/fe69851d-4ce1-468f-8e74-91b935103c43-ring-data-devices\") pod \"swift-ring-rebalance-debug-sh5w2\" (UID: \"fe69851d-4ce1-468f-8e74-91b935103c43\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sh5w2" Mar 07 15:19:31 crc kubenswrapper[4943]: I0307 15:19:31.610770 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/fe69851d-4ce1-468f-8e74-91b935103c43-dispersionconf\") pod \"swift-ring-rebalance-debug-sh5w2\" (UID: \"fe69851d-4ce1-468f-8e74-91b935103c43\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sh5w2" Mar 07 15:19:31 crc kubenswrapper[4943]: I0307 15:19:31.610842 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bhfn\" (UniqueName: \"kubernetes.io/projected/fe69851d-4ce1-468f-8e74-91b935103c43-kube-api-access-5bhfn\") pod \"swift-ring-rebalance-debug-sh5w2\" (UID: \"fe69851d-4ce1-468f-8e74-91b935103c43\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sh5w2" Mar 07 15:19:31 crc kubenswrapper[4943]: I0307 15:19:31.712344 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bhfn\" (UniqueName: \"kubernetes.io/projected/fe69851d-4ce1-468f-8e74-91b935103c43-kube-api-access-5bhfn\") pod \"swift-ring-rebalance-debug-sh5w2\" (UID: \"fe69851d-4ce1-468f-8e74-91b935103c43\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sh5w2" Mar 07 15:19:31 crc kubenswrapper[4943]: I0307 15:19:31.712500 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fe69851d-4ce1-468f-8e74-91b935103c43-scripts\") pod \"swift-ring-rebalance-debug-sh5w2\" (UID: \"fe69851d-4ce1-468f-8e74-91b935103c43\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sh5w2" Mar 07 15:19:31 crc kubenswrapper[4943]: I0307 15:19:31.712555 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/fe69851d-4ce1-468f-8e74-91b935103c43-etc-swift\") pod \"swift-ring-rebalance-debug-sh5w2\" (UID: \"fe69851d-4ce1-468f-8e74-91b935103c43\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sh5w2" Mar 07 15:19:31 crc kubenswrapper[4943]: I0307 15:19:31.712596 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/fe69851d-4ce1-468f-8e74-91b935103c43-swiftconf\") pod \"swift-ring-rebalance-debug-sh5w2\" (UID: \"fe69851d-4ce1-468f-8e74-91b935103c43\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sh5w2" Mar 07 15:19:31 crc kubenswrapper[4943]: I0307 15:19:31.712628 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/fe69851d-4ce1-468f-8e74-91b935103c43-ring-data-devices\") pod \"swift-ring-rebalance-debug-sh5w2\" (UID: \"fe69851d-4ce1-468f-8e74-91b935103c43\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sh5w2" Mar 07 15:19:31 crc kubenswrapper[4943]: I0307 15:19:31.712726 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/fe69851d-4ce1-468f-8e74-91b935103c43-dispersionconf\") pod \"swift-ring-rebalance-debug-sh5w2\" (UID: \"fe69851d-4ce1-468f-8e74-91b935103c43\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sh5w2" Mar 07 15:19:31 crc kubenswrapper[4943]: I0307 15:19:31.713372 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/fe69851d-4ce1-468f-8e74-91b935103c43-etc-swift\") pod \"swift-ring-rebalance-debug-sh5w2\" (UID: \"fe69851d-4ce1-468f-8e74-91b935103c43\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sh5w2" Mar 07 15:19:31 crc kubenswrapper[4943]: I0307 15:19:31.714364 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fe69851d-4ce1-468f-8e74-91b935103c43-scripts\") pod \"swift-ring-rebalance-debug-sh5w2\" (UID: \"fe69851d-4ce1-468f-8e74-91b935103c43\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sh5w2" Mar 07 15:19:31 crc kubenswrapper[4943]: I0307 15:19:31.715127 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/fe69851d-4ce1-468f-8e74-91b935103c43-ring-data-devices\") pod \"swift-ring-rebalance-debug-sh5w2\" (UID: \"fe69851d-4ce1-468f-8e74-91b935103c43\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sh5w2" Mar 07 15:19:31 crc kubenswrapper[4943]: I0307 15:19:31.718058 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/fe69851d-4ce1-468f-8e74-91b935103c43-dispersionconf\") pod \"swift-ring-rebalance-debug-sh5w2\" (UID: \"fe69851d-4ce1-468f-8e74-91b935103c43\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sh5w2" Mar 07 15:19:31 crc kubenswrapper[4943]: I0307 15:19:31.718547 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/fe69851d-4ce1-468f-8e74-91b935103c43-swiftconf\") pod \"swift-ring-rebalance-debug-sh5w2\" (UID: \"fe69851d-4ce1-468f-8e74-91b935103c43\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sh5w2" Mar 07 15:19:31 crc kubenswrapper[4943]: I0307 15:19:31.741612 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bhfn\" (UniqueName: \"kubernetes.io/projected/fe69851d-4ce1-468f-8e74-91b935103c43-kube-api-access-5bhfn\") pod \"swift-ring-rebalance-debug-sh5w2\" (UID: \"fe69851d-4ce1-468f-8e74-91b935103c43\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sh5w2" Mar 07 15:19:31 crc kubenswrapper[4943]: I0307 15:19:31.824286 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-sh5w2" Mar 07 15:19:32 crc kubenswrapper[4943]: I0307 15:19:32.310963 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-sh5w2"] Mar 07 15:19:32 crc kubenswrapper[4943]: W0307 15:19:32.325197 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfe69851d_4ce1_468f_8e74_91b935103c43.slice/crio-72f048510073d8b744b6a5e3939ebabee4d18dbc6c85fc8d89cb5512e1962a81 WatchSource:0}: Error finding container 72f048510073d8b744b6a5e3939ebabee4d18dbc6c85fc8d89cb5512e1962a81: Status 404 returned error can't find the container with id 72f048510073d8b744b6a5e3939ebabee4d18dbc6c85fc8d89cb5512e1962a81 Mar 07 15:19:32 crc kubenswrapper[4943]: I0307 15:19:32.975918 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-sh5w2" event={"ID":"fe69851d-4ce1-468f-8e74-91b935103c43","Type":"ContainerStarted","Data":"96ee9016de2326c8c0501131b1ba67d2ea036f5a37256c42afb745bfadb03d25"} Mar 07 15:19:32 crc kubenswrapper[4943]: I0307 15:19:32.976441 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-sh5w2" event={"ID":"fe69851d-4ce1-468f-8e74-91b935103c43","Type":"ContainerStarted","Data":"72f048510073d8b744b6a5e3939ebabee4d18dbc6c85fc8d89cb5512e1962a81"} Mar 07 15:19:33 crc kubenswrapper[4943]: I0307 15:19:33.009323 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-sh5w2" podStartSLOduration=2.00930347 podStartE2EDuration="2.00930347s" podCreationTimestamp="2026-03-07 15:19:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:19:33.00438053 +0000 UTC m=+2414.956517038" watchObservedRunningTime="2026-03-07 15:19:33.00930347 +0000 UTC m=+2414.961439968" Mar 07 15:19:33 crc kubenswrapper[4943]: I0307 15:19:33.990359 4943 generic.go:334] "Generic (PLEG): container finished" podID="fe69851d-4ce1-468f-8e74-91b935103c43" containerID="96ee9016de2326c8c0501131b1ba67d2ea036f5a37256c42afb745bfadb03d25" exitCode=0 Mar 07 15:19:33 crc kubenswrapper[4943]: I0307 15:19:33.990638 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-sh5w2" event={"ID":"fe69851d-4ce1-468f-8e74-91b935103c43","Type":"ContainerDied","Data":"96ee9016de2326c8c0501131b1ba67d2ea036f5a37256c42afb745bfadb03d25"} Mar 07 15:19:35 crc kubenswrapper[4943]: I0307 15:19:35.335365 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-sh5w2" Mar 07 15:19:35 crc kubenswrapper[4943]: I0307 15:19:35.386057 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-sh5w2"] Mar 07 15:19:35 crc kubenswrapper[4943]: I0307 15:19:35.395103 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-sh5w2"] Mar 07 15:19:35 crc kubenswrapper[4943]: I0307 15:19:35.478539 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/fe69851d-4ce1-468f-8e74-91b935103c43-etc-swift\") pod \"fe69851d-4ce1-468f-8e74-91b935103c43\" (UID: \"fe69851d-4ce1-468f-8e74-91b935103c43\") " Mar 07 15:19:35 crc kubenswrapper[4943]: I0307 15:19:35.478673 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5bhfn\" (UniqueName: \"kubernetes.io/projected/fe69851d-4ce1-468f-8e74-91b935103c43-kube-api-access-5bhfn\") pod \"fe69851d-4ce1-468f-8e74-91b935103c43\" (UID: \"fe69851d-4ce1-468f-8e74-91b935103c43\") " Mar 07 15:19:35 crc kubenswrapper[4943]: I0307 15:19:35.478842 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/fe69851d-4ce1-468f-8e74-91b935103c43-dispersionconf\") pod \"fe69851d-4ce1-468f-8e74-91b935103c43\" (UID: \"fe69851d-4ce1-468f-8e74-91b935103c43\") " Mar 07 15:19:35 crc kubenswrapper[4943]: I0307 15:19:35.478908 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fe69851d-4ce1-468f-8e74-91b935103c43-scripts\") pod \"fe69851d-4ce1-468f-8e74-91b935103c43\" (UID: \"fe69851d-4ce1-468f-8e74-91b935103c43\") " Mar 07 15:19:35 crc kubenswrapper[4943]: I0307 15:19:35.479031 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/fe69851d-4ce1-468f-8e74-91b935103c43-swiftconf\") pod \"fe69851d-4ce1-468f-8e74-91b935103c43\" (UID: \"fe69851d-4ce1-468f-8e74-91b935103c43\") " Mar 07 15:19:35 crc kubenswrapper[4943]: I0307 15:19:35.479113 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/fe69851d-4ce1-468f-8e74-91b935103c43-ring-data-devices\") pod \"fe69851d-4ce1-468f-8e74-91b935103c43\" (UID: \"fe69851d-4ce1-468f-8e74-91b935103c43\") " Mar 07 15:19:35 crc kubenswrapper[4943]: I0307 15:19:35.479801 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fe69851d-4ce1-468f-8e74-91b935103c43-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "fe69851d-4ce1-468f-8e74-91b935103c43" (UID: "fe69851d-4ce1-468f-8e74-91b935103c43"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:19:35 crc kubenswrapper[4943]: I0307 15:19:35.480258 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe69851d-4ce1-468f-8e74-91b935103c43-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "fe69851d-4ce1-468f-8e74-91b935103c43" (UID: "fe69851d-4ce1-468f-8e74-91b935103c43"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:19:35 crc kubenswrapper[4943]: I0307 15:19:35.485013 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe69851d-4ce1-468f-8e74-91b935103c43-kube-api-access-5bhfn" (OuterVolumeSpecName: "kube-api-access-5bhfn") pod "fe69851d-4ce1-468f-8e74-91b935103c43" (UID: "fe69851d-4ce1-468f-8e74-91b935103c43"). InnerVolumeSpecName "kube-api-access-5bhfn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:19:35 crc kubenswrapper[4943]: I0307 15:19:35.513763 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe69851d-4ce1-468f-8e74-91b935103c43-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "fe69851d-4ce1-468f-8e74-91b935103c43" (UID: "fe69851d-4ce1-468f-8e74-91b935103c43"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:19:35 crc kubenswrapper[4943]: I0307 15:19:35.514257 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fe69851d-4ce1-468f-8e74-91b935103c43-scripts" (OuterVolumeSpecName: "scripts") pod "fe69851d-4ce1-468f-8e74-91b935103c43" (UID: "fe69851d-4ce1-468f-8e74-91b935103c43"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:19:35 crc kubenswrapper[4943]: I0307 15:19:35.520961 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe69851d-4ce1-468f-8e74-91b935103c43-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "fe69851d-4ce1-468f-8e74-91b935103c43" (UID: "fe69851d-4ce1-468f-8e74-91b935103c43"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:19:35 crc kubenswrapper[4943]: I0307 15:19:35.581264 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/fe69851d-4ce1-468f-8e74-91b935103c43-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:35 crc kubenswrapper[4943]: I0307 15:19:35.581361 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/fe69851d-4ce1-468f-8e74-91b935103c43-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:35 crc kubenswrapper[4943]: I0307 15:19:35.581425 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5bhfn\" (UniqueName: \"kubernetes.io/projected/fe69851d-4ce1-468f-8e74-91b935103c43-kube-api-access-5bhfn\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:35 crc kubenswrapper[4943]: I0307 15:19:35.581456 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/fe69851d-4ce1-468f-8e74-91b935103c43-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:35 crc kubenswrapper[4943]: I0307 15:19:35.581476 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fe69851d-4ce1-468f-8e74-91b935103c43-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:35 crc kubenswrapper[4943]: I0307 15:19:35.581525 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/fe69851d-4ce1-468f-8e74-91b935103c43-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:36 crc kubenswrapper[4943]: I0307 15:19:36.009527 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="72f048510073d8b744b6a5e3939ebabee4d18dbc6c85fc8d89cb5512e1962a81" Mar 07 15:19:36 crc kubenswrapper[4943]: I0307 15:19:36.009624 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-sh5w2" Mar 07 15:19:36 crc kubenswrapper[4943]: I0307 15:19:36.531910 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-nr8zq"] Mar 07 15:19:36 crc kubenswrapper[4943]: E0307 15:19:36.532244 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe69851d-4ce1-468f-8e74-91b935103c43" containerName="swift-ring-rebalance" Mar 07 15:19:36 crc kubenswrapper[4943]: I0307 15:19:36.532256 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe69851d-4ce1-468f-8e74-91b935103c43" containerName="swift-ring-rebalance" Mar 07 15:19:36 crc kubenswrapper[4943]: I0307 15:19:36.532415 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe69851d-4ce1-468f-8e74-91b935103c43" containerName="swift-ring-rebalance" Mar 07 15:19:36 crc kubenswrapper[4943]: I0307 15:19:36.532939 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-nr8zq" Mar 07 15:19:36 crc kubenswrapper[4943]: I0307 15:19:36.535025 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:19:36 crc kubenswrapper[4943]: I0307 15:19:36.535077 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:19:36 crc kubenswrapper[4943]: I0307 15:19:36.538663 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-nr8zq"] Mar 07 15:19:36 crc kubenswrapper[4943]: I0307 15:19:36.713187 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c4262493-a091-4725-9a51-958224e1d697-scripts\") pod \"swift-ring-rebalance-debug-nr8zq\" (UID: \"c4262493-a091-4725-9a51-958224e1d697\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-nr8zq" Mar 07 15:19:36 crc kubenswrapper[4943]: I0307 15:19:36.713272 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c4262493-a091-4725-9a51-958224e1d697-swiftconf\") pod \"swift-ring-rebalance-debug-nr8zq\" (UID: \"c4262493-a091-4725-9a51-958224e1d697\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-nr8zq" Mar 07 15:19:36 crc kubenswrapper[4943]: I0307 15:19:36.713321 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c4262493-a091-4725-9a51-958224e1d697-etc-swift\") pod \"swift-ring-rebalance-debug-nr8zq\" (UID: \"c4262493-a091-4725-9a51-958224e1d697\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-nr8zq" Mar 07 15:19:36 crc kubenswrapper[4943]: I0307 15:19:36.713403 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c4262493-a091-4725-9a51-958224e1d697-ring-data-devices\") pod \"swift-ring-rebalance-debug-nr8zq\" (UID: \"c4262493-a091-4725-9a51-958224e1d697\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-nr8zq" Mar 07 15:19:36 crc kubenswrapper[4943]: I0307 15:19:36.713721 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rcfv\" (UniqueName: \"kubernetes.io/projected/c4262493-a091-4725-9a51-958224e1d697-kube-api-access-6rcfv\") pod \"swift-ring-rebalance-debug-nr8zq\" (UID: \"c4262493-a091-4725-9a51-958224e1d697\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-nr8zq" Mar 07 15:19:36 crc kubenswrapper[4943]: I0307 15:19:36.713821 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c4262493-a091-4725-9a51-958224e1d697-dispersionconf\") pod \"swift-ring-rebalance-debug-nr8zq\" (UID: \"c4262493-a091-4725-9a51-958224e1d697\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-nr8zq" Mar 07 15:19:36 crc kubenswrapper[4943]: I0307 15:19:36.765435 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe69851d-4ce1-468f-8e74-91b935103c43" path="/var/lib/kubelet/pods/fe69851d-4ce1-468f-8e74-91b935103c43/volumes" Mar 07 15:19:36 crc kubenswrapper[4943]: I0307 15:19:36.814275 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c4262493-a091-4725-9a51-958224e1d697-ring-data-devices\") pod \"swift-ring-rebalance-debug-nr8zq\" (UID: \"c4262493-a091-4725-9a51-958224e1d697\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-nr8zq" Mar 07 15:19:36 crc kubenswrapper[4943]: I0307 15:19:36.814345 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rcfv\" (UniqueName: \"kubernetes.io/projected/c4262493-a091-4725-9a51-958224e1d697-kube-api-access-6rcfv\") pod \"swift-ring-rebalance-debug-nr8zq\" (UID: \"c4262493-a091-4725-9a51-958224e1d697\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-nr8zq" Mar 07 15:19:36 crc kubenswrapper[4943]: I0307 15:19:36.814387 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c4262493-a091-4725-9a51-958224e1d697-dispersionconf\") pod \"swift-ring-rebalance-debug-nr8zq\" (UID: \"c4262493-a091-4725-9a51-958224e1d697\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-nr8zq" Mar 07 15:19:36 crc kubenswrapper[4943]: I0307 15:19:36.814416 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c4262493-a091-4725-9a51-958224e1d697-scripts\") pod \"swift-ring-rebalance-debug-nr8zq\" (UID: \"c4262493-a091-4725-9a51-958224e1d697\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-nr8zq" Mar 07 15:19:36 crc kubenswrapper[4943]: I0307 15:19:36.814452 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c4262493-a091-4725-9a51-958224e1d697-swiftconf\") pod \"swift-ring-rebalance-debug-nr8zq\" (UID: \"c4262493-a091-4725-9a51-958224e1d697\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-nr8zq" Mar 07 15:19:36 crc kubenswrapper[4943]: I0307 15:19:36.814484 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c4262493-a091-4725-9a51-958224e1d697-etc-swift\") pod \"swift-ring-rebalance-debug-nr8zq\" (UID: \"c4262493-a091-4725-9a51-958224e1d697\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-nr8zq" Mar 07 15:19:36 crc kubenswrapper[4943]: I0307 15:19:36.814935 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c4262493-a091-4725-9a51-958224e1d697-etc-swift\") pod \"swift-ring-rebalance-debug-nr8zq\" (UID: \"c4262493-a091-4725-9a51-958224e1d697\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-nr8zq" Mar 07 15:19:36 crc kubenswrapper[4943]: I0307 15:19:36.815523 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c4262493-a091-4725-9a51-958224e1d697-ring-data-devices\") pod \"swift-ring-rebalance-debug-nr8zq\" (UID: \"c4262493-a091-4725-9a51-958224e1d697\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-nr8zq" Mar 07 15:19:36 crc kubenswrapper[4943]: I0307 15:19:36.818957 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c4262493-a091-4725-9a51-958224e1d697-scripts\") pod \"swift-ring-rebalance-debug-nr8zq\" (UID: \"c4262493-a091-4725-9a51-958224e1d697\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-nr8zq" Mar 07 15:19:36 crc kubenswrapper[4943]: I0307 15:19:36.821006 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c4262493-a091-4725-9a51-958224e1d697-dispersionconf\") pod \"swift-ring-rebalance-debug-nr8zq\" (UID: \"c4262493-a091-4725-9a51-958224e1d697\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-nr8zq" Mar 07 15:19:36 crc kubenswrapper[4943]: I0307 15:19:36.824001 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c4262493-a091-4725-9a51-958224e1d697-swiftconf\") pod \"swift-ring-rebalance-debug-nr8zq\" (UID: \"c4262493-a091-4725-9a51-958224e1d697\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-nr8zq" Mar 07 15:19:36 crc kubenswrapper[4943]: I0307 15:19:36.832394 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rcfv\" (UniqueName: \"kubernetes.io/projected/c4262493-a091-4725-9a51-958224e1d697-kube-api-access-6rcfv\") pod \"swift-ring-rebalance-debug-nr8zq\" (UID: \"c4262493-a091-4725-9a51-958224e1d697\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-nr8zq" Mar 07 15:19:36 crc kubenswrapper[4943]: I0307 15:19:36.851215 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-nr8zq" Mar 07 15:19:37 crc kubenswrapper[4943]: I0307 15:19:37.265392 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-nr8zq"] Mar 07 15:19:37 crc kubenswrapper[4943]: W0307 15:19:37.281152 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc4262493_a091_4725_9a51_958224e1d697.slice/crio-64eb7d9a7ddcac20f36829d59845b1eef55d8eda91564a86152a0a751bc2dd8a WatchSource:0}: Error finding container 64eb7d9a7ddcac20f36829d59845b1eef55d8eda91564a86152a0a751bc2dd8a: Status 404 returned error can't find the container with id 64eb7d9a7ddcac20f36829d59845b1eef55d8eda91564a86152a0a751bc2dd8a Mar 07 15:19:38 crc kubenswrapper[4943]: I0307 15:19:38.036128 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-nr8zq" event={"ID":"c4262493-a091-4725-9a51-958224e1d697","Type":"ContainerStarted","Data":"9ac9868825e1815aca9d9330a4c0faf79884d5b3741d270439dccaa556536326"} Mar 07 15:19:38 crc kubenswrapper[4943]: I0307 15:19:38.036505 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-nr8zq" event={"ID":"c4262493-a091-4725-9a51-958224e1d697","Type":"ContainerStarted","Data":"64eb7d9a7ddcac20f36829d59845b1eef55d8eda91564a86152a0a751bc2dd8a"} Mar 07 15:19:38 crc kubenswrapper[4943]: I0307 15:19:38.059117 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-nr8zq" podStartSLOduration=2.059098888 podStartE2EDuration="2.059098888s" podCreationTimestamp="2026-03-07 15:19:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:19:38.055816087 +0000 UTC m=+2420.007952635" watchObservedRunningTime="2026-03-07 15:19:38.059098888 +0000 UTC m=+2420.011235386" Mar 07 15:19:38 crc kubenswrapper[4943]: I0307 15:19:38.765353 4943 scope.go:117] "RemoveContainer" containerID="448308ad6e34ac922d66b3c90e66d0927407a94dd28633ec350502d69e4ca8aa" Mar 07 15:19:38 crc kubenswrapper[4943]: E0307 15:19:38.765732 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cpdmm_openshift-machine-config-operator(4c1f18ea-e557-41fc-95dd-ce4e1722a61e)\"" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" Mar 07 15:19:39 crc kubenswrapper[4943]: I0307 15:19:39.053620 4943 generic.go:334] "Generic (PLEG): container finished" podID="c4262493-a091-4725-9a51-958224e1d697" containerID="9ac9868825e1815aca9d9330a4c0faf79884d5b3741d270439dccaa556536326" exitCode=0 Mar 07 15:19:39 crc kubenswrapper[4943]: I0307 15:19:39.053687 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-nr8zq" event={"ID":"c4262493-a091-4725-9a51-958224e1d697","Type":"ContainerDied","Data":"9ac9868825e1815aca9d9330a4c0faf79884d5b3741d270439dccaa556536326"} Mar 07 15:19:40 crc kubenswrapper[4943]: I0307 15:19:40.342250 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-nr8zq" Mar 07 15:19:40 crc kubenswrapper[4943]: I0307 15:19:40.403470 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-nr8zq"] Mar 07 15:19:40 crc kubenswrapper[4943]: I0307 15:19:40.409786 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-nr8zq"] Mar 07 15:19:40 crc kubenswrapper[4943]: I0307 15:19:40.473187 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c4262493-a091-4725-9a51-958224e1d697-swiftconf\") pod \"c4262493-a091-4725-9a51-958224e1d697\" (UID: \"c4262493-a091-4725-9a51-958224e1d697\") " Mar 07 15:19:40 crc kubenswrapper[4943]: I0307 15:19:40.473260 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c4262493-a091-4725-9a51-958224e1d697-ring-data-devices\") pod \"c4262493-a091-4725-9a51-958224e1d697\" (UID: \"c4262493-a091-4725-9a51-958224e1d697\") " Mar 07 15:19:40 crc kubenswrapper[4943]: I0307 15:19:40.473319 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c4262493-a091-4725-9a51-958224e1d697-scripts\") pod \"c4262493-a091-4725-9a51-958224e1d697\" (UID: \"c4262493-a091-4725-9a51-958224e1d697\") " Mar 07 15:19:40 crc kubenswrapper[4943]: I0307 15:19:40.473348 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c4262493-a091-4725-9a51-958224e1d697-dispersionconf\") pod \"c4262493-a091-4725-9a51-958224e1d697\" (UID: \"c4262493-a091-4725-9a51-958224e1d697\") " Mar 07 15:19:40 crc kubenswrapper[4943]: I0307 15:19:40.473381 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6rcfv\" (UniqueName: \"kubernetes.io/projected/c4262493-a091-4725-9a51-958224e1d697-kube-api-access-6rcfv\") pod \"c4262493-a091-4725-9a51-958224e1d697\" (UID: \"c4262493-a091-4725-9a51-958224e1d697\") " Mar 07 15:19:40 crc kubenswrapper[4943]: I0307 15:19:40.473470 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c4262493-a091-4725-9a51-958224e1d697-etc-swift\") pod \"c4262493-a091-4725-9a51-958224e1d697\" (UID: \"c4262493-a091-4725-9a51-958224e1d697\") " Mar 07 15:19:40 crc kubenswrapper[4943]: I0307 15:19:40.473984 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4262493-a091-4725-9a51-958224e1d697-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "c4262493-a091-4725-9a51-958224e1d697" (UID: "c4262493-a091-4725-9a51-958224e1d697"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:19:40 crc kubenswrapper[4943]: I0307 15:19:40.474551 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4262493-a091-4725-9a51-958224e1d697-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "c4262493-a091-4725-9a51-958224e1d697" (UID: "c4262493-a091-4725-9a51-958224e1d697"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:19:40 crc kubenswrapper[4943]: I0307 15:19:40.479097 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4262493-a091-4725-9a51-958224e1d697-kube-api-access-6rcfv" (OuterVolumeSpecName: "kube-api-access-6rcfv") pod "c4262493-a091-4725-9a51-958224e1d697" (UID: "c4262493-a091-4725-9a51-958224e1d697"). InnerVolumeSpecName "kube-api-access-6rcfv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:19:40 crc kubenswrapper[4943]: I0307 15:19:40.493500 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4262493-a091-4725-9a51-958224e1d697-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "c4262493-a091-4725-9a51-958224e1d697" (UID: "c4262493-a091-4725-9a51-958224e1d697"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:19:40 crc kubenswrapper[4943]: I0307 15:19:40.497855 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4262493-a091-4725-9a51-958224e1d697-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "c4262493-a091-4725-9a51-958224e1d697" (UID: "c4262493-a091-4725-9a51-958224e1d697"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:19:40 crc kubenswrapper[4943]: I0307 15:19:40.506711 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4262493-a091-4725-9a51-958224e1d697-scripts" (OuterVolumeSpecName: "scripts") pod "c4262493-a091-4725-9a51-958224e1d697" (UID: "c4262493-a091-4725-9a51-958224e1d697"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:19:40 crc kubenswrapper[4943]: I0307 15:19:40.575477 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c4262493-a091-4725-9a51-958224e1d697-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:40 crc kubenswrapper[4943]: I0307 15:19:40.575526 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c4262493-a091-4725-9a51-958224e1d697-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:40 crc kubenswrapper[4943]: I0307 15:19:40.575538 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c4262493-a091-4725-9a51-958224e1d697-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:40 crc kubenswrapper[4943]: I0307 15:19:40.575554 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c4262493-a091-4725-9a51-958224e1d697-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:40 crc kubenswrapper[4943]: I0307 15:19:40.575565 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c4262493-a091-4725-9a51-958224e1d697-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:40 crc kubenswrapper[4943]: I0307 15:19:40.575578 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6rcfv\" (UniqueName: \"kubernetes.io/projected/c4262493-a091-4725-9a51-958224e1d697-kube-api-access-6rcfv\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:40 crc kubenswrapper[4943]: I0307 15:19:40.768066 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4262493-a091-4725-9a51-958224e1d697" path="/var/lib/kubelet/pods/c4262493-a091-4725-9a51-958224e1d697/volumes" Mar 07 15:19:41 crc kubenswrapper[4943]: I0307 15:19:41.077120 4943 scope.go:117] "RemoveContainer" containerID="9ac9868825e1815aca9d9330a4c0faf79884d5b3741d270439dccaa556536326" Mar 07 15:19:41 crc kubenswrapper[4943]: I0307 15:19:41.077166 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-nr8zq" Mar 07 15:19:41 crc kubenswrapper[4943]: I0307 15:19:41.573840 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-krtxp"] Mar 07 15:19:41 crc kubenswrapper[4943]: E0307 15:19:41.574358 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4262493-a091-4725-9a51-958224e1d697" containerName="swift-ring-rebalance" Mar 07 15:19:41 crc kubenswrapper[4943]: I0307 15:19:41.574398 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4262493-a091-4725-9a51-958224e1d697" containerName="swift-ring-rebalance" Mar 07 15:19:41 crc kubenswrapper[4943]: I0307 15:19:41.574712 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4262493-a091-4725-9a51-958224e1d697" containerName="swift-ring-rebalance" Mar 07 15:19:41 crc kubenswrapper[4943]: I0307 15:19:41.575902 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-krtxp" Mar 07 15:19:41 crc kubenswrapper[4943]: I0307 15:19:41.580896 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:19:41 crc kubenswrapper[4943]: I0307 15:19:41.584819 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:19:41 crc kubenswrapper[4943]: I0307 15:19:41.593145 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-krtxp"] Mar 07 15:19:41 crc kubenswrapper[4943]: I0307 15:19:41.690582 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/74504775-550e-4eb7-986f-b9affc029d07-scripts\") pod \"swift-ring-rebalance-debug-krtxp\" (UID: \"74504775-550e-4eb7-986f-b9affc029d07\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-krtxp" Mar 07 15:19:41 crc kubenswrapper[4943]: I0307 15:19:41.690694 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/74504775-550e-4eb7-986f-b9affc029d07-swiftconf\") pod \"swift-ring-rebalance-debug-krtxp\" (UID: \"74504775-550e-4eb7-986f-b9affc029d07\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-krtxp" Mar 07 15:19:41 crc kubenswrapper[4943]: I0307 15:19:41.690761 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgz94\" (UniqueName: \"kubernetes.io/projected/74504775-550e-4eb7-986f-b9affc029d07-kube-api-access-jgz94\") pod \"swift-ring-rebalance-debug-krtxp\" (UID: \"74504775-550e-4eb7-986f-b9affc029d07\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-krtxp" Mar 07 15:19:41 crc kubenswrapper[4943]: I0307 15:19:41.690795 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/74504775-550e-4eb7-986f-b9affc029d07-ring-data-devices\") pod \"swift-ring-rebalance-debug-krtxp\" (UID: \"74504775-550e-4eb7-986f-b9affc029d07\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-krtxp" Mar 07 15:19:41 crc kubenswrapper[4943]: I0307 15:19:41.691870 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/74504775-550e-4eb7-986f-b9affc029d07-etc-swift\") pod \"swift-ring-rebalance-debug-krtxp\" (UID: \"74504775-550e-4eb7-986f-b9affc029d07\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-krtxp" Mar 07 15:19:41 crc kubenswrapper[4943]: I0307 15:19:41.692004 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/74504775-550e-4eb7-986f-b9affc029d07-dispersionconf\") pod \"swift-ring-rebalance-debug-krtxp\" (UID: \"74504775-550e-4eb7-986f-b9affc029d07\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-krtxp" Mar 07 15:19:41 crc kubenswrapper[4943]: I0307 15:19:41.794227 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/74504775-550e-4eb7-986f-b9affc029d07-swiftconf\") pod \"swift-ring-rebalance-debug-krtxp\" (UID: \"74504775-550e-4eb7-986f-b9affc029d07\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-krtxp" Mar 07 15:19:41 crc kubenswrapper[4943]: I0307 15:19:41.794342 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgz94\" (UniqueName: \"kubernetes.io/projected/74504775-550e-4eb7-986f-b9affc029d07-kube-api-access-jgz94\") pod \"swift-ring-rebalance-debug-krtxp\" (UID: \"74504775-550e-4eb7-986f-b9affc029d07\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-krtxp" Mar 07 15:19:41 crc kubenswrapper[4943]: I0307 15:19:41.794397 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/74504775-550e-4eb7-986f-b9affc029d07-ring-data-devices\") pod \"swift-ring-rebalance-debug-krtxp\" (UID: \"74504775-550e-4eb7-986f-b9affc029d07\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-krtxp" Mar 07 15:19:41 crc kubenswrapper[4943]: I0307 15:19:41.794490 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/74504775-550e-4eb7-986f-b9affc029d07-etc-swift\") pod \"swift-ring-rebalance-debug-krtxp\" (UID: \"74504775-550e-4eb7-986f-b9affc029d07\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-krtxp" Mar 07 15:19:41 crc kubenswrapper[4943]: I0307 15:19:41.794552 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/74504775-550e-4eb7-986f-b9affc029d07-dispersionconf\") pod \"swift-ring-rebalance-debug-krtxp\" (UID: \"74504775-550e-4eb7-986f-b9affc029d07\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-krtxp" Mar 07 15:19:41 crc kubenswrapper[4943]: I0307 15:19:41.794636 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/74504775-550e-4eb7-986f-b9affc029d07-scripts\") pod \"swift-ring-rebalance-debug-krtxp\" (UID: \"74504775-550e-4eb7-986f-b9affc029d07\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-krtxp" Mar 07 15:19:41 crc kubenswrapper[4943]: I0307 15:19:41.795595 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/74504775-550e-4eb7-986f-b9affc029d07-etc-swift\") pod \"swift-ring-rebalance-debug-krtxp\" (UID: \"74504775-550e-4eb7-986f-b9affc029d07\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-krtxp" Mar 07 15:19:41 crc kubenswrapper[4943]: I0307 15:19:41.795990 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/74504775-550e-4eb7-986f-b9affc029d07-ring-data-devices\") pod \"swift-ring-rebalance-debug-krtxp\" (UID: \"74504775-550e-4eb7-986f-b9affc029d07\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-krtxp" Mar 07 15:19:41 crc kubenswrapper[4943]: I0307 15:19:41.796109 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/74504775-550e-4eb7-986f-b9affc029d07-scripts\") pod \"swift-ring-rebalance-debug-krtxp\" (UID: \"74504775-550e-4eb7-986f-b9affc029d07\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-krtxp" Mar 07 15:19:41 crc kubenswrapper[4943]: I0307 15:19:41.801042 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/74504775-550e-4eb7-986f-b9affc029d07-dispersionconf\") pod \"swift-ring-rebalance-debug-krtxp\" (UID: \"74504775-550e-4eb7-986f-b9affc029d07\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-krtxp" Mar 07 15:19:41 crc kubenswrapper[4943]: I0307 15:19:41.801586 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/74504775-550e-4eb7-986f-b9affc029d07-swiftconf\") pod \"swift-ring-rebalance-debug-krtxp\" (UID: \"74504775-550e-4eb7-986f-b9affc029d07\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-krtxp" Mar 07 15:19:41 crc kubenswrapper[4943]: I0307 15:19:41.820921 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgz94\" (UniqueName: \"kubernetes.io/projected/74504775-550e-4eb7-986f-b9affc029d07-kube-api-access-jgz94\") pod \"swift-ring-rebalance-debug-krtxp\" (UID: \"74504775-550e-4eb7-986f-b9affc029d07\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-krtxp" Mar 07 15:19:41 crc kubenswrapper[4943]: I0307 15:19:41.940859 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-krtxp" Mar 07 15:19:42 crc kubenswrapper[4943]: I0307 15:19:42.453256 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-krtxp"] Mar 07 15:19:43 crc kubenswrapper[4943]: I0307 15:19:43.109683 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-krtxp" event={"ID":"74504775-550e-4eb7-986f-b9affc029d07","Type":"ContainerStarted","Data":"27a0e2a50f061bddad7d687ea67d3df7313113490b76d8d044db002cc063ca08"} Mar 07 15:19:43 crc kubenswrapper[4943]: I0307 15:19:43.110091 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-krtxp" event={"ID":"74504775-550e-4eb7-986f-b9affc029d07","Type":"ContainerStarted","Data":"afa4feae54b7d4afd2d6203786cfb8ffd1b39cdf55b348eb0baa2703506d30c0"} Mar 07 15:19:43 crc kubenswrapper[4943]: I0307 15:19:43.143988 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-krtxp" podStartSLOduration=2.143923274 podStartE2EDuration="2.143923274s" podCreationTimestamp="2026-03-07 15:19:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:19:43.133072418 +0000 UTC m=+2425.085208956" watchObservedRunningTime="2026-03-07 15:19:43.143923274 +0000 UTC m=+2425.096059812" Mar 07 15:19:44 crc kubenswrapper[4943]: I0307 15:19:44.125162 4943 generic.go:334] "Generic (PLEG): container finished" podID="74504775-550e-4eb7-986f-b9affc029d07" containerID="27a0e2a50f061bddad7d687ea67d3df7313113490b76d8d044db002cc063ca08" exitCode=0 Mar 07 15:19:44 crc kubenswrapper[4943]: I0307 15:19:44.125292 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-krtxp" event={"ID":"74504775-550e-4eb7-986f-b9affc029d07","Type":"ContainerDied","Data":"27a0e2a50f061bddad7d687ea67d3df7313113490b76d8d044db002cc063ca08"} Mar 07 15:19:45 crc kubenswrapper[4943]: I0307 15:19:45.527218 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-krtxp" Mar 07 15:19:45 crc kubenswrapper[4943]: I0307 15:19:45.574406 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-krtxp"] Mar 07 15:19:45 crc kubenswrapper[4943]: I0307 15:19:45.583509 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-krtxp"] Mar 07 15:19:45 crc kubenswrapper[4943]: I0307 15:19:45.661812 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/74504775-550e-4eb7-986f-b9affc029d07-etc-swift\") pod \"74504775-550e-4eb7-986f-b9affc029d07\" (UID: \"74504775-550e-4eb7-986f-b9affc029d07\") " Mar 07 15:19:45 crc kubenswrapper[4943]: I0307 15:19:45.661910 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/74504775-550e-4eb7-986f-b9affc029d07-ring-data-devices\") pod \"74504775-550e-4eb7-986f-b9affc029d07\" (UID: \"74504775-550e-4eb7-986f-b9affc029d07\") " Mar 07 15:19:45 crc kubenswrapper[4943]: I0307 15:19:45.662065 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/74504775-550e-4eb7-986f-b9affc029d07-scripts\") pod \"74504775-550e-4eb7-986f-b9affc029d07\" (UID: \"74504775-550e-4eb7-986f-b9affc029d07\") " Mar 07 15:19:45 crc kubenswrapper[4943]: I0307 15:19:45.662100 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/74504775-550e-4eb7-986f-b9affc029d07-dispersionconf\") pod \"74504775-550e-4eb7-986f-b9affc029d07\" (UID: \"74504775-550e-4eb7-986f-b9affc029d07\") " Mar 07 15:19:45 crc kubenswrapper[4943]: I0307 15:19:45.662162 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jgz94\" (UniqueName: \"kubernetes.io/projected/74504775-550e-4eb7-986f-b9affc029d07-kube-api-access-jgz94\") pod \"74504775-550e-4eb7-986f-b9affc029d07\" (UID: \"74504775-550e-4eb7-986f-b9affc029d07\") " Mar 07 15:19:45 crc kubenswrapper[4943]: I0307 15:19:45.662211 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/74504775-550e-4eb7-986f-b9affc029d07-swiftconf\") pod \"74504775-550e-4eb7-986f-b9affc029d07\" (UID: \"74504775-550e-4eb7-986f-b9affc029d07\") " Mar 07 15:19:45 crc kubenswrapper[4943]: I0307 15:19:45.664121 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74504775-550e-4eb7-986f-b9affc029d07-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "74504775-550e-4eb7-986f-b9affc029d07" (UID: "74504775-550e-4eb7-986f-b9affc029d07"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:19:45 crc kubenswrapper[4943]: I0307 15:19:45.664577 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/74504775-550e-4eb7-986f-b9affc029d07-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:45 crc kubenswrapper[4943]: I0307 15:19:45.665421 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74504775-550e-4eb7-986f-b9affc029d07-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "74504775-550e-4eb7-986f-b9affc029d07" (UID: "74504775-550e-4eb7-986f-b9affc029d07"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:19:45 crc kubenswrapper[4943]: I0307 15:19:45.670366 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74504775-550e-4eb7-986f-b9affc029d07-kube-api-access-jgz94" (OuterVolumeSpecName: "kube-api-access-jgz94") pod "74504775-550e-4eb7-986f-b9affc029d07" (UID: "74504775-550e-4eb7-986f-b9affc029d07"). InnerVolumeSpecName "kube-api-access-jgz94". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:19:45 crc kubenswrapper[4943]: I0307 15:19:45.691847 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74504775-550e-4eb7-986f-b9affc029d07-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "74504775-550e-4eb7-986f-b9affc029d07" (UID: "74504775-550e-4eb7-986f-b9affc029d07"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:19:45 crc kubenswrapper[4943]: I0307 15:19:45.702093 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74504775-550e-4eb7-986f-b9affc029d07-scripts" (OuterVolumeSpecName: "scripts") pod "74504775-550e-4eb7-986f-b9affc029d07" (UID: "74504775-550e-4eb7-986f-b9affc029d07"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:19:45 crc kubenswrapper[4943]: I0307 15:19:45.702754 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74504775-550e-4eb7-986f-b9affc029d07-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "74504775-550e-4eb7-986f-b9affc029d07" (UID: "74504775-550e-4eb7-986f-b9affc029d07"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:19:45 crc kubenswrapper[4943]: I0307 15:19:45.766173 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/74504775-550e-4eb7-986f-b9affc029d07-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:45 crc kubenswrapper[4943]: I0307 15:19:45.766450 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/74504775-550e-4eb7-986f-b9affc029d07-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:45 crc kubenswrapper[4943]: I0307 15:19:45.766579 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jgz94\" (UniqueName: \"kubernetes.io/projected/74504775-550e-4eb7-986f-b9affc029d07-kube-api-access-jgz94\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:45 crc kubenswrapper[4943]: I0307 15:19:45.766695 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/74504775-550e-4eb7-986f-b9affc029d07-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:45 crc kubenswrapper[4943]: I0307 15:19:45.766808 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/74504775-550e-4eb7-986f-b9affc029d07-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:46 crc kubenswrapper[4943]: I0307 15:19:46.154863 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="afa4feae54b7d4afd2d6203786cfb8ffd1b39cdf55b348eb0baa2703506d30c0" Mar 07 15:19:46 crc kubenswrapper[4943]: I0307 15:19:46.155319 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-krtxp" Mar 07 15:19:46 crc kubenswrapper[4943]: I0307 15:19:46.752160 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-rmbq7"] Mar 07 15:19:46 crc kubenswrapper[4943]: E0307 15:19:46.752543 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74504775-550e-4eb7-986f-b9affc029d07" containerName="swift-ring-rebalance" Mar 07 15:19:46 crc kubenswrapper[4943]: I0307 15:19:46.752559 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="74504775-550e-4eb7-986f-b9affc029d07" containerName="swift-ring-rebalance" Mar 07 15:19:46 crc kubenswrapper[4943]: I0307 15:19:46.752757 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="74504775-550e-4eb7-986f-b9affc029d07" containerName="swift-ring-rebalance" Mar 07 15:19:46 crc kubenswrapper[4943]: I0307 15:19:46.753376 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-rmbq7" Mar 07 15:19:46 crc kubenswrapper[4943]: I0307 15:19:46.756729 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:19:46 crc kubenswrapper[4943]: I0307 15:19:46.757442 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:19:46 crc kubenswrapper[4943]: I0307 15:19:46.775036 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74504775-550e-4eb7-986f-b9affc029d07" path="/var/lib/kubelet/pods/74504775-550e-4eb7-986f-b9affc029d07/volumes" Mar 07 15:19:46 crc kubenswrapper[4943]: I0307 15:19:46.775608 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-rmbq7"] Mar 07 15:19:46 crc kubenswrapper[4943]: I0307 15:19:46.802516 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6b0b8ea3-e635-4dcf-8b35-04abc468d724-ring-data-devices\") pod \"swift-ring-rebalance-debug-rmbq7\" (UID: \"6b0b8ea3-e635-4dcf-8b35-04abc468d724\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rmbq7" Mar 07 15:19:46 crc kubenswrapper[4943]: I0307 15:19:46.802667 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6b0b8ea3-e635-4dcf-8b35-04abc468d724-scripts\") pod \"swift-ring-rebalance-debug-rmbq7\" (UID: \"6b0b8ea3-e635-4dcf-8b35-04abc468d724\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rmbq7" Mar 07 15:19:46 crc kubenswrapper[4943]: I0307 15:19:46.802709 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6b0b8ea3-e635-4dcf-8b35-04abc468d724-etc-swift\") pod \"swift-ring-rebalance-debug-rmbq7\" (UID: \"6b0b8ea3-e635-4dcf-8b35-04abc468d724\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rmbq7" Mar 07 15:19:46 crc kubenswrapper[4943]: I0307 15:19:46.802979 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44pq5\" (UniqueName: \"kubernetes.io/projected/6b0b8ea3-e635-4dcf-8b35-04abc468d724-kube-api-access-44pq5\") pod \"swift-ring-rebalance-debug-rmbq7\" (UID: \"6b0b8ea3-e635-4dcf-8b35-04abc468d724\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rmbq7" Mar 07 15:19:46 crc kubenswrapper[4943]: I0307 15:19:46.803047 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6b0b8ea3-e635-4dcf-8b35-04abc468d724-swiftconf\") pod \"swift-ring-rebalance-debug-rmbq7\" (UID: \"6b0b8ea3-e635-4dcf-8b35-04abc468d724\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rmbq7" Mar 07 15:19:46 crc kubenswrapper[4943]: I0307 15:19:46.803094 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6b0b8ea3-e635-4dcf-8b35-04abc468d724-dispersionconf\") pod \"swift-ring-rebalance-debug-rmbq7\" (UID: \"6b0b8ea3-e635-4dcf-8b35-04abc468d724\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rmbq7" Mar 07 15:19:46 crc kubenswrapper[4943]: I0307 15:19:46.903725 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44pq5\" (UniqueName: \"kubernetes.io/projected/6b0b8ea3-e635-4dcf-8b35-04abc468d724-kube-api-access-44pq5\") pod \"swift-ring-rebalance-debug-rmbq7\" (UID: \"6b0b8ea3-e635-4dcf-8b35-04abc468d724\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rmbq7" Mar 07 15:19:46 crc kubenswrapper[4943]: I0307 15:19:46.903803 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6b0b8ea3-e635-4dcf-8b35-04abc468d724-swiftconf\") pod \"swift-ring-rebalance-debug-rmbq7\" (UID: \"6b0b8ea3-e635-4dcf-8b35-04abc468d724\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rmbq7" Mar 07 15:19:46 crc kubenswrapper[4943]: I0307 15:19:46.903843 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6b0b8ea3-e635-4dcf-8b35-04abc468d724-dispersionconf\") pod \"swift-ring-rebalance-debug-rmbq7\" (UID: \"6b0b8ea3-e635-4dcf-8b35-04abc468d724\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rmbq7" Mar 07 15:19:46 crc kubenswrapper[4943]: I0307 15:19:46.903920 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6b0b8ea3-e635-4dcf-8b35-04abc468d724-ring-data-devices\") pod \"swift-ring-rebalance-debug-rmbq7\" (UID: \"6b0b8ea3-e635-4dcf-8b35-04abc468d724\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rmbq7" Mar 07 15:19:46 crc kubenswrapper[4943]: I0307 15:19:46.904005 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6b0b8ea3-e635-4dcf-8b35-04abc468d724-scripts\") pod \"swift-ring-rebalance-debug-rmbq7\" (UID: \"6b0b8ea3-e635-4dcf-8b35-04abc468d724\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rmbq7" Mar 07 15:19:46 crc kubenswrapper[4943]: I0307 15:19:46.904043 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6b0b8ea3-e635-4dcf-8b35-04abc468d724-etc-swift\") pod \"swift-ring-rebalance-debug-rmbq7\" (UID: \"6b0b8ea3-e635-4dcf-8b35-04abc468d724\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rmbq7" Mar 07 15:19:46 crc kubenswrapper[4943]: I0307 15:19:46.905767 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6b0b8ea3-e635-4dcf-8b35-04abc468d724-etc-swift\") pod \"swift-ring-rebalance-debug-rmbq7\" (UID: \"6b0b8ea3-e635-4dcf-8b35-04abc468d724\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rmbq7" Mar 07 15:19:46 crc kubenswrapper[4943]: I0307 15:19:46.906125 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6b0b8ea3-e635-4dcf-8b35-04abc468d724-ring-data-devices\") pod \"swift-ring-rebalance-debug-rmbq7\" (UID: \"6b0b8ea3-e635-4dcf-8b35-04abc468d724\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rmbq7" Mar 07 15:19:46 crc kubenswrapper[4943]: I0307 15:19:46.906313 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6b0b8ea3-e635-4dcf-8b35-04abc468d724-scripts\") pod \"swift-ring-rebalance-debug-rmbq7\" (UID: \"6b0b8ea3-e635-4dcf-8b35-04abc468d724\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rmbq7" Mar 07 15:19:46 crc kubenswrapper[4943]: I0307 15:19:46.908102 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6b0b8ea3-e635-4dcf-8b35-04abc468d724-dispersionconf\") pod \"swift-ring-rebalance-debug-rmbq7\" (UID: \"6b0b8ea3-e635-4dcf-8b35-04abc468d724\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rmbq7" Mar 07 15:19:46 crc kubenswrapper[4943]: I0307 15:19:46.909454 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6b0b8ea3-e635-4dcf-8b35-04abc468d724-swiftconf\") pod \"swift-ring-rebalance-debug-rmbq7\" (UID: \"6b0b8ea3-e635-4dcf-8b35-04abc468d724\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rmbq7" Mar 07 15:19:46 crc kubenswrapper[4943]: I0307 15:19:46.925883 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44pq5\" (UniqueName: \"kubernetes.io/projected/6b0b8ea3-e635-4dcf-8b35-04abc468d724-kube-api-access-44pq5\") pod \"swift-ring-rebalance-debug-rmbq7\" (UID: \"6b0b8ea3-e635-4dcf-8b35-04abc468d724\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rmbq7" Mar 07 15:19:47 crc kubenswrapper[4943]: I0307 15:19:47.077988 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-rmbq7" Mar 07 15:19:47 crc kubenswrapper[4943]: I0307 15:19:47.358166 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-rmbq7"] Mar 07 15:19:48 crc kubenswrapper[4943]: I0307 15:19:48.208992 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-rmbq7" event={"ID":"6b0b8ea3-e635-4dcf-8b35-04abc468d724","Type":"ContainerStarted","Data":"98c57785988edac00b070e088586579b62685a8b1d3d70d1725cb6cc2cd845bb"} Mar 07 15:19:48 crc kubenswrapper[4943]: I0307 15:19:48.212200 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-rmbq7" event={"ID":"6b0b8ea3-e635-4dcf-8b35-04abc468d724","Type":"ContainerStarted","Data":"5f62f24a27acc856fd439b5b3eb3d390880937279f54043d14c52a87389229b5"} Mar 07 15:19:48 crc kubenswrapper[4943]: I0307 15:19:48.246215 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-rmbq7" podStartSLOduration=2.246188357 podStartE2EDuration="2.246188357s" podCreationTimestamp="2026-03-07 15:19:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:19:48.231303042 +0000 UTC m=+2430.183439590" watchObservedRunningTime="2026-03-07 15:19:48.246188357 +0000 UTC m=+2430.198324885" Mar 07 15:19:49 crc kubenswrapper[4943]: I0307 15:19:49.225654 4943 generic.go:334] "Generic (PLEG): container finished" podID="6b0b8ea3-e635-4dcf-8b35-04abc468d724" containerID="98c57785988edac00b070e088586579b62685a8b1d3d70d1725cb6cc2cd845bb" exitCode=0 Mar 07 15:19:49 crc kubenswrapper[4943]: I0307 15:19:49.226032 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-rmbq7" event={"ID":"6b0b8ea3-e635-4dcf-8b35-04abc468d724","Type":"ContainerDied","Data":"98c57785988edac00b070e088586579b62685a8b1d3d70d1725cb6cc2cd845bb"} Mar 07 15:19:50 crc kubenswrapper[4943]: I0307 15:19:50.614784 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-rmbq7" Mar 07 15:19:50 crc kubenswrapper[4943]: I0307 15:19:50.663785 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-rmbq7"] Mar 07 15:19:50 crc kubenswrapper[4943]: I0307 15:19:50.668815 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-rmbq7"] Mar 07 15:19:50 crc kubenswrapper[4943]: I0307 15:19:50.673760 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6b0b8ea3-e635-4dcf-8b35-04abc468d724-dispersionconf\") pod \"6b0b8ea3-e635-4dcf-8b35-04abc468d724\" (UID: \"6b0b8ea3-e635-4dcf-8b35-04abc468d724\") " Mar 07 15:19:50 crc kubenswrapper[4943]: I0307 15:19:50.673831 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6b0b8ea3-e635-4dcf-8b35-04abc468d724-ring-data-devices\") pod \"6b0b8ea3-e635-4dcf-8b35-04abc468d724\" (UID: \"6b0b8ea3-e635-4dcf-8b35-04abc468d724\") " Mar 07 15:19:50 crc kubenswrapper[4943]: I0307 15:19:50.673880 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6b0b8ea3-e635-4dcf-8b35-04abc468d724-scripts\") pod \"6b0b8ea3-e635-4dcf-8b35-04abc468d724\" (UID: \"6b0b8ea3-e635-4dcf-8b35-04abc468d724\") " Mar 07 15:19:50 crc kubenswrapper[4943]: I0307 15:19:50.673957 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6b0b8ea3-e635-4dcf-8b35-04abc468d724-etc-swift\") pod \"6b0b8ea3-e635-4dcf-8b35-04abc468d724\" (UID: \"6b0b8ea3-e635-4dcf-8b35-04abc468d724\") " Mar 07 15:19:50 crc kubenswrapper[4943]: I0307 15:19:50.673986 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6b0b8ea3-e635-4dcf-8b35-04abc468d724-swiftconf\") pod \"6b0b8ea3-e635-4dcf-8b35-04abc468d724\" (UID: \"6b0b8ea3-e635-4dcf-8b35-04abc468d724\") " Mar 07 15:19:50 crc kubenswrapper[4943]: I0307 15:19:50.674025 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-44pq5\" (UniqueName: \"kubernetes.io/projected/6b0b8ea3-e635-4dcf-8b35-04abc468d724-kube-api-access-44pq5\") pod \"6b0b8ea3-e635-4dcf-8b35-04abc468d724\" (UID: \"6b0b8ea3-e635-4dcf-8b35-04abc468d724\") " Mar 07 15:19:50 crc kubenswrapper[4943]: I0307 15:19:50.675535 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b0b8ea3-e635-4dcf-8b35-04abc468d724-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "6b0b8ea3-e635-4dcf-8b35-04abc468d724" (UID: "6b0b8ea3-e635-4dcf-8b35-04abc468d724"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:19:50 crc kubenswrapper[4943]: I0307 15:19:50.676195 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b0b8ea3-e635-4dcf-8b35-04abc468d724-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "6b0b8ea3-e635-4dcf-8b35-04abc468d724" (UID: "6b0b8ea3-e635-4dcf-8b35-04abc468d724"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:19:50 crc kubenswrapper[4943]: I0307 15:19:50.699738 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b0b8ea3-e635-4dcf-8b35-04abc468d724-kube-api-access-44pq5" (OuterVolumeSpecName: "kube-api-access-44pq5") pod "6b0b8ea3-e635-4dcf-8b35-04abc468d724" (UID: "6b0b8ea3-e635-4dcf-8b35-04abc468d724"). InnerVolumeSpecName "kube-api-access-44pq5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:19:50 crc kubenswrapper[4943]: I0307 15:19:50.702616 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b0b8ea3-e635-4dcf-8b35-04abc468d724-scripts" (OuterVolumeSpecName: "scripts") pod "6b0b8ea3-e635-4dcf-8b35-04abc468d724" (UID: "6b0b8ea3-e635-4dcf-8b35-04abc468d724"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:19:50 crc kubenswrapper[4943]: I0307 15:19:50.704659 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b0b8ea3-e635-4dcf-8b35-04abc468d724-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "6b0b8ea3-e635-4dcf-8b35-04abc468d724" (UID: "6b0b8ea3-e635-4dcf-8b35-04abc468d724"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:19:50 crc kubenswrapper[4943]: I0307 15:19:50.712018 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b0b8ea3-e635-4dcf-8b35-04abc468d724-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "6b0b8ea3-e635-4dcf-8b35-04abc468d724" (UID: "6b0b8ea3-e635-4dcf-8b35-04abc468d724"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:19:50 crc kubenswrapper[4943]: I0307 15:19:50.773045 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b0b8ea3-e635-4dcf-8b35-04abc468d724" path="/var/lib/kubelet/pods/6b0b8ea3-e635-4dcf-8b35-04abc468d724/volumes" Mar 07 15:19:50 crc kubenswrapper[4943]: I0307 15:19:50.775688 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6b0b8ea3-e635-4dcf-8b35-04abc468d724-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:50 crc kubenswrapper[4943]: I0307 15:19:50.775804 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6b0b8ea3-e635-4dcf-8b35-04abc468d724-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:50 crc kubenswrapper[4943]: I0307 15:19:50.775881 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-44pq5\" (UniqueName: \"kubernetes.io/projected/6b0b8ea3-e635-4dcf-8b35-04abc468d724-kube-api-access-44pq5\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:50 crc kubenswrapper[4943]: I0307 15:19:50.775999 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6b0b8ea3-e635-4dcf-8b35-04abc468d724-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:50 crc kubenswrapper[4943]: I0307 15:19:50.776534 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6b0b8ea3-e635-4dcf-8b35-04abc468d724-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:50 crc kubenswrapper[4943]: I0307 15:19:50.776645 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6b0b8ea3-e635-4dcf-8b35-04abc468d724-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:50 crc kubenswrapper[4943]: I0307 15:19:50.799054 4943 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod4585133d-e3c5-4530-a96d-d7d4f28a4d20"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod4585133d-e3c5-4530-a96d-d7d4f28a4d20] : Timed out while waiting for systemd to remove kubepods-besteffort-pod4585133d_e3c5_4530_a96d_d7d4f28a4d20.slice" Mar 07 15:19:51 crc kubenswrapper[4943]: I0307 15:19:51.252648 4943 scope.go:117] "RemoveContainer" containerID="98c57785988edac00b070e088586579b62685a8b1d3d70d1725cb6cc2cd845bb" Mar 07 15:19:51 crc kubenswrapper[4943]: I0307 15:19:51.252699 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-rmbq7" Mar 07 15:19:51 crc kubenswrapper[4943]: I0307 15:19:51.827143 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-j6n95"] Mar 07 15:19:51 crc kubenswrapper[4943]: E0307 15:19:51.827671 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b0b8ea3-e635-4dcf-8b35-04abc468d724" containerName="swift-ring-rebalance" Mar 07 15:19:51 crc kubenswrapper[4943]: I0307 15:19:51.827699 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b0b8ea3-e635-4dcf-8b35-04abc468d724" containerName="swift-ring-rebalance" Mar 07 15:19:51 crc kubenswrapper[4943]: I0307 15:19:51.828302 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b0b8ea3-e635-4dcf-8b35-04abc468d724" containerName="swift-ring-rebalance" Mar 07 15:19:51 crc kubenswrapper[4943]: I0307 15:19:51.829242 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j6n95" Mar 07 15:19:51 crc kubenswrapper[4943]: I0307 15:19:51.833375 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:19:51 crc kubenswrapper[4943]: I0307 15:19:51.833386 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:19:51 crc kubenswrapper[4943]: I0307 15:19:51.861308 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-j6n95"] Mar 07 15:19:51 crc kubenswrapper[4943]: I0307 15:19:51.895831 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d-ring-data-devices\") pod \"swift-ring-rebalance-debug-j6n95\" (UID: \"1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j6n95" Mar 07 15:19:51 crc kubenswrapper[4943]: I0307 15:19:51.896137 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d-swiftconf\") pod \"swift-ring-rebalance-debug-j6n95\" (UID: \"1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j6n95" Mar 07 15:19:51 crc kubenswrapper[4943]: I0307 15:19:51.896254 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d-dispersionconf\") pod \"swift-ring-rebalance-debug-j6n95\" (UID: \"1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j6n95" Mar 07 15:19:51 crc kubenswrapper[4943]: I0307 15:19:51.896369 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d-scripts\") pod \"swift-ring-rebalance-debug-j6n95\" (UID: \"1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j6n95" Mar 07 15:19:51 crc kubenswrapper[4943]: I0307 15:19:51.896578 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p69gd\" (UniqueName: \"kubernetes.io/projected/1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d-kube-api-access-p69gd\") pod \"swift-ring-rebalance-debug-j6n95\" (UID: \"1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j6n95" Mar 07 15:19:51 crc kubenswrapper[4943]: I0307 15:19:51.896691 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d-etc-swift\") pod \"swift-ring-rebalance-debug-j6n95\" (UID: \"1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j6n95" Mar 07 15:19:51 crc kubenswrapper[4943]: I0307 15:19:51.998261 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d-ring-data-devices\") pod \"swift-ring-rebalance-debug-j6n95\" (UID: \"1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j6n95" Mar 07 15:19:51 crc kubenswrapper[4943]: I0307 15:19:51.998690 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d-swiftconf\") pod \"swift-ring-rebalance-debug-j6n95\" (UID: \"1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j6n95" Mar 07 15:19:51 crc kubenswrapper[4943]: I0307 15:19:51.998917 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d-dispersionconf\") pod \"swift-ring-rebalance-debug-j6n95\" (UID: \"1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j6n95" Mar 07 15:19:51 crc kubenswrapper[4943]: I0307 15:19:51.999141 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d-scripts\") pod \"swift-ring-rebalance-debug-j6n95\" (UID: \"1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j6n95" Mar 07 15:19:51 crc kubenswrapper[4943]: I0307 15:19:51.999341 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p69gd\" (UniqueName: \"kubernetes.io/projected/1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d-kube-api-access-p69gd\") pod \"swift-ring-rebalance-debug-j6n95\" (UID: \"1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j6n95" Mar 07 15:19:51 crc kubenswrapper[4943]: I0307 15:19:51.999547 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d-etc-swift\") pod \"swift-ring-rebalance-debug-j6n95\" (UID: \"1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j6n95" Mar 07 15:19:51 crc kubenswrapper[4943]: I0307 15:19:51.999172 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d-ring-data-devices\") pod \"swift-ring-rebalance-debug-j6n95\" (UID: \"1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j6n95" Mar 07 15:19:51 crc kubenswrapper[4943]: I0307 15:19:51.999732 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d-scripts\") pod \"swift-ring-rebalance-debug-j6n95\" (UID: \"1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j6n95" Mar 07 15:19:52 crc kubenswrapper[4943]: I0307 15:19:52.000577 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d-etc-swift\") pod \"swift-ring-rebalance-debug-j6n95\" (UID: \"1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j6n95" Mar 07 15:19:52 crc kubenswrapper[4943]: I0307 15:19:52.006864 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d-swiftconf\") pod \"swift-ring-rebalance-debug-j6n95\" (UID: \"1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j6n95" Mar 07 15:19:52 crc kubenswrapper[4943]: I0307 15:19:52.006978 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d-dispersionconf\") pod \"swift-ring-rebalance-debug-j6n95\" (UID: \"1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j6n95" Mar 07 15:19:52 crc kubenswrapper[4943]: I0307 15:19:52.029719 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p69gd\" (UniqueName: \"kubernetes.io/projected/1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d-kube-api-access-p69gd\") pod \"swift-ring-rebalance-debug-j6n95\" (UID: \"1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j6n95" Mar 07 15:19:52 crc kubenswrapper[4943]: I0307 15:19:52.155950 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j6n95" Mar 07 15:19:52 crc kubenswrapper[4943]: I0307 15:19:52.445953 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-j6n95"] Mar 07 15:19:52 crc kubenswrapper[4943]: W0307 15:19:52.451074 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1cdb40f7_6880_4b66_b7c2_a0445b8a9f3d.slice/crio-970a8a2aa99c6684b395a6eac588226ee11902cd33965b5d5a5a3fdc9db7aa57 WatchSource:0}: Error finding container 970a8a2aa99c6684b395a6eac588226ee11902cd33965b5d5a5a3fdc9db7aa57: Status 404 returned error can't find the container with id 970a8a2aa99c6684b395a6eac588226ee11902cd33965b5d5a5a3fdc9db7aa57 Mar 07 15:19:53 crc kubenswrapper[4943]: I0307 15:19:53.299354 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j6n95" event={"ID":"1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d","Type":"ContainerStarted","Data":"d442999f3351d46e800c4d2e91511ae83e9ab8d0dd60e769149db07b2ca7024e"} Mar 07 15:19:53 crc kubenswrapper[4943]: I0307 15:19:53.299731 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j6n95" event={"ID":"1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d","Type":"ContainerStarted","Data":"970a8a2aa99c6684b395a6eac588226ee11902cd33965b5d5a5a3fdc9db7aa57"} Mar 07 15:19:53 crc kubenswrapper[4943]: I0307 15:19:53.333789 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j6n95" podStartSLOduration=2.333760819 podStartE2EDuration="2.333760819s" podCreationTimestamp="2026-03-07 15:19:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:19:53.322022322 +0000 UTC m=+2435.274158870" watchObservedRunningTime="2026-03-07 15:19:53.333760819 +0000 UTC m=+2435.285897347" Mar 07 15:19:53 crc kubenswrapper[4943]: I0307 15:19:53.755294 4943 scope.go:117] "RemoveContainer" containerID="448308ad6e34ac922d66b3c90e66d0927407a94dd28633ec350502d69e4ca8aa" Mar 07 15:19:53 crc kubenswrapper[4943]: E0307 15:19:53.755974 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cpdmm_openshift-machine-config-operator(4c1f18ea-e557-41fc-95dd-ce4e1722a61e)\"" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" Mar 07 15:19:54 crc kubenswrapper[4943]: I0307 15:19:54.315351 4943 generic.go:334] "Generic (PLEG): container finished" podID="1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d" containerID="d442999f3351d46e800c4d2e91511ae83e9ab8d0dd60e769149db07b2ca7024e" exitCode=0 Mar 07 15:19:54 crc kubenswrapper[4943]: I0307 15:19:54.315430 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j6n95" event={"ID":"1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d","Type":"ContainerDied","Data":"d442999f3351d46e800c4d2e91511ae83e9ab8d0dd60e769149db07b2ca7024e"} Mar 07 15:19:55 crc kubenswrapper[4943]: I0307 15:19:55.703050 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j6n95" Mar 07 15:19:55 crc kubenswrapper[4943]: I0307 15:19:55.746320 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-j6n95"] Mar 07 15:19:55 crc kubenswrapper[4943]: I0307 15:19:55.754363 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-j6n95"] Mar 07 15:19:55 crc kubenswrapper[4943]: I0307 15:19:55.865887 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d-scripts\") pod \"1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d\" (UID: \"1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d\") " Mar 07 15:19:55 crc kubenswrapper[4943]: I0307 15:19:55.867151 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p69gd\" (UniqueName: \"kubernetes.io/projected/1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d-kube-api-access-p69gd\") pod \"1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d\" (UID: \"1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d\") " Mar 07 15:19:55 crc kubenswrapper[4943]: I0307 15:19:55.867232 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d-swiftconf\") pod \"1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d\" (UID: \"1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d\") " Mar 07 15:19:55 crc kubenswrapper[4943]: I0307 15:19:55.867333 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d-etc-swift\") pod \"1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d\" (UID: \"1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d\") " Mar 07 15:19:55 crc kubenswrapper[4943]: I0307 15:19:55.867368 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d-dispersionconf\") pod \"1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d\" (UID: \"1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d\") " Mar 07 15:19:55 crc kubenswrapper[4943]: I0307 15:19:55.867415 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d-ring-data-devices\") pod \"1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d\" (UID: \"1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d\") " Mar 07 15:19:55 crc kubenswrapper[4943]: I0307 15:19:55.868805 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d" (UID: "1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:19:55 crc kubenswrapper[4943]: I0307 15:19:55.870271 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d" (UID: "1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:19:55 crc kubenswrapper[4943]: I0307 15:19:55.874851 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d-kube-api-access-p69gd" (OuterVolumeSpecName: "kube-api-access-p69gd") pod "1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d" (UID: "1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d"). InnerVolumeSpecName "kube-api-access-p69gd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:19:55 crc kubenswrapper[4943]: I0307 15:19:55.894711 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d" (UID: "1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:19:55 crc kubenswrapper[4943]: E0307 15:19:55.901491 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d-scripts podName:1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d nodeName:}" failed. No retries permitted until 2026-03-07 15:19:56.40145779 +0000 UTC m=+2438.353594318 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "scripts" (UniqueName: "kubernetes.io/configmap/1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d-scripts") pod "1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d" (UID: "1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d") : error deleting /var/lib/kubelet/pods/1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d/volume-subpaths: remove /var/lib/kubelet/pods/1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d/volume-subpaths: no such file or directory Mar 07 15:19:55 crc kubenswrapper[4943]: I0307 15:19:55.904996 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d" (UID: "1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:19:55 crc kubenswrapper[4943]: I0307 15:19:55.970240 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:55 crc kubenswrapper[4943]: I0307 15:19:55.970291 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:55 crc kubenswrapper[4943]: I0307 15:19:55.970312 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:55 crc kubenswrapper[4943]: I0307 15:19:55.970331 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p69gd\" (UniqueName: \"kubernetes.io/projected/1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d-kube-api-access-p69gd\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:55 crc kubenswrapper[4943]: I0307 15:19:55.970350 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:56 crc kubenswrapper[4943]: I0307 15:19:56.342629 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="970a8a2aa99c6684b395a6eac588226ee11902cd33965b5d5a5a3fdc9db7aa57" Mar 07 15:19:56 crc kubenswrapper[4943]: I0307 15:19:56.342723 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j6n95" Mar 07 15:19:56 crc kubenswrapper[4943]: I0307 15:19:56.478045 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d-scripts\") pod \"1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d\" (UID: \"1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d\") " Mar 07 15:19:56 crc kubenswrapper[4943]: I0307 15:19:56.478738 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d-scripts" (OuterVolumeSpecName: "scripts") pod "1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d" (UID: "1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:19:56 crc kubenswrapper[4943]: I0307 15:19:56.581005 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:19:56 crc kubenswrapper[4943]: I0307 15:19:56.772618 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d" path="/var/lib/kubelet/pods/1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d/volumes" Mar 07 15:19:56 crc kubenswrapper[4943]: I0307 15:19:56.973840 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-wgjk5"] Mar 07 15:19:56 crc kubenswrapper[4943]: E0307 15:19:56.974183 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d" containerName="swift-ring-rebalance" Mar 07 15:19:56 crc kubenswrapper[4943]: I0307 15:19:56.974197 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d" containerName="swift-ring-rebalance" Mar 07 15:19:56 crc kubenswrapper[4943]: I0307 15:19:56.974419 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="1cdb40f7-6880-4b66-b7c2-a0445b8a9f3d" containerName="swift-ring-rebalance" Mar 07 15:19:56 crc kubenswrapper[4943]: I0307 15:19:56.975077 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wgjk5" Mar 07 15:19:56 crc kubenswrapper[4943]: I0307 15:19:56.981567 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:19:56 crc kubenswrapper[4943]: I0307 15:19:56.981755 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:19:56 crc kubenswrapper[4943]: I0307 15:19:56.996050 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-wgjk5"] Mar 07 15:19:57 crc kubenswrapper[4943]: I0307 15:19:57.097473 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d5bf8088-fc2f-4a26-a0d1-469e56f912b5-swiftconf\") pod \"swift-ring-rebalance-debug-wgjk5\" (UID: \"d5bf8088-fc2f-4a26-a0d1-469e56f912b5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wgjk5" Mar 07 15:19:57 crc kubenswrapper[4943]: I0307 15:19:57.097594 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d5bf8088-fc2f-4a26-a0d1-469e56f912b5-ring-data-devices\") pod \"swift-ring-rebalance-debug-wgjk5\" (UID: \"d5bf8088-fc2f-4a26-a0d1-469e56f912b5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wgjk5" Mar 07 15:19:57 crc kubenswrapper[4943]: I0307 15:19:57.097622 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d5bf8088-fc2f-4a26-a0d1-469e56f912b5-etc-swift\") pod \"swift-ring-rebalance-debug-wgjk5\" (UID: \"d5bf8088-fc2f-4a26-a0d1-469e56f912b5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wgjk5" Mar 07 15:19:57 crc kubenswrapper[4943]: I0307 15:19:57.097646 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhgsg\" (UniqueName: \"kubernetes.io/projected/d5bf8088-fc2f-4a26-a0d1-469e56f912b5-kube-api-access-nhgsg\") pod \"swift-ring-rebalance-debug-wgjk5\" (UID: \"d5bf8088-fc2f-4a26-a0d1-469e56f912b5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wgjk5" Mar 07 15:19:57 crc kubenswrapper[4943]: I0307 15:19:57.097687 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d5bf8088-fc2f-4a26-a0d1-469e56f912b5-dispersionconf\") pod \"swift-ring-rebalance-debug-wgjk5\" (UID: \"d5bf8088-fc2f-4a26-a0d1-469e56f912b5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wgjk5" Mar 07 15:19:57 crc kubenswrapper[4943]: I0307 15:19:57.097885 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d5bf8088-fc2f-4a26-a0d1-469e56f912b5-scripts\") pod \"swift-ring-rebalance-debug-wgjk5\" (UID: \"d5bf8088-fc2f-4a26-a0d1-469e56f912b5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wgjk5" Mar 07 15:19:57 crc kubenswrapper[4943]: I0307 15:19:57.199622 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d5bf8088-fc2f-4a26-a0d1-469e56f912b5-swiftconf\") pod \"swift-ring-rebalance-debug-wgjk5\" (UID: \"d5bf8088-fc2f-4a26-a0d1-469e56f912b5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wgjk5" Mar 07 15:19:57 crc kubenswrapper[4943]: I0307 15:19:57.200140 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d5bf8088-fc2f-4a26-a0d1-469e56f912b5-ring-data-devices\") pod \"swift-ring-rebalance-debug-wgjk5\" (UID: \"d5bf8088-fc2f-4a26-a0d1-469e56f912b5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wgjk5" Mar 07 15:19:57 crc kubenswrapper[4943]: I0307 15:19:57.200188 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d5bf8088-fc2f-4a26-a0d1-469e56f912b5-etc-swift\") pod \"swift-ring-rebalance-debug-wgjk5\" (UID: \"d5bf8088-fc2f-4a26-a0d1-469e56f912b5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wgjk5" Mar 07 15:19:57 crc kubenswrapper[4943]: I0307 15:19:57.200214 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhgsg\" (UniqueName: \"kubernetes.io/projected/d5bf8088-fc2f-4a26-a0d1-469e56f912b5-kube-api-access-nhgsg\") pod \"swift-ring-rebalance-debug-wgjk5\" (UID: \"d5bf8088-fc2f-4a26-a0d1-469e56f912b5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wgjk5" Mar 07 15:19:57 crc kubenswrapper[4943]: I0307 15:19:57.200267 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d5bf8088-fc2f-4a26-a0d1-469e56f912b5-dispersionconf\") pod \"swift-ring-rebalance-debug-wgjk5\" (UID: \"d5bf8088-fc2f-4a26-a0d1-469e56f912b5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wgjk5" Mar 07 15:19:57 crc kubenswrapper[4943]: I0307 15:19:57.200313 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d5bf8088-fc2f-4a26-a0d1-469e56f912b5-scripts\") pod \"swift-ring-rebalance-debug-wgjk5\" (UID: \"d5bf8088-fc2f-4a26-a0d1-469e56f912b5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wgjk5" Mar 07 15:19:57 crc kubenswrapper[4943]: I0307 15:19:57.201651 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d5bf8088-fc2f-4a26-a0d1-469e56f912b5-etc-swift\") pod \"swift-ring-rebalance-debug-wgjk5\" (UID: \"d5bf8088-fc2f-4a26-a0d1-469e56f912b5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wgjk5" Mar 07 15:19:57 crc kubenswrapper[4943]: I0307 15:19:57.202833 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d5bf8088-fc2f-4a26-a0d1-469e56f912b5-ring-data-devices\") pod \"swift-ring-rebalance-debug-wgjk5\" (UID: \"d5bf8088-fc2f-4a26-a0d1-469e56f912b5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wgjk5" Mar 07 15:19:57 crc kubenswrapper[4943]: I0307 15:19:57.202999 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d5bf8088-fc2f-4a26-a0d1-469e56f912b5-scripts\") pod \"swift-ring-rebalance-debug-wgjk5\" (UID: \"d5bf8088-fc2f-4a26-a0d1-469e56f912b5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wgjk5" Mar 07 15:19:57 crc kubenswrapper[4943]: I0307 15:19:57.204701 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d5bf8088-fc2f-4a26-a0d1-469e56f912b5-swiftconf\") pod \"swift-ring-rebalance-debug-wgjk5\" (UID: \"d5bf8088-fc2f-4a26-a0d1-469e56f912b5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wgjk5" Mar 07 15:19:57 crc kubenswrapper[4943]: I0307 15:19:57.208488 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d5bf8088-fc2f-4a26-a0d1-469e56f912b5-dispersionconf\") pod \"swift-ring-rebalance-debug-wgjk5\" (UID: \"d5bf8088-fc2f-4a26-a0d1-469e56f912b5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wgjk5" Mar 07 15:19:57 crc kubenswrapper[4943]: I0307 15:19:57.219644 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhgsg\" (UniqueName: \"kubernetes.io/projected/d5bf8088-fc2f-4a26-a0d1-469e56f912b5-kube-api-access-nhgsg\") pod \"swift-ring-rebalance-debug-wgjk5\" (UID: \"d5bf8088-fc2f-4a26-a0d1-469e56f912b5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wgjk5" Mar 07 15:19:57 crc kubenswrapper[4943]: I0307 15:19:57.307693 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wgjk5" Mar 07 15:19:57 crc kubenswrapper[4943]: I0307 15:19:57.783909 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-wgjk5"] Mar 07 15:19:57 crc kubenswrapper[4943]: W0307 15:19:57.791716 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd5bf8088_fc2f_4a26_a0d1_469e56f912b5.slice/crio-7683ec2794d22528df7d3bc9493ca29cd7468f769ec5ff52b727d489adcede4f WatchSource:0}: Error finding container 7683ec2794d22528df7d3bc9493ca29cd7468f769ec5ff52b727d489adcede4f: Status 404 returned error can't find the container with id 7683ec2794d22528df7d3bc9493ca29cd7468f769ec5ff52b727d489adcede4f Mar 07 15:19:58 crc kubenswrapper[4943]: I0307 15:19:58.368427 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wgjk5" event={"ID":"d5bf8088-fc2f-4a26-a0d1-469e56f912b5","Type":"ContainerStarted","Data":"9aa118624744f146b373f21bda642afdae46954f0c5a6428e17c76aaf87843a9"} Mar 07 15:19:58 crc kubenswrapper[4943]: I0307 15:19:58.370450 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wgjk5" event={"ID":"d5bf8088-fc2f-4a26-a0d1-469e56f912b5","Type":"ContainerStarted","Data":"7683ec2794d22528df7d3bc9493ca29cd7468f769ec5ff52b727d489adcede4f"} Mar 07 15:19:58 crc kubenswrapper[4943]: I0307 15:19:58.395591 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wgjk5" podStartSLOduration=2.395567939 podStartE2EDuration="2.395567939s" podCreationTimestamp="2026-03-07 15:19:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:19:58.391197712 +0000 UTC m=+2440.343334280" watchObservedRunningTime="2026-03-07 15:19:58.395567939 +0000 UTC m=+2440.347704477" Mar 07 15:19:59 crc kubenswrapper[4943]: I0307 15:19:59.383675 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wgjk5" event={"ID":"d5bf8088-fc2f-4a26-a0d1-469e56f912b5","Type":"ContainerDied","Data":"9aa118624744f146b373f21bda642afdae46954f0c5a6428e17c76aaf87843a9"} Mar 07 15:19:59 crc kubenswrapper[4943]: I0307 15:19:59.383486 4943 generic.go:334] "Generic (PLEG): container finished" podID="d5bf8088-fc2f-4a26-a0d1-469e56f912b5" containerID="9aa118624744f146b373f21bda642afdae46954f0c5a6428e17c76aaf87843a9" exitCode=0 Mar 07 15:20:00 crc kubenswrapper[4943]: I0307 15:20:00.151200 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29548280-q7zxj"] Mar 07 15:20:00 crc kubenswrapper[4943]: I0307 15:20:00.152661 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548280-q7zxj" Mar 07 15:20:00 crc kubenswrapper[4943]: I0307 15:20:00.155319 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 07 15:20:00 crc kubenswrapper[4943]: I0307 15:20:00.155338 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-qqlk5" Mar 07 15:20:00 crc kubenswrapper[4943]: I0307 15:20:00.157215 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 07 15:20:00 crc kubenswrapper[4943]: I0307 15:20:00.166542 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29548280-q7zxj"] Mar 07 15:20:00 crc kubenswrapper[4943]: I0307 15:20:00.254418 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wmjz\" (UniqueName: \"kubernetes.io/projected/a63975e7-b7d9-4ab0-aee1-fcdd68e805cf-kube-api-access-8wmjz\") pod \"auto-csr-approver-29548280-q7zxj\" (UID: \"a63975e7-b7d9-4ab0-aee1-fcdd68e805cf\") " pod="openshift-infra/auto-csr-approver-29548280-q7zxj" Mar 07 15:20:00 crc kubenswrapper[4943]: I0307 15:20:00.355865 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wmjz\" (UniqueName: \"kubernetes.io/projected/a63975e7-b7d9-4ab0-aee1-fcdd68e805cf-kube-api-access-8wmjz\") pod \"auto-csr-approver-29548280-q7zxj\" (UID: \"a63975e7-b7d9-4ab0-aee1-fcdd68e805cf\") " pod="openshift-infra/auto-csr-approver-29548280-q7zxj" Mar 07 15:20:00 crc kubenswrapper[4943]: I0307 15:20:00.392020 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wmjz\" (UniqueName: \"kubernetes.io/projected/a63975e7-b7d9-4ab0-aee1-fcdd68e805cf-kube-api-access-8wmjz\") pod \"auto-csr-approver-29548280-q7zxj\" (UID: \"a63975e7-b7d9-4ab0-aee1-fcdd68e805cf\") " pod="openshift-infra/auto-csr-approver-29548280-q7zxj" Mar 07 15:20:00 crc kubenswrapper[4943]: I0307 15:20:00.489758 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548280-q7zxj" Mar 07 15:20:00 crc kubenswrapper[4943]: I0307 15:20:00.774303 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wgjk5" Mar 07 15:20:00 crc kubenswrapper[4943]: I0307 15:20:00.813625 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-wgjk5"] Mar 07 15:20:00 crc kubenswrapper[4943]: I0307 15:20:00.820377 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-wgjk5"] Mar 07 15:20:00 crc kubenswrapper[4943]: I0307 15:20:00.863173 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d5bf8088-fc2f-4a26-a0d1-469e56f912b5-swiftconf\") pod \"d5bf8088-fc2f-4a26-a0d1-469e56f912b5\" (UID: \"d5bf8088-fc2f-4a26-a0d1-469e56f912b5\") " Mar 07 15:20:00 crc kubenswrapper[4943]: I0307 15:20:00.863336 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d5bf8088-fc2f-4a26-a0d1-469e56f912b5-ring-data-devices\") pod \"d5bf8088-fc2f-4a26-a0d1-469e56f912b5\" (UID: \"d5bf8088-fc2f-4a26-a0d1-469e56f912b5\") " Mar 07 15:20:00 crc kubenswrapper[4943]: I0307 15:20:00.863471 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nhgsg\" (UniqueName: \"kubernetes.io/projected/d5bf8088-fc2f-4a26-a0d1-469e56f912b5-kube-api-access-nhgsg\") pod \"d5bf8088-fc2f-4a26-a0d1-469e56f912b5\" (UID: \"d5bf8088-fc2f-4a26-a0d1-469e56f912b5\") " Mar 07 15:20:00 crc kubenswrapper[4943]: I0307 15:20:00.863527 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d5bf8088-fc2f-4a26-a0d1-469e56f912b5-dispersionconf\") pod \"d5bf8088-fc2f-4a26-a0d1-469e56f912b5\" (UID: \"d5bf8088-fc2f-4a26-a0d1-469e56f912b5\") " Mar 07 15:20:00 crc kubenswrapper[4943]: I0307 15:20:00.863557 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d5bf8088-fc2f-4a26-a0d1-469e56f912b5-scripts\") pod \"d5bf8088-fc2f-4a26-a0d1-469e56f912b5\" (UID: \"d5bf8088-fc2f-4a26-a0d1-469e56f912b5\") " Mar 07 15:20:00 crc kubenswrapper[4943]: I0307 15:20:00.863589 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d5bf8088-fc2f-4a26-a0d1-469e56f912b5-etc-swift\") pod \"d5bf8088-fc2f-4a26-a0d1-469e56f912b5\" (UID: \"d5bf8088-fc2f-4a26-a0d1-469e56f912b5\") " Mar 07 15:20:00 crc kubenswrapper[4943]: I0307 15:20:00.864218 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5bf8088-fc2f-4a26-a0d1-469e56f912b5-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "d5bf8088-fc2f-4a26-a0d1-469e56f912b5" (UID: "d5bf8088-fc2f-4a26-a0d1-469e56f912b5"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:20:00 crc kubenswrapper[4943]: I0307 15:20:00.864390 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5bf8088-fc2f-4a26-a0d1-469e56f912b5-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "d5bf8088-fc2f-4a26-a0d1-469e56f912b5" (UID: "d5bf8088-fc2f-4a26-a0d1-469e56f912b5"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:20:00 crc kubenswrapper[4943]: I0307 15:20:00.866118 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d5bf8088-fc2f-4a26-a0d1-469e56f912b5-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:20:00 crc kubenswrapper[4943]: I0307 15:20:00.866145 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d5bf8088-fc2f-4a26-a0d1-469e56f912b5-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:20:00 crc kubenswrapper[4943]: I0307 15:20:00.868629 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5bf8088-fc2f-4a26-a0d1-469e56f912b5-kube-api-access-nhgsg" (OuterVolumeSpecName: "kube-api-access-nhgsg") pod "d5bf8088-fc2f-4a26-a0d1-469e56f912b5" (UID: "d5bf8088-fc2f-4a26-a0d1-469e56f912b5"). InnerVolumeSpecName "kube-api-access-nhgsg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:20:00 crc kubenswrapper[4943]: I0307 15:20:00.890130 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5bf8088-fc2f-4a26-a0d1-469e56f912b5-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "d5bf8088-fc2f-4a26-a0d1-469e56f912b5" (UID: "d5bf8088-fc2f-4a26-a0d1-469e56f912b5"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:20:00 crc kubenswrapper[4943]: I0307 15:20:00.897159 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5bf8088-fc2f-4a26-a0d1-469e56f912b5-scripts" (OuterVolumeSpecName: "scripts") pod "d5bf8088-fc2f-4a26-a0d1-469e56f912b5" (UID: "d5bf8088-fc2f-4a26-a0d1-469e56f912b5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:20:00 crc kubenswrapper[4943]: I0307 15:20:00.903112 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5bf8088-fc2f-4a26-a0d1-469e56f912b5-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "d5bf8088-fc2f-4a26-a0d1-469e56f912b5" (UID: "d5bf8088-fc2f-4a26-a0d1-469e56f912b5"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:20:00 crc kubenswrapper[4943]: I0307 15:20:00.967807 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nhgsg\" (UniqueName: \"kubernetes.io/projected/d5bf8088-fc2f-4a26-a0d1-469e56f912b5-kube-api-access-nhgsg\") on node \"crc\" DevicePath \"\"" Mar 07 15:20:00 crc kubenswrapper[4943]: I0307 15:20:00.967851 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d5bf8088-fc2f-4a26-a0d1-469e56f912b5-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:20:00 crc kubenswrapper[4943]: I0307 15:20:00.967866 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d5bf8088-fc2f-4a26-a0d1-469e56f912b5-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:20:00 crc kubenswrapper[4943]: I0307 15:20:00.967879 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d5bf8088-fc2f-4a26-a0d1-469e56f912b5-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:20:00 crc kubenswrapper[4943]: I0307 15:20:00.986439 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29548280-q7zxj"] Mar 07 15:20:00 crc kubenswrapper[4943]: W0307 15:20:00.995192 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda63975e7_b7d9_4ab0_aee1_fcdd68e805cf.slice/crio-1df0e534c7fd78b23b6000c86d844b604109c79689f96c43b566d6eaaab2d5e9 WatchSource:0}: Error finding container 1df0e534c7fd78b23b6000c86d844b604109c79689f96c43b566d6eaaab2d5e9: Status 404 returned error can't find the container with id 1df0e534c7fd78b23b6000c86d844b604109c79689f96c43b566d6eaaab2d5e9 Mar 07 15:20:01 crc kubenswrapper[4943]: I0307 15:20:01.413335 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7683ec2794d22528df7d3bc9493ca29cd7468f769ec5ff52b727d489adcede4f" Mar 07 15:20:01 crc kubenswrapper[4943]: I0307 15:20:01.413613 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wgjk5" Mar 07 15:20:01 crc kubenswrapper[4943]: I0307 15:20:01.416261 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548280-q7zxj" event={"ID":"a63975e7-b7d9-4ab0-aee1-fcdd68e805cf","Type":"ContainerStarted","Data":"1df0e534c7fd78b23b6000c86d844b604109c79689f96c43b566d6eaaab2d5e9"} Mar 07 15:20:02 crc kubenswrapper[4943]: I0307 15:20:02.018708 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-k6zdr"] Mar 07 15:20:02 crc kubenswrapper[4943]: E0307 15:20:02.019259 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5bf8088-fc2f-4a26-a0d1-469e56f912b5" containerName="swift-ring-rebalance" Mar 07 15:20:02 crc kubenswrapper[4943]: I0307 15:20:02.019292 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5bf8088-fc2f-4a26-a0d1-469e56f912b5" containerName="swift-ring-rebalance" Mar 07 15:20:02 crc kubenswrapper[4943]: I0307 15:20:02.019588 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5bf8088-fc2f-4a26-a0d1-469e56f912b5" containerName="swift-ring-rebalance" Mar 07 15:20:02 crc kubenswrapper[4943]: I0307 15:20:02.020347 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-k6zdr" Mar 07 15:20:02 crc kubenswrapper[4943]: I0307 15:20:02.025060 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:20:02 crc kubenswrapper[4943]: I0307 15:20:02.025191 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:20:02 crc kubenswrapper[4943]: I0307 15:20:02.056728 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-k6zdr"] Mar 07 15:20:02 crc kubenswrapper[4943]: I0307 15:20:02.100430 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba-dispersionconf\") pod \"swift-ring-rebalance-debug-k6zdr\" (UID: \"b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-k6zdr" Mar 07 15:20:02 crc kubenswrapper[4943]: I0307 15:20:02.100565 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba-ring-data-devices\") pod \"swift-ring-rebalance-debug-k6zdr\" (UID: \"b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-k6zdr" Mar 07 15:20:02 crc kubenswrapper[4943]: I0307 15:20:02.100839 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba-scripts\") pod \"swift-ring-rebalance-debug-k6zdr\" (UID: \"b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-k6zdr" Mar 07 15:20:02 crc kubenswrapper[4943]: I0307 15:20:02.100896 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba-etc-swift\") pod \"swift-ring-rebalance-debug-k6zdr\" (UID: \"b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-k6zdr" Mar 07 15:20:02 crc kubenswrapper[4943]: I0307 15:20:02.101200 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ch2np\" (UniqueName: \"kubernetes.io/projected/b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba-kube-api-access-ch2np\") pod \"swift-ring-rebalance-debug-k6zdr\" (UID: \"b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-k6zdr" Mar 07 15:20:02 crc kubenswrapper[4943]: I0307 15:20:02.105191 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba-swiftconf\") pod \"swift-ring-rebalance-debug-k6zdr\" (UID: \"b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-k6zdr" Mar 07 15:20:02 crc kubenswrapper[4943]: I0307 15:20:02.206953 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba-scripts\") pod \"swift-ring-rebalance-debug-k6zdr\" (UID: \"b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-k6zdr" Mar 07 15:20:02 crc kubenswrapper[4943]: I0307 15:20:02.207041 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba-etc-swift\") pod \"swift-ring-rebalance-debug-k6zdr\" (UID: \"b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-k6zdr" Mar 07 15:20:02 crc kubenswrapper[4943]: I0307 15:20:02.207132 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ch2np\" (UniqueName: \"kubernetes.io/projected/b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba-kube-api-access-ch2np\") pod \"swift-ring-rebalance-debug-k6zdr\" (UID: \"b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-k6zdr" Mar 07 15:20:02 crc kubenswrapper[4943]: I0307 15:20:02.207223 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba-swiftconf\") pod \"swift-ring-rebalance-debug-k6zdr\" (UID: \"b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-k6zdr" Mar 07 15:20:02 crc kubenswrapper[4943]: I0307 15:20:02.207272 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba-dispersionconf\") pod \"swift-ring-rebalance-debug-k6zdr\" (UID: \"b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-k6zdr" Mar 07 15:20:02 crc kubenswrapper[4943]: I0307 15:20:02.207333 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba-ring-data-devices\") pod \"swift-ring-rebalance-debug-k6zdr\" (UID: \"b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-k6zdr" Mar 07 15:20:02 crc kubenswrapper[4943]: I0307 15:20:02.207525 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba-etc-swift\") pod \"swift-ring-rebalance-debug-k6zdr\" (UID: \"b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-k6zdr" Mar 07 15:20:02 crc kubenswrapper[4943]: I0307 15:20:02.208326 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba-scripts\") pod \"swift-ring-rebalance-debug-k6zdr\" (UID: \"b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-k6zdr" Mar 07 15:20:02 crc kubenswrapper[4943]: I0307 15:20:02.210877 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba-ring-data-devices\") pod \"swift-ring-rebalance-debug-k6zdr\" (UID: \"b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-k6zdr" Mar 07 15:20:02 crc kubenswrapper[4943]: I0307 15:20:02.211398 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba-swiftconf\") pod \"swift-ring-rebalance-debug-k6zdr\" (UID: \"b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-k6zdr" Mar 07 15:20:02 crc kubenswrapper[4943]: I0307 15:20:02.215356 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba-dispersionconf\") pod \"swift-ring-rebalance-debug-k6zdr\" (UID: \"b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-k6zdr" Mar 07 15:20:02 crc kubenswrapper[4943]: I0307 15:20:02.228388 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ch2np\" (UniqueName: \"kubernetes.io/projected/b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba-kube-api-access-ch2np\") pod \"swift-ring-rebalance-debug-k6zdr\" (UID: \"b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-k6zdr" Mar 07 15:20:02 crc kubenswrapper[4943]: I0307 15:20:02.348071 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-k6zdr" Mar 07 15:20:02 crc kubenswrapper[4943]: I0307 15:20:02.426988 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548280-q7zxj" event={"ID":"a63975e7-b7d9-4ab0-aee1-fcdd68e805cf","Type":"ContainerStarted","Data":"a15c43c8f498a82a124d4b4e6c272172c48df774a1fa978357f13c98aa911040"} Mar 07 15:20:02 crc kubenswrapper[4943]: I0307 15:20:02.449135 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29548280-q7zxj" podStartSLOduration=1.440775984 podStartE2EDuration="2.449108788s" podCreationTimestamp="2026-03-07 15:20:00 +0000 UTC" firstStartedPulling="2026-03-07 15:20:00.997038389 +0000 UTC m=+2442.949174897" lastFinishedPulling="2026-03-07 15:20:02.005371173 +0000 UTC m=+2443.957507701" observedRunningTime="2026-03-07 15:20:02.443339057 +0000 UTC m=+2444.395475595" watchObservedRunningTime="2026-03-07 15:20:02.449108788 +0000 UTC m=+2444.401245326" Mar 07 15:20:02 crc kubenswrapper[4943]: I0307 15:20:02.642584 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-k6zdr"] Mar 07 15:20:02 crc kubenswrapper[4943]: W0307 15:20:02.653509 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb27e7fa4_ff8e_4fcd_bdd5_b9f4f9965eba.slice/crio-8569a283c36f42d2ddcb23151e793e20e3a73745744ee6163c40f468c9003087 WatchSource:0}: Error finding container 8569a283c36f42d2ddcb23151e793e20e3a73745744ee6163c40f468c9003087: Status 404 returned error can't find the container with id 8569a283c36f42d2ddcb23151e793e20e3a73745744ee6163c40f468c9003087 Mar 07 15:20:02 crc kubenswrapper[4943]: I0307 15:20:02.769087 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5bf8088-fc2f-4a26-a0d1-469e56f912b5" path="/var/lib/kubelet/pods/d5bf8088-fc2f-4a26-a0d1-469e56f912b5/volumes" Mar 07 15:20:03 crc kubenswrapper[4943]: I0307 15:20:03.439591 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-k6zdr" event={"ID":"b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba","Type":"ContainerStarted","Data":"e4cee84db2592c06ad43026174ee52161982e207a05d76fb520c5751c15115a7"} Mar 07 15:20:03 crc kubenswrapper[4943]: I0307 15:20:03.440010 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-k6zdr" event={"ID":"b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba","Type":"ContainerStarted","Data":"8569a283c36f42d2ddcb23151e793e20e3a73745744ee6163c40f468c9003087"} Mar 07 15:20:03 crc kubenswrapper[4943]: I0307 15:20:03.441703 4943 generic.go:334] "Generic (PLEG): container finished" podID="a63975e7-b7d9-4ab0-aee1-fcdd68e805cf" containerID="a15c43c8f498a82a124d4b4e6c272172c48df774a1fa978357f13c98aa911040" exitCode=0 Mar 07 15:20:03 crc kubenswrapper[4943]: I0307 15:20:03.441752 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548280-q7zxj" event={"ID":"a63975e7-b7d9-4ab0-aee1-fcdd68e805cf","Type":"ContainerDied","Data":"a15c43c8f498a82a124d4b4e6c272172c48df774a1fa978357f13c98aa911040"} Mar 07 15:20:03 crc kubenswrapper[4943]: I0307 15:20:03.471254 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-k6zdr" podStartSLOduration=2.47123264 podStartE2EDuration="2.47123264s" podCreationTimestamp="2026-03-07 15:20:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:20:03.463546522 +0000 UTC m=+2445.415683050" watchObservedRunningTime="2026-03-07 15:20:03.47123264 +0000 UTC m=+2445.423369168" Mar 07 15:20:04 crc kubenswrapper[4943]: I0307 15:20:04.465594 4943 generic.go:334] "Generic (PLEG): container finished" podID="b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba" containerID="e4cee84db2592c06ad43026174ee52161982e207a05d76fb520c5751c15115a7" exitCode=0 Mar 07 15:20:04 crc kubenswrapper[4943]: I0307 15:20:04.465718 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-k6zdr" event={"ID":"b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba","Type":"ContainerDied","Data":"e4cee84db2592c06ad43026174ee52161982e207a05d76fb520c5751c15115a7"} Mar 07 15:20:04 crc kubenswrapper[4943]: I0307 15:20:04.861372 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548280-q7zxj" Mar 07 15:20:04 crc kubenswrapper[4943]: I0307 15:20:04.953805 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8wmjz\" (UniqueName: \"kubernetes.io/projected/a63975e7-b7d9-4ab0-aee1-fcdd68e805cf-kube-api-access-8wmjz\") pod \"a63975e7-b7d9-4ab0-aee1-fcdd68e805cf\" (UID: \"a63975e7-b7d9-4ab0-aee1-fcdd68e805cf\") " Mar 07 15:20:04 crc kubenswrapper[4943]: I0307 15:20:04.959511 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a63975e7-b7d9-4ab0-aee1-fcdd68e805cf-kube-api-access-8wmjz" (OuterVolumeSpecName: "kube-api-access-8wmjz") pod "a63975e7-b7d9-4ab0-aee1-fcdd68e805cf" (UID: "a63975e7-b7d9-4ab0-aee1-fcdd68e805cf"). InnerVolumeSpecName "kube-api-access-8wmjz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:20:05 crc kubenswrapper[4943]: I0307 15:20:05.056210 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8wmjz\" (UniqueName: \"kubernetes.io/projected/a63975e7-b7d9-4ab0-aee1-fcdd68e805cf-kube-api-access-8wmjz\") on node \"crc\" DevicePath \"\"" Mar 07 15:20:05 crc kubenswrapper[4943]: I0307 15:20:05.484365 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548280-q7zxj" Mar 07 15:20:05 crc kubenswrapper[4943]: I0307 15:20:05.484405 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548280-q7zxj" event={"ID":"a63975e7-b7d9-4ab0-aee1-fcdd68e805cf","Type":"ContainerDied","Data":"1df0e534c7fd78b23b6000c86d844b604109c79689f96c43b566d6eaaab2d5e9"} Mar 07 15:20:05 crc kubenswrapper[4943]: I0307 15:20:05.484464 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1df0e534c7fd78b23b6000c86d844b604109c79689f96c43b566d6eaaab2d5e9" Mar 07 15:20:05 crc kubenswrapper[4943]: I0307 15:20:05.560783 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29548274-9dcnx"] Mar 07 15:20:05 crc kubenswrapper[4943]: I0307 15:20:05.570063 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29548274-9dcnx"] Mar 07 15:20:05 crc kubenswrapper[4943]: I0307 15:20:05.927680 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-k6zdr" Mar 07 15:20:05 crc kubenswrapper[4943]: I0307 15:20:05.973209 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-k6zdr"] Mar 07 15:20:05 crc kubenswrapper[4943]: I0307 15:20:05.981093 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-k6zdr"] Mar 07 15:20:06 crc kubenswrapper[4943]: I0307 15:20:06.074914 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ch2np\" (UniqueName: \"kubernetes.io/projected/b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba-kube-api-access-ch2np\") pod \"b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba\" (UID: \"b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba\") " Mar 07 15:20:06 crc kubenswrapper[4943]: I0307 15:20:06.074990 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba-ring-data-devices\") pod \"b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba\" (UID: \"b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba\") " Mar 07 15:20:06 crc kubenswrapper[4943]: I0307 15:20:06.075031 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba-etc-swift\") pod \"b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba\" (UID: \"b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba\") " Mar 07 15:20:06 crc kubenswrapper[4943]: I0307 15:20:06.075053 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba-dispersionconf\") pod \"b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba\" (UID: \"b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba\") " Mar 07 15:20:06 crc kubenswrapper[4943]: I0307 15:20:06.075084 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba-swiftconf\") pod \"b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba\" (UID: \"b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba\") " Mar 07 15:20:06 crc kubenswrapper[4943]: I0307 15:20:06.075118 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba-scripts\") pod \"b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba\" (UID: \"b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba\") " Mar 07 15:20:06 crc kubenswrapper[4943]: I0307 15:20:06.076238 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba" (UID: "b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:20:06 crc kubenswrapper[4943]: I0307 15:20:06.076503 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba" (UID: "b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:20:06 crc kubenswrapper[4943]: I0307 15:20:06.081847 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba-kube-api-access-ch2np" (OuterVolumeSpecName: "kube-api-access-ch2np") pod "b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba" (UID: "b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba"). InnerVolumeSpecName "kube-api-access-ch2np". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:20:06 crc kubenswrapper[4943]: I0307 15:20:06.093355 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba-scripts" (OuterVolumeSpecName: "scripts") pod "b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba" (UID: "b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:20:06 crc kubenswrapper[4943]: I0307 15:20:06.112335 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba" (UID: "b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:20:06 crc kubenswrapper[4943]: I0307 15:20:06.124108 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba" (UID: "b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:20:06 crc kubenswrapper[4943]: I0307 15:20:06.178145 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:20:06 crc kubenswrapper[4943]: I0307 15:20:06.178197 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:20:06 crc kubenswrapper[4943]: I0307 15:20:06.178222 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:20:06 crc kubenswrapper[4943]: I0307 15:20:06.178245 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ch2np\" (UniqueName: \"kubernetes.io/projected/b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba-kube-api-access-ch2np\") on node \"crc\" DevicePath \"\"" Mar 07 15:20:06 crc kubenswrapper[4943]: I0307 15:20:06.178271 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:20:06 crc kubenswrapper[4943]: I0307 15:20:06.178292 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:20:06 crc kubenswrapper[4943]: I0307 15:20:06.500828 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8569a283c36f42d2ddcb23151e793e20e3a73745744ee6163c40f468c9003087" Mar 07 15:20:06 crc kubenswrapper[4943]: I0307 15:20:06.500886 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-k6zdr" Mar 07 15:20:06 crc kubenswrapper[4943]: I0307 15:20:06.773017 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b797be7-6317-4ddc-95dd-652f99ba6a32" path="/var/lib/kubelet/pods/9b797be7-6317-4ddc-95dd-652f99ba6a32/volumes" Mar 07 15:20:06 crc kubenswrapper[4943]: I0307 15:20:06.773999 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba" path="/var/lib/kubelet/pods/b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba/volumes" Mar 07 15:20:07 crc kubenswrapper[4943]: I0307 15:20:07.199171 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-bbh8q"] Mar 07 15:20:07 crc kubenswrapper[4943]: E0307 15:20:07.199821 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba" containerName="swift-ring-rebalance" Mar 07 15:20:07 crc kubenswrapper[4943]: I0307 15:20:07.199842 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba" containerName="swift-ring-rebalance" Mar 07 15:20:07 crc kubenswrapper[4943]: E0307 15:20:07.199863 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a63975e7-b7d9-4ab0-aee1-fcdd68e805cf" containerName="oc" Mar 07 15:20:07 crc kubenswrapper[4943]: I0307 15:20:07.199873 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="a63975e7-b7d9-4ab0-aee1-fcdd68e805cf" containerName="oc" Mar 07 15:20:07 crc kubenswrapper[4943]: I0307 15:20:07.200077 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="a63975e7-b7d9-4ab0-aee1-fcdd68e805cf" containerName="oc" Mar 07 15:20:07 crc kubenswrapper[4943]: I0307 15:20:07.200107 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="b27e7fa4-ff8e-4fcd-bdd5-b9f4f9965eba" containerName="swift-ring-rebalance" Mar 07 15:20:07 crc kubenswrapper[4943]: I0307 15:20:07.200694 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-bbh8q" Mar 07 15:20:07 crc kubenswrapper[4943]: I0307 15:20:07.203332 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:20:07 crc kubenswrapper[4943]: I0307 15:20:07.206471 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:20:07 crc kubenswrapper[4943]: I0307 15:20:07.219065 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-bbh8q"] Mar 07 15:20:07 crc kubenswrapper[4943]: I0307 15:20:07.297097 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/07ac3ac5-33ba-436c-af70-0e7976486164-dispersionconf\") pod \"swift-ring-rebalance-debug-bbh8q\" (UID: \"07ac3ac5-33ba-436c-af70-0e7976486164\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bbh8q" Mar 07 15:20:07 crc kubenswrapper[4943]: I0307 15:20:07.297423 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljq9m\" (UniqueName: \"kubernetes.io/projected/07ac3ac5-33ba-436c-af70-0e7976486164-kube-api-access-ljq9m\") pod \"swift-ring-rebalance-debug-bbh8q\" (UID: \"07ac3ac5-33ba-436c-af70-0e7976486164\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bbh8q" Mar 07 15:20:07 crc kubenswrapper[4943]: I0307 15:20:07.297731 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/07ac3ac5-33ba-436c-af70-0e7976486164-etc-swift\") pod \"swift-ring-rebalance-debug-bbh8q\" (UID: \"07ac3ac5-33ba-436c-af70-0e7976486164\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bbh8q" Mar 07 15:20:07 crc kubenswrapper[4943]: I0307 15:20:07.297905 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/07ac3ac5-33ba-436c-af70-0e7976486164-ring-data-devices\") pod \"swift-ring-rebalance-debug-bbh8q\" (UID: \"07ac3ac5-33ba-436c-af70-0e7976486164\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bbh8q" Mar 07 15:20:07 crc kubenswrapper[4943]: I0307 15:20:07.298207 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/07ac3ac5-33ba-436c-af70-0e7976486164-swiftconf\") pod \"swift-ring-rebalance-debug-bbh8q\" (UID: \"07ac3ac5-33ba-436c-af70-0e7976486164\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bbh8q" Mar 07 15:20:07 crc kubenswrapper[4943]: I0307 15:20:07.298456 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/07ac3ac5-33ba-436c-af70-0e7976486164-scripts\") pod \"swift-ring-rebalance-debug-bbh8q\" (UID: \"07ac3ac5-33ba-436c-af70-0e7976486164\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bbh8q" Mar 07 15:20:07 crc kubenswrapper[4943]: I0307 15:20:07.400283 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/07ac3ac5-33ba-436c-af70-0e7976486164-etc-swift\") pod \"swift-ring-rebalance-debug-bbh8q\" (UID: \"07ac3ac5-33ba-436c-af70-0e7976486164\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bbh8q" Mar 07 15:20:07 crc kubenswrapper[4943]: I0307 15:20:07.400343 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/07ac3ac5-33ba-436c-af70-0e7976486164-ring-data-devices\") pod \"swift-ring-rebalance-debug-bbh8q\" (UID: \"07ac3ac5-33ba-436c-af70-0e7976486164\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bbh8q" Mar 07 15:20:07 crc kubenswrapper[4943]: I0307 15:20:07.400493 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/07ac3ac5-33ba-436c-af70-0e7976486164-swiftconf\") pod \"swift-ring-rebalance-debug-bbh8q\" (UID: \"07ac3ac5-33ba-436c-af70-0e7976486164\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bbh8q" Mar 07 15:20:07 crc kubenswrapper[4943]: I0307 15:20:07.400532 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/07ac3ac5-33ba-436c-af70-0e7976486164-scripts\") pod \"swift-ring-rebalance-debug-bbh8q\" (UID: \"07ac3ac5-33ba-436c-af70-0e7976486164\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bbh8q" Mar 07 15:20:07 crc kubenswrapper[4943]: I0307 15:20:07.400606 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/07ac3ac5-33ba-436c-af70-0e7976486164-dispersionconf\") pod \"swift-ring-rebalance-debug-bbh8q\" (UID: \"07ac3ac5-33ba-436c-af70-0e7976486164\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bbh8q" Mar 07 15:20:07 crc kubenswrapper[4943]: I0307 15:20:07.400643 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljq9m\" (UniqueName: \"kubernetes.io/projected/07ac3ac5-33ba-436c-af70-0e7976486164-kube-api-access-ljq9m\") pod \"swift-ring-rebalance-debug-bbh8q\" (UID: \"07ac3ac5-33ba-436c-af70-0e7976486164\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bbh8q" Mar 07 15:20:07 crc kubenswrapper[4943]: I0307 15:20:07.401023 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/07ac3ac5-33ba-436c-af70-0e7976486164-etc-swift\") pod \"swift-ring-rebalance-debug-bbh8q\" (UID: \"07ac3ac5-33ba-436c-af70-0e7976486164\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bbh8q" Mar 07 15:20:07 crc kubenswrapper[4943]: I0307 15:20:07.401733 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/07ac3ac5-33ba-436c-af70-0e7976486164-scripts\") pod \"swift-ring-rebalance-debug-bbh8q\" (UID: \"07ac3ac5-33ba-436c-af70-0e7976486164\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bbh8q" Mar 07 15:20:07 crc kubenswrapper[4943]: I0307 15:20:07.403255 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/07ac3ac5-33ba-436c-af70-0e7976486164-ring-data-devices\") pod \"swift-ring-rebalance-debug-bbh8q\" (UID: \"07ac3ac5-33ba-436c-af70-0e7976486164\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bbh8q" Mar 07 15:20:07 crc kubenswrapper[4943]: I0307 15:20:07.408538 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/07ac3ac5-33ba-436c-af70-0e7976486164-dispersionconf\") pod \"swift-ring-rebalance-debug-bbh8q\" (UID: \"07ac3ac5-33ba-436c-af70-0e7976486164\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bbh8q" Mar 07 15:20:07 crc kubenswrapper[4943]: I0307 15:20:07.409373 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/07ac3ac5-33ba-436c-af70-0e7976486164-swiftconf\") pod \"swift-ring-rebalance-debug-bbh8q\" (UID: \"07ac3ac5-33ba-436c-af70-0e7976486164\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bbh8q" Mar 07 15:20:07 crc kubenswrapper[4943]: I0307 15:20:07.423871 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljq9m\" (UniqueName: \"kubernetes.io/projected/07ac3ac5-33ba-436c-af70-0e7976486164-kube-api-access-ljq9m\") pod \"swift-ring-rebalance-debug-bbh8q\" (UID: \"07ac3ac5-33ba-436c-af70-0e7976486164\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bbh8q" Mar 07 15:20:07 crc kubenswrapper[4943]: I0307 15:20:07.522768 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-bbh8q" Mar 07 15:20:08 crc kubenswrapper[4943]: I0307 15:20:08.047210 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-bbh8q"] Mar 07 15:20:08 crc kubenswrapper[4943]: I0307 15:20:08.525984 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-bbh8q" event={"ID":"07ac3ac5-33ba-436c-af70-0e7976486164","Type":"ContainerStarted","Data":"a928e8c9e1805599cfcc0243ab9d559652e0cc68f084c5ad20b90e7af50975ca"} Mar 07 15:20:08 crc kubenswrapper[4943]: I0307 15:20:08.526366 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-bbh8q" event={"ID":"07ac3ac5-33ba-436c-af70-0e7976486164","Type":"ContainerStarted","Data":"93d9bd101d82d865eb49b58060997dbd4e3fb4fda25bbeaa22bef1f3c36a61e1"} Mar 07 15:20:08 crc kubenswrapper[4943]: I0307 15:20:08.549118 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-bbh8q" podStartSLOduration=1.549091796 podStartE2EDuration="1.549091796s" podCreationTimestamp="2026-03-07 15:20:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:20:08.547653441 +0000 UTC m=+2450.499789979" watchObservedRunningTime="2026-03-07 15:20:08.549091796 +0000 UTC m=+2450.501228324" Mar 07 15:20:08 crc kubenswrapper[4943]: I0307 15:20:08.767418 4943 scope.go:117] "RemoveContainer" containerID="448308ad6e34ac922d66b3c90e66d0927407a94dd28633ec350502d69e4ca8aa" Mar 07 15:20:08 crc kubenswrapper[4943]: E0307 15:20:08.768275 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cpdmm_openshift-machine-config-operator(4c1f18ea-e557-41fc-95dd-ce4e1722a61e)\"" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" Mar 07 15:20:09 crc kubenswrapper[4943]: I0307 15:20:09.538303 4943 generic.go:334] "Generic (PLEG): container finished" podID="07ac3ac5-33ba-436c-af70-0e7976486164" containerID="a928e8c9e1805599cfcc0243ab9d559652e0cc68f084c5ad20b90e7af50975ca" exitCode=0 Mar 07 15:20:09 crc kubenswrapper[4943]: I0307 15:20:09.538427 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-bbh8q" event={"ID":"07ac3ac5-33ba-436c-af70-0e7976486164","Type":"ContainerDied","Data":"a928e8c9e1805599cfcc0243ab9d559652e0cc68f084c5ad20b90e7af50975ca"} Mar 07 15:20:10 crc kubenswrapper[4943]: I0307 15:20:10.909878 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-bbh8q" Mar 07 15:20:10 crc kubenswrapper[4943]: I0307 15:20:10.944779 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-bbh8q"] Mar 07 15:20:10 crc kubenswrapper[4943]: I0307 15:20:10.954876 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-bbh8q"] Mar 07 15:20:11 crc kubenswrapper[4943]: I0307 15:20:11.061727 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/07ac3ac5-33ba-436c-af70-0e7976486164-etc-swift\") pod \"07ac3ac5-33ba-436c-af70-0e7976486164\" (UID: \"07ac3ac5-33ba-436c-af70-0e7976486164\") " Mar 07 15:20:11 crc kubenswrapper[4943]: I0307 15:20:11.061832 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ljq9m\" (UniqueName: \"kubernetes.io/projected/07ac3ac5-33ba-436c-af70-0e7976486164-kube-api-access-ljq9m\") pod \"07ac3ac5-33ba-436c-af70-0e7976486164\" (UID: \"07ac3ac5-33ba-436c-af70-0e7976486164\") " Mar 07 15:20:11 crc kubenswrapper[4943]: I0307 15:20:11.061948 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/07ac3ac5-33ba-436c-af70-0e7976486164-scripts\") pod \"07ac3ac5-33ba-436c-af70-0e7976486164\" (UID: \"07ac3ac5-33ba-436c-af70-0e7976486164\") " Mar 07 15:20:11 crc kubenswrapper[4943]: I0307 15:20:11.061994 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/07ac3ac5-33ba-436c-af70-0e7976486164-swiftconf\") pod \"07ac3ac5-33ba-436c-af70-0e7976486164\" (UID: \"07ac3ac5-33ba-436c-af70-0e7976486164\") " Mar 07 15:20:11 crc kubenswrapper[4943]: I0307 15:20:11.062043 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/07ac3ac5-33ba-436c-af70-0e7976486164-ring-data-devices\") pod \"07ac3ac5-33ba-436c-af70-0e7976486164\" (UID: \"07ac3ac5-33ba-436c-af70-0e7976486164\") " Mar 07 15:20:11 crc kubenswrapper[4943]: I0307 15:20:11.062082 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/07ac3ac5-33ba-436c-af70-0e7976486164-dispersionconf\") pod \"07ac3ac5-33ba-436c-af70-0e7976486164\" (UID: \"07ac3ac5-33ba-436c-af70-0e7976486164\") " Mar 07 15:20:11 crc kubenswrapper[4943]: I0307 15:20:11.062745 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07ac3ac5-33ba-436c-af70-0e7976486164-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "07ac3ac5-33ba-436c-af70-0e7976486164" (UID: "07ac3ac5-33ba-436c-af70-0e7976486164"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:20:11 crc kubenswrapper[4943]: I0307 15:20:11.063387 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/07ac3ac5-33ba-436c-af70-0e7976486164-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "07ac3ac5-33ba-436c-af70-0e7976486164" (UID: "07ac3ac5-33ba-436c-af70-0e7976486164"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:20:11 crc kubenswrapper[4943]: I0307 15:20:11.068732 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07ac3ac5-33ba-436c-af70-0e7976486164-kube-api-access-ljq9m" (OuterVolumeSpecName: "kube-api-access-ljq9m") pod "07ac3ac5-33ba-436c-af70-0e7976486164" (UID: "07ac3ac5-33ba-436c-af70-0e7976486164"). InnerVolumeSpecName "kube-api-access-ljq9m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:20:11 crc kubenswrapper[4943]: I0307 15:20:11.087751 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07ac3ac5-33ba-436c-af70-0e7976486164-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "07ac3ac5-33ba-436c-af70-0e7976486164" (UID: "07ac3ac5-33ba-436c-af70-0e7976486164"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:20:11 crc kubenswrapper[4943]: I0307 15:20:11.089455 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07ac3ac5-33ba-436c-af70-0e7976486164-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "07ac3ac5-33ba-436c-af70-0e7976486164" (UID: "07ac3ac5-33ba-436c-af70-0e7976486164"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:20:11 crc kubenswrapper[4943]: I0307 15:20:11.114842 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/07ac3ac5-33ba-436c-af70-0e7976486164-scripts" (OuterVolumeSpecName: "scripts") pod "07ac3ac5-33ba-436c-af70-0e7976486164" (UID: "07ac3ac5-33ba-436c-af70-0e7976486164"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:20:11 crc kubenswrapper[4943]: I0307 15:20:11.163989 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/07ac3ac5-33ba-436c-af70-0e7976486164-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:20:11 crc kubenswrapper[4943]: I0307 15:20:11.164012 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/07ac3ac5-33ba-436c-af70-0e7976486164-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:20:11 crc kubenswrapper[4943]: I0307 15:20:11.164022 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/07ac3ac5-33ba-436c-af70-0e7976486164-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:20:11 crc kubenswrapper[4943]: I0307 15:20:11.164031 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/07ac3ac5-33ba-436c-af70-0e7976486164-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:20:11 crc kubenswrapper[4943]: I0307 15:20:11.164040 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/07ac3ac5-33ba-436c-af70-0e7976486164-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:20:11 crc kubenswrapper[4943]: I0307 15:20:11.164050 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ljq9m\" (UniqueName: \"kubernetes.io/projected/07ac3ac5-33ba-436c-af70-0e7976486164-kube-api-access-ljq9m\") on node \"crc\" DevicePath \"\"" Mar 07 15:20:11 crc kubenswrapper[4943]: I0307 15:20:11.569028 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="93d9bd101d82d865eb49b58060997dbd4e3fb4fda25bbeaa22bef1f3c36a61e1" Mar 07 15:20:11 crc kubenswrapper[4943]: I0307 15:20:11.569091 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-bbh8q" Mar 07 15:20:12 crc kubenswrapper[4943]: I0307 15:20:12.138712 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-bwfsl"] Mar 07 15:20:12 crc kubenswrapper[4943]: E0307 15:20:12.139263 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07ac3ac5-33ba-436c-af70-0e7976486164" containerName="swift-ring-rebalance" Mar 07 15:20:12 crc kubenswrapper[4943]: I0307 15:20:12.139286 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="07ac3ac5-33ba-436c-af70-0e7976486164" containerName="swift-ring-rebalance" Mar 07 15:20:12 crc kubenswrapper[4943]: I0307 15:20:12.139571 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="07ac3ac5-33ba-436c-af70-0e7976486164" containerName="swift-ring-rebalance" Mar 07 15:20:12 crc kubenswrapper[4943]: I0307 15:20:12.140361 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-bwfsl" Mar 07 15:20:12 crc kubenswrapper[4943]: I0307 15:20:12.144767 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 07 15:20:12 crc kubenswrapper[4943]: I0307 15:20:12.144767 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 07 15:20:12 crc kubenswrapper[4943]: I0307 15:20:12.161043 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-bwfsl"] Mar 07 15:20:12 crc kubenswrapper[4943]: I0307 15:20:12.280340 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e4f22006-4983-4395-b42e-432e4e5cbe59-scripts\") pod \"swift-ring-rebalance-debug-bwfsl\" (UID: \"e4f22006-4983-4395-b42e-432e4e5cbe59\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bwfsl" Mar 07 15:20:12 crc kubenswrapper[4943]: I0307 15:20:12.280410 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e4f22006-4983-4395-b42e-432e4e5cbe59-dispersionconf\") pod \"swift-ring-rebalance-debug-bwfsl\" (UID: \"e4f22006-4983-4395-b42e-432e4e5cbe59\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bwfsl" Mar 07 15:20:12 crc kubenswrapper[4943]: I0307 15:20:12.280438 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e4f22006-4983-4395-b42e-432e4e5cbe59-etc-swift\") pod \"swift-ring-rebalance-debug-bwfsl\" (UID: \"e4f22006-4983-4395-b42e-432e4e5cbe59\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bwfsl" Mar 07 15:20:12 crc kubenswrapper[4943]: I0307 15:20:12.280518 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e4f22006-4983-4395-b42e-432e4e5cbe59-swiftconf\") pod \"swift-ring-rebalance-debug-bwfsl\" (UID: \"e4f22006-4983-4395-b42e-432e4e5cbe59\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bwfsl" Mar 07 15:20:12 crc kubenswrapper[4943]: I0307 15:20:12.280545 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e4f22006-4983-4395-b42e-432e4e5cbe59-ring-data-devices\") pod \"swift-ring-rebalance-debug-bwfsl\" (UID: \"e4f22006-4983-4395-b42e-432e4e5cbe59\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bwfsl" Mar 07 15:20:12 crc kubenswrapper[4943]: I0307 15:20:12.280624 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnpw5\" (UniqueName: \"kubernetes.io/projected/e4f22006-4983-4395-b42e-432e4e5cbe59-kube-api-access-wnpw5\") pod \"swift-ring-rebalance-debug-bwfsl\" (UID: \"e4f22006-4983-4395-b42e-432e4e5cbe59\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bwfsl" Mar 07 15:20:12 crc kubenswrapper[4943]: I0307 15:20:12.381695 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e4f22006-4983-4395-b42e-432e4e5cbe59-etc-swift\") pod \"swift-ring-rebalance-debug-bwfsl\" (UID: \"e4f22006-4983-4395-b42e-432e4e5cbe59\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bwfsl" Mar 07 15:20:12 crc kubenswrapper[4943]: I0307 15:20:12.382113 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e4f22006-4983-4395-b42e-432e4e5cbe59-dispersionconf\") pod \"swift-ring-rebalance-debug-bwfsl\" (UID: \"e4f22006-4983-4395-b42e-432e4e5cbe59\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bwfsl" Mar 07 15:20:12 crc kubenswrapper[4943]: I0307 15:20:12.382290 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e4f22006-4983-4395-b42e-432e4e5cbe59-etc-swift\") pod \"swift-ring-rebalance-debug-bwfsl\" (UID: \"e4f22006-4983-4395-b42e-432e4e5cbe59\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bwfsl" Mar 07 15:20:12 crc kubenswrapper[4943]: I0307 15:20:12.382431 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e4f22006-4983-4395-b42e-432e4e5cbe59-swiftconf\") pod \"swift-ring-rebalance-debug-bwfsl\" (UID: \"e4f22006-4983-4395-b42e-432e4e5cbe59\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bwfsl" Mar 07 15:20:12 crc kubenswrapper[4943]: I0307 15:20:12.382585 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e4f22006-4983-4395-b42e-432e4e5cbe59-ring-data-devices\") pod \"swift-ring-rebalance-debug-bwfsl\" (UID: \"e4f22006-4983-4395-b42e-432e4e5cbe59\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bwfsl" Mar 07 15:20:12 crc kubenswrapper[4943]: I0307 15:20:12.382758 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnpw5\" (UniqueName: \"kubernetes.io/projected/e4f22006-4983-4395-b42e-432e4e5cbe59-kube-api-access-wnpw5\") pod \"swift-ring-rebalance-debug-bwfsl\" (UID: \"e4f22006-4983-4395-b42e-432e4e5cbe59\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bwfsl" Mar 07 15:20:12 crc kubenswrapper[4943]: I0307 15:20:12.382990 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e4f22006-4983-4395-b42e-432e4e5cbe59-scripts\") pod \"swift-ring-rebalance-debug-bwfsl\" (UID: \"e4f22006-4983-4395-b42e-432e4e5cbe59\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bwfsl" Mar 07 15:20:12 crc kubenswrapper[4943]: I0307 15:20:12.383710 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e4f22006-4983-4395-b42e-432e4e5cbe59-ring-data-devices\") pod \"swift-ring-rebalance-debug-bwfsl\" (UID: \"e4f22006-4983-4395-b42e-432e4e5cbe59\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bwfsl" Mar 07 15:20:12 crc kubenswrapper[4943]: I0307 15:20:12.383822 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e4f22006-4983-4395-b42e-432e4e5cbe59-scripts\") pod \"swift-ring-rebalance-debug-bwfsl\" (UID: \"e4f22006-4983-4395-b42e-432e4e5cbe59\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bwfsl" Mar 07 15:20:12 crc kubenswrapper[4943]: I0307 15:20:12.389810 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e4f22006-4983-4395-b42e-432e4e5cbe59-swiftconf\") pod \"swift-ring-rebalance-debug-bwfsl\" (UID: \"e4f22006-4983-4395-b42e-432e4e5cbe59\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bwfsl" Mar 07 15:20:12 crc kubenswrapper[4943]: I0307 15:20:12.391738 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e4f22006-4983-4395-b42e-432e4e5cbe59-dispersionconf\") pod \"swift-ring-rebalance-debug-bwfsl\" (UID: \"e4f22006-4983-4395-b42e-432e4e5cbe59\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bwfsl" Mar 07 15:20:12 crc kubenswrapper[4943]: I0307 15:20:12.407313 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnpw5\" (UniqueName: \"kubernetes.io/projected/e4f22006-4983-4395-b42e-432e4e5cbe59-kube-api-access-wnpw5\") pod \"swift-ring-rebalance-debug-bwfsl\" (UID: \"e4f22006-4983-4395-b42e-432e4e5cbe59\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bwfsl" Mar 07 15:20:12 crc kubenswrapper[4943]: I0307 15:20:12.475550 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-bwfsl" Mar 07 15:20:12 crc kubenswrapper[4943]: I0307 15:20:12.765061 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07ac3ac5-33ba-436c-af70-0e7976486164" path="/var/lib/kubelet/pods/07ac3ac5-33ba-436c-af70-0e7976486164/volumes" Mar 07 15:20:12 crc kubenswrapper[4943]: I0307 15:20:12.939076 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-bwfsl"] Mar 07 15:20:13 crc kubenswrapper[4943]: I0307 15:20:13.595060 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-bwfsl" event={"ID":"e4f22006-4983-4395-b42e-432e4e5cbe59","Type":"ContainerStarted","Data":"a87fab2fc350e01e1c8b6cb2c885db9037b0892ac597bcda669e2b0df98b51eb"} Mar 07 15:20:13 crc kubenswrapper[4943]: I0307 15:20:13.595158 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-bwfsl" event={"ID":"e4f22006-4983-4395-b42e-432e4e5cbe59","Type":"ContainerStarted","Data":"79510d4a00c9a86a19172438fd5133bd3455a070604faf9fc85dcfb60c925123"} Mar 07 15:20:13 crc kubenswrapper[4943]: I0307 15:20:13.624661 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-bwfsl" podStartSLOduration=1.6246355540000001 podStartE2EDuration="1.624635554s" podCreationTimestamp="2026-03-07 15:20:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 15:20:13.614773953 +0000 UTC m=+2455.566910471" watchObservedRunningTime="2026-03-07 15:20:13.624635554 +0000 UTC m=+2455.576772082" Mar 07 15:20:14 crc kubenswrapper[4943]: I0307 15:20:14.610502 4943 generic.go:334] "Generic (PLEG): container finished" podID="e4f22006-4983-4395-b42e-432e4e5cbe59" containerID="a87fab2fc350e01e1c8b6cb2c885db9037b0892ac597bcda669e2b0df98b51eb" exitCode=0 Mar 07 15:20:14 crc kubenswrapper[4943]: I0307 15:20:14.610620 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-bwfsl" event={"ID":"e4f22006-4983-4395-b42e-432e4e5cbe59","Type":"ContainerDied","Data":"a87fab2fc350e01e1c8b6cb2c885db9037b0892ac597bcda669e2b0df98b51eb"} Mar 07 15:20:15 crc kubenswrapper[4943]: I0307 15:20:15.934559 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-bwfsl" Mar 07 15:20:15 crc kubenswrapper[4943]: I0307 15:20:15.966598 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-bwfsl"] Mar 07 15:20:15 crc kubenswrapper[4943]: I0307 15:20:15.973799 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-bwfsl"] Mar 07 15:20:16 crc kubenswrapper[4943]: I0307 15:20:16.053839 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e4f22006-4983-4395-b42e-432e4e5cbe59-dispersionconf\") pod \"e4f22006-4983-4395-b42e-432e4e5cbe59\" (UID: \"e4f22006-4983-4395-b42e-432e4e5cbe59\") " Mar 07 15:20:16 crc kubenswrapper[4943]: I0307 15:20:16.053890 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e4f22006-4983-4395-b42e-432e4e5cbe59-swiftconf\") pod \"e4f22006-4983-4395-b42e-432e4e5cbe59\" (UID: \"e4f22006-4983-4395-b42e-432e4e5cbe59\") " Mar 07 15:20:16 crc kubenswrapper[4943]: I0307 15:20:16.054021 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e4f22006-4983-4395-b42e-432e4e5cbe59-ring-data-devices\") pod \"e4f22006-4983-4395-b42e-432e4e5cbe59\" (UID: \"e4f22006-4983-4395-b42e-432e4e5cbe59\") " Mar 07 15:20:16 crc kubenswrapper[4943]: I0307 15:20:16.054060 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wnpw5\" (UniqueName: \"kubernetes.io/projected/e4f22006-4983-4395-b42e-432e4e5cbe59-kube-api-access-wnpw5\") pod \"e4f22006-4983-4395-b42e-432e4e5cbe59\" (UID: \"e4f22006-4983-4395-b42e-432e4e5cbe59\") " Mar 07 15:20:16 crc kubenswrapper[4943]: I0307 15:20:16.054087 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e4f22006-4983-4395-b42e-432e4e5cbe59-scripts\") pod \"e4f22006-4983-4395-b42e-432e4e5cbe59\" (UID: \"e4f22006-4983-4395-b42e-432e4e5cbe59\") " Mar 07 15:20:16 crc kubenswrapper[4943]: I0307 15:20:16.054195 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e4f22006-4983-4395-b42e-432e4e5cbe59-etc-swift\") pod \"e4f22006-4983-4395-b42e-432e4e5cbe59\" (UID: \"e4f22006-4983-4395-b42e-432e4e5cbe59\") " Mar 07 15:20:16 crc kubenswrapper[4943]: I0307 15:20:16.054567 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4f22006-4983-4395-b42e-432e4e5cbe59-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "e4f22006-4983-4395-b42e-432e4e5cbe59" (UID: "e4f22006-4983-4395-b42e-432e4e5cbe59"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:20:16 crc kubenswrapper[4943]: I0307 15:20:16.055231 4943 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e4f22006-4983-4395-b42e-432e4e5cbe59-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 07 15:20:16 crc kubenswrapper[4943]: I0307 15:20:16.055229 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e4f22006-4983-4395-b42e-432e4e5cbe59-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "e4f22006-4983-4395-b42e-432e4e5cbe59" (UID: "e4f22006-4983-4395-b42e-432e4e5cbe59"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:20:16 crc kubenswrapper[4943]: I0307 15:20:16.060483 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4f22006-4983-4395-b42e-432e4e5cbe59-kube-api-access-wnpw5" (OuterVolumeSpecName: "kube-api-access-wnpw5") pod "e4f22006-4983-4395-b42e-432e4e5cbe59" (UID: "e4f22006-4983-4395-b42e-432e4e5cbe59"). InnerVolumeSpecName "kube-api-access-wnpw5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:20:16 crc kubenswrapper[4943]: I0307 15:20:16.080092 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4f22006-4983-4395-b42e-432e4e5cbe59-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "e4f22006-4983-4395-b42e-432e4e5cbe59" (UID: "e4f22006-4983-4395-b42e-432e4e5cbe59"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:20:16 crc kubenswrapper[4943]: I0307 15:20:16.090438 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4f22006-4983-4395-b42e-432e4e5cbe59-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "e4f22006-4983-4395-b42e-432e4e5cbe59" (UID: "e4f22006-4983-4395-b42e-432e4e5cbe59"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 07 15:20:16 crc kubenswrapper[4943]: I0307 15:20:16.092497 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4f22006-4983-4395-b42e-432e4e5cbe59-scripts" (OuterVolumeSpecName: "scripts") pod "e4f22006-4983-4395-b42e-432e4e5cbe59" (UID: "e4f22006-4983-4395-b42e-432e4e5cbe59"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 07 15:20:16 crc kubenswrapper[4943]: I0307 15:20:16.156772 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wnpw5\" (UniqueName: \"kubernetes.io/projected/e4f22006-4983-4395-b42e-432e4e5cbe59-kube-api-access-wnpw5\") on node \"crc\" DevicePath \"\"" Mar 07 15:20:16 crc kubenswrapper[4943]: I0307 15:20:16.156824 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e4f22006-4983-4395-b42e-432e4e5cbe59-scripts\") on node \"crc\" DevicePath \"\"" Mar 07 15:20:16 crc kubenswrapper[4943]: I0307 15:20:16.156862 4943 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e4f22006-4983-4395-b42e-432e4e5cbe59-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 07 15:20:16 crc kubenswrapper[4943]: I0307 15:20:16.156881 4943 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e4f22006-4983-4395-b42e-432e4e5cbe59-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:20:16 crc kubenswrapper[4943]: I0307 15:20:16.156896 4943 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e4f22006-4983-4395-b42e-432e4e5cbe59-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 07 15:20:16 crc kubenswrapper[4943]: I0307 15:20:16.642381 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="79510d4a00c9a86a19172438fd5133bd3455a070604faf9fc85dcfb60c925123" Mar 07 15:20:16 crc kubenswrapper[4943]: I0307 15:20:16.642451 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-bwfsl" Mar 07 15:20:16 crc kubenswrapper[4943]: I0307 15:20:16.769711 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4f22006-4983-4395-b42e-432e4e5cbe59" path="/var/lib/kubelet/pods/e4f22006-4983-4395-b42e-432e4e5cbe59/volumes" Mar 07 15:20:20 crc kubenswrapper[4943]: I0307 15:20:20.755737 4943 scope.go:117] "RemoveContainer" containerID="448308ad6e34ac922d66b3c90e66d0927407a94dd28633ec350502d69e4ca8aa" Mar 07 15:20:20 crc kubenswrapper[4943]: E0307 15:20:20.756626 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cpdmm_openshift-machine-config-operator(4c1f18ea-e557-41fc-95dd-ce4e1722a61e)\"" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" Mar 07 15:20:24 crc kubenswrapper[4943]: I0307 15:20:24.192378 4943 scope.go:117] "RemoveContainer" containerID="f5f943654c0799359019ede37edc2fc94b884ba56d5f2f4d65ae0f659a310e73" Mar 07 15:20:24 crc kubenswrapper[4943]: I0307 15:20:24.260918 4943 scope.go:117] "RemoveContainer" containerID="4c521497d06b0b4870529d7fe6ba7b461ebfa9e691dd8007da6a3c477d54a0b4" Mar 07 15:20:24 crc kubenswrapper[4943]: I0307 15:20:24.301813 4943 scope.go:117] "RemoveContainer" containerID="b301320a41d48cca02200a810c67ad9d75cc4daaa821bf5cd1ad024d1933c225" Mar 07 15:20:24 crc kubenswrapper[4943]: I0307 15:20:24.349422 4943 scope.go:117] "RemoveContainer" containerID="15062d0afbc2b0fce1e0d349ad9933e0354c4de994c64b443f3dc6f13774de2e" Mar 07 15:20:24 crc kubenswrapper[4943]: I0307 15:20:24.387648 4943 scope.go:117] "RemoveContainer" containerID="2800f30d0d51f86586364cf6aa464f7fe6d5bca06dd80bd35c8d82cccb70b9a3" Mar 07 15:20:24 crc kubenswrapper[4943]: I0307 15:20:24.431393 4943 scope.go:117] "RemoveContainer" containerID="19aeee34f55b8a67a45502efa5df20f87d64a97e23eb243996a57ad16dda0bb1" Mar 07 15:20:24 crc kubenswrapper[4943]: I0307 15:20:24.469064 4943 scope.go:117] "RemoveContainer" containerID="7091e809442e2336edbc5d300cf4a15937fc81d3519536fbf5a5d0cbbc5bb824" Mar 07 15:20:24 crc kubenswrapper[4943]: I0307 15:20:24.512317 4943 scope.go:117] "RemoveContainer" containerID="5e35fd56c2804ea758b6d8b42abb3f040da0696fcaec51f9e79d144dacd0b9c9" Mar 07 15:20:33 crc kubenswrapper[4943]: I0307 15:20:33.755314 4943 scope.go:117] "RemoveContainer" containerID="448308ad6e34ac922d66b3c90e66d0927407a94dd28633ec350502d69e4ca8aa" Mar 07 15:20:33 crc kubenswrapper[4943]: E0307 15:20:33.756531 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cpdmm_openshift-machine-config-operator(4c1f18ea-e557-41fc-95dd-ce4e1722a61e)\"" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" Mar 07 15:20:45 crc kubenswrapper[4943]: I0307 15:20:45.755920 4943 scope.go:117] "RemoveContainer" containerID="448308ad6e34ac922d66b3c90e66d0927407a94dd28633ec350502d69e4ca8aa" Mar 07 15:20:45 crc kubenswrapper[4943]: E0307 15:20:45.758591 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cpdmm_openshift-machine-config-operator(4c1f18ea-e557-41fc-95dd-ce4e1722a61e)\"" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" Mar 07 15:20:48 crc kubenswrapper[4943]: I0307 15:20:48.060839 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-lxzm4/must-gather-lrs4f"] Mar 07 15:20:48 crc kubenswrapper[4943]: E0307 15:20:48.061501 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4f22006-4983-4395-b42e-432e4e5cbe59" containerName="swift-ring-rebalance" Mar 07 15:20:48 crc kubenswrapper[4943]: I0307 15:20:48.061517 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4f22006-4983-4395-b42e-432e4e5cbe59" containerName="swift-ring-rebalance" Mar 07 15:20:48 crc kubenswrapper[4943]: I0307 15:20:48.061765 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4f22006-4983-4395-b42e-432e4e5cbe59" containerName="swift-ring-rebalance" Mar 07 15:20:48 crc kubenswrapper[4943]: I0307 15:20:48.062698 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lxzm4/must-gather-lrs4f" Mar 07 15:20:48 crc kubenswrapper[4943]: I0307 15:20:48.065516 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-lxzm4"/"openshift-service-ca.crt" Mar 07 15:20:48 crc kubenswrapper[4943]: I0307 15:20:48.065720 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-lxzm4"/"kube-root-ca.crt" Mar 07 15:20:48 crc kubenswrapper[4943]: I0307 15:20:48.074225 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-lxzm4/must-gather-lrs4f"] Mar 07 15:20:48 crc kubenswrapper[4943]: I0307 15:20:48.199362 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gllrc\" (UniqueName: \"kubernetes.io/projected/112000fc-ea09-47ab-b27b-d70c6f37a8a0-kube-api-access-gllrc\") pod \"must-gather-lrs4f\" (UID: \"112000fc-ea09-47ab-b27b-d70c6f37a8a0\") " pod="openshift-must-gather-lxzm4/must-gather-lrs4f" Mar 07 15:20:48 crc kubenswrapper[4943]: I0307 15:20:48.199414 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/112000fc-ea09-47ab-b27b-d70c6f37a8a0-must-gather-output\") pod \"must-gather-lrs4f\" (UID: \"112000fc-ea09-47ab-b27b-d70c6f37a8a0\") " pod="openshift-must-gather-lxzm4/must-gather-lrs4f" Mar 07 15:20:48 crc kubenswrapper[4943]: I0307 15:20:48.300617 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gllrc\" (UniqueName: \"kubernetes.io/projected/112000fc-ea09-47ab-b27b-d70c6f37a8a0-kube-api-access-gllrc\") pod \"must-gather-lrs4f\" (UID: \"112000fc-ea09-47ab-b27b-d70c6f37a8a0\") " pod="openshift-must-gather-lxzm4/must-gather-lrs4f" Mar 07 15:20:48 crc kubenswrapper[4943]: I0307 15:20:48.300677 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/112000fc-ea09-47ab-b27b-d70c6f37a8a0-must-gather-output\") pod \"must-gather-lrs4f\" (UID: \"112000fc-ea09-47ab-b27b-d70c6f37a8a0\") " pod="openshift-must-gather-lxzm4/must-gather-lrs4f" Mar 07 15:20:48 crc kubenswrapper[4943]: I0307 15:20:48.301437 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/112000fc-ea09-47ab-b27b-d70c6f37a8a0-must-gather-output\") pod \"must-gather-lrs4f\" (UID: \"112000fc-ea09-47ab-b27b-d70c6f37a8a0\") " pod="openshift-must-gather-lxzm4/must-gather-lrs4f" Mar 07 15:20:48 crc kubenswrapper[4943]: I0307 15:20:48.320273 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gllrc\" (UniqueName: \"kubernetes.io/projected/112000fc-ea09-47ab-b27b-d70c6f37a8a0-kube-api-access-gllrc\") pod \"must-gather-lrs4f\" (UID: \"112000fc-ea09-47ab-b27b-d70c6f37a8a0\") " pod="openshift-must-gather-lxzm4/must-gather-lrs4f" Mar 07 15:20:48 crc kubenswrapper[4943]: I0307 15:20:48.381374 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lxzm4/must-gather-lrs4f" Mar 07 15:20:48 crc kubenswrapper[4943]: I0307 15:20:48.819918 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-lxzm4/must-gather-lrs4f"] Mar 07 15:20:49 crc kubenswrapper[4943]: I0307 15:20:49.071679 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lxzm4/must-gather-lrs4f" event={"ID":"112000fc-ea09-47ab-b27b-d70c6f37a8a0","Type":"ContainerStarted","Data":"854cc019e38f40c28d3f3c4eabfd1c5fdd3cf4e5367729e020be32606f87d0f2"} Mar 07 15:20:55 crc kubenswrapper[4943]: I0307 15:20:55.126394 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lxzm4/must-gather-lrs4f" event={"ID":"112000fc-ea09-47ab-b27b-d70c6f37a8a0","Type":"ContainerStarted","Data":"f0b490bf5811307fe9ab134a8894c055a93b3e8cdb5783c6012ee854501f994c"} Mar 07 15:20:55 crc kubenswrapper[4943]: I0307 15:20:55.127261 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lxzm4/must-gather-lrs4f" event={"ID":"112000fc-ea09-47ab-b27b-d70c6f37a8a0","Type":"ContainerStarted","Data":"ba5ca406c42a6e37bb34abce909357f13ae1eaf861c725d47e64e7f968a77817"} Mar 07 15:20:55 crc kubenswrapper[4943]: I0307 15:20:55.159247 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-lxzm4/must-gather-lrs4f" podStartSLOduration=1.626117729 podStartE2EDuration="7.159218143s" podCreationTimestamp="2026-03-07 15:20:48 +0000 UTC" firstStartedPulling="2026-03-07 15:20:48.811887913 +0000 UTC m=+2490.764024411" lastFinishedPulling="2026-03-07 15:20:54.344988277 +0000 UTC m=+2496.297124825" observedRunningTime="2026-03-07 15:20:55.14766428 +0000 UTC m=+2497.099800848" watchObservedRunningTime="2026-03-07 15:20:55.159218143 +0000 UTC m=+2497.111354681" Mar 07 15:20:58 crc kubenswrapper[4943]: I0307 15:20:58.759550 4943 scope.go:117] "RemoveContainer" containerID="448308ad6e34ac922d66b3c90e66d0927407a94dd28633ec350502d69e4ca8aa" Mar 07 15:20:58 crc kubenswrapper[4943]: E0307 15:20:58.760276 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cpdmm_openshift-machine-config-operator(4c1f18ea-e557-41fc-95dd-ce4e1722a61e)\"" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" Mar 07 15:21:13 crc kubenswrapper[4943]: I0307 15:21:13.755418 4943 scope.go:117] "RemoveContainer" containerID="448308ad6e34ac922d66b3c90e66d0927407a94dd28633ec350502d69e4ca8aa" Mar 07 15:21:13 crc kubenswrapper[4943]: E0307 15:21:13.756148 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cpdmm_openshift-machine-config-operator(4c1f18ea-e557-41fc-95dd-ce4e1722a61e)\"" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" Mar 07 15:21:24 crc kubenswrapper[4943]: I0307 15:21:24.841796 4943 scope.go:117] "RemoveContainer" containerID="4a27c160db96ac0508bd3c9253afe31401289535ce5e53a860e3771b2587e6da" Mar 07 15:21:24 crc kubenswrapper[4943]: I0307 15:21:24.921774 4943 scope.go:117] "RemoveContainer" containerID="81ad0427c89f20cf442bd3e43a1c4ec1ee825f288babfc2bff62a6c5418ed361" Mar 07 15:21:24 crc kubenswrapper[4943]: I0307 15:21:24.961561 4943 scope.go:117] "RemoveContainer" containerID="c4c7288d2582ac16fdf710eca40a5eccdaa8f97875f3ce161802f9ef99ca3ed9" Mar 07 15:21:24 crc kubenswrapper[4943]: I0307 15:21:24.995953 4943 scope.go:117] "RemoveContainer" containerID="6b42dfcd362de484f87d1214e9c529ee4264aa022e5f08f253e6268435f5a2de" Mar 07 15:21:25 crc kubenswrapper[4943]: I0307 15:21:25.042592 4943 scope.go:117] "RemoveContainer" containerID="e5bc4c5fb3c660735c7885799282c09732ec1f0a5848df8f03dbe19751851494" Mar 07 15:21:25 crc kubenswrapper[4943]: I0307 15:21:25.090670 4943 scope.go:117] "RemoveContainer" containerID="f4752e484e0ccbe4760ca08e61be7c10dc834d665b03dd98b780ea21f8da2243" Mar 07 15:21:25 crc kubenswrapper[4943]: I0307 15:21:25.122124 4943 scope.go:117] "RemoveContainer" containerID="a0eaca674ddb1faa49197ac8b9583a9de6ff525edaa8faa0e36b3a34246dd60d" Mar 07 15:21:25 crc kubenswrapper[4943]: I0307 15:21:25.142219 4943 scope.go:117] "RemoveContainer" containerID="75b39e0895817d203e2ba4b4ad29a861fd66d7b554f98549272c08813e57cc6f" Mar 07 15:21:25 crc kubenswrapper[4943]: I0307 15:21:25.166287 4943 scope.go:117] "RemoveContainer" containerID="79267f52fc465e59ae9257b9280f1da6610e63b4134f932f0f536fa3f529ec0d" Mar 07 15:21:26 crc kubenswrapper[4943]: I0307 15:21:26.756744 4943 scope.go:117] "RemoveContainer" containerID="448308ad6e34ac922d66b3c90e66d0927407a94dd28633ec350502d69e4ca8aa" Mar 07 15:21:26 crc kubenswrapper[4943]: E0307 15:21:26.757205 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cpdmm_openshift-machine-config-operator(4c1f18ea-e557-41fc-95dd-ce4e1722a61e)\"" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" Mar 07 15:21:38 crc kubenswrapper[4943]: I0307 15:21:38.763864 4943 scope.go:117] "RemoveContainer" containerID="448308ad6e34ac922d66b3c90e66d0927407a94dd28633ec350502d69e4ca8aa" Mar 07 15:21:39 crc kubenswrapper[4943]: I0307 15:21:39.458085 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" event={"ID":"4c1f18ea-e557-41fc-95dd-ce4e1722a61e","Type":"ContainerStarted","Data":"600a051b71dd164e5cc8344d91e9c5c354cdc3c5514cd0b1a1a3e3062c02fe1d"} Mar 07 15:21:41 crc kubenswrapper[4943]: I0307 15:21:41.213703 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar4jd2_3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1/util/0.log" Mar 07 15:21:41 crc kubenswrapper[4943]: I0307 15:21:41.354746 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar4jd2_3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1/util/0.log" Mar 07 15:21:41 crc kubenswrapper[4943]: I0307 15:21:41.362500 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar4jd2_3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1/pull/0.log" Mar 07 15:21:41 crc kubenswrapper[4943]: I0307 15:21:41.396398 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar4jd2_3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1/pull/0.log" Mar 07 15:21:41 crc kubenswrapper[4943]: I0307 15:21:41.556035 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar4jd2_3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1/util/0.log" Mar 07 15:21:41 crc kubenswrapper[4943]: I0307 15:21:41.579169 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar4jd2_3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1/pull/0.log" Mar 07 15:21:41 crc kubenswrapper[4943]: I0307 15:21:41.627883 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar4jd2_3bcdb55d-458b-4bb1-b9d2-fe1d98fc75e1/extract/0.log" Mar 07 15:21:41 crc kubenswrapper[4943]: I0307 15:21:41.724039 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8e67c57d4916ae7e348662cb38dd5155bd37a096e90f723dff4c6ec243ptqwj_7f65812b-f0ff-4e87-845d-eafb6cc93743/util/0.log" Mar 07 15:21:41 crc kubenswrapper[4943]: I0307 15:21:41.877353 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8e67c57d4916ae7e348662cb38dd5155bd37a096e90f723dff4c6ec243ptqwj_7f65812b-f0ff-4e87-845d-eafb6cc93743/util/0.log" Mar 07 15:21:41 crc kubenswrapper[4943]: I0307 15:21:41.925231 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8e67c57d4916ae7e348662cb38dd5155bd37a096e90f723dff4c6ec243ptqwj_7f65812b-f0ff-4e87-845d-eafb6cc93743/pull/0.log" Mar 07 15:21:41 crc kubenswrapper[4943]: I0307 15:21:41.937828 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8e67c57d4916ae7e348662cb38dd5155bd37a096e90f723dff4c6ec243ptqwj_7f65812b-f0ff-4e87-845d-eafb6cc93743/pull/0.log" Mar 07 15:21:42 crc kubenswrapper[4943]: I0307 15:21:42.088005 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8e67c57d4916ae7e348662cb38dd5155bd37a096e90f723dff4c6ec243ptqwj_7f65812b-f0ff-4e87-845d-eafb6cc93743/pull/0.log" Mar 07 15:21:42 crc kubenswrapper[4943]: I0307 15:21:42.107941 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8e67c57d4916ae7e348662cb38dd5155bd37a096e90f723dff4c6ec243ptqwj_7f65812b-f0ff-4e87-845d-eafb6cc93743/extract/0.log" Mar 07 15:21:42 crc kubenswrapper[4943]: I0307 15:21:42.121095 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8e67c57d4916ae7e348662cb38dd5155bd37a096e90f723dff4c6ec243ptqwj_7f65812b-f0ff-4e87-845d-eafb6cc93743/util/0.log" Mar 07 15:21:42 crc kubenswrapper[4943]: I0307 15:21:42.259458 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590mkxwv_e6788dcd-2850-4113-94e7-204d8028d32a/util/0.log" Mar 07 15:21:42 crc kubenswrapper[4943]: I0307 15:21:42.409564 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590mkxwv_e6788dcd-2850-4113-94e7-204d8028d32a/util/0.log" Mar 07 15:21:42 crc kubenswrapper[4943]: I0307 15:21:42.419727 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590mkxwv_e6788dcd-2850-4113-94e7-204d8028d32a/pull/0.log" Mar 07 15:21:42 crc kubenswrapper[4943]: I0307 15:21:42.441198 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590mkxwv_e6788dcd-2850-4113-94e7-204d8028d32a/pull/0.log" Mar 07 15:21:42 crc kubenswrapper[4943]: I0307 15:21:42.626413 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590mkxwv_e6788dcd-2850-4113-94e7-204d8028d32a/util/0.log" Mar 07 15:21:42 crc kubenswrapper[4943]: I0307 15:21:42.651448 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590mkxwv_e6788dcd-2850-4113-94e7-204d8028d32a/extract/0.log" Mar 07 15:21:42 crc kubenswrapper[4943]: I0307 15:21:42.652901 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590mkxwv_e6788dcd-2850-4113-94e7-204d8028d32a/pull/0.log" Mar 07 15:21:42 crc kubenswrapper[4943]: I0307 15:21:42.935174 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40frz9l_fb540778-9fd9-48c8-b3f8-45327871b647/util/0.log" Mar 07 15:21:43 crc kubenswrapper[4943]: I0307 15:21:43.025239 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40frz9l_fb540778-9fd9-48c8-b3f8-45327871b647/pull/0.log" Mar 07 15:21:43 crc kubenswrapper[4943]: I0307 15:21:43.044267 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40frz9l_fb540778-9fd9-48c8-b3f8-45327871b647/util/0.log" Mar 07 15:21:43 crc kubenswrapper[4943]: I0307 15:21:43.047653 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40frz9l_fb540778-9fd9-48c8-b3f8-45327871b647/pull/0.log" Mar 07 15:21:43 crc kubenswrapper[4943]: I0307 15:21:43.247213 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40frz9l_fb540778-9fd9-48c8-b3f8-45327871b647/util/0.log" Mar 07 15:21:43 crc kubenswrapper[4943]: I0307 15:21:43.249523 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40frz9l_fb540778-9fd9-48c8-b3f8-45327871b647/pull/0.log" Mar 07 15:21:43 crc kubenswrapper[4943]: I0307 15:21:43.284246 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40frz9l_fb540778-9fd9-48c8-b3f8-45327871b647/extract/0.log" Mar 07 15:21:43 crc kubenswrapper[4943]: I0307 15:21:43.416075 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da67slpc_a880ad69-0f85-45ec-a5f2-e27ce2ce850f/util/0.log" Mar 07 15:21:43 crc kubenswrapper[4943]: I0307 15:21:43.608718 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da67slpc_a880ad69-0f85-45ec-a5f2-e27ce2ce850f/pull/0.log" Mar 07 15:21:43 crc kubenswrapper[4943]: I0307 15:21:43.617584 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da67slpc_a880ad69-0f85-45ec-a5f2-e27ce2ce850f/util/0.log" Mar 07 15:21:43 crc kubenswrapper[4943]: I0307 15:21:43.642388 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da67slpc_a880ad69-0f85-45ec-a5f2-e27ce2ce850f/pull/0.log" Mar 07 15:21:43 crc kubenswrapper[4943]: I0307 15:21:43.804364 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da67slpc_a880ad69-0f85-45ec-a5f2-e27ce2ce850f/extract/0.log" Mar 07 15:21:43 crc kubenswrapper[4943]: I0307 15:21:43.850146 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da67slpc_a880ad69-0f85-45ec-a5f2-e27ce2ce850f/util/0.log" Mar 07 15:21:43 crc kubenswrapper[4943]: I0307 15:21:43.857707 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da67slpc_a880ad69-0f85-45ec-a5f2-e27ce2ce850f/pull/0.log" Mar 07 15:21:44 crc kubenswrapper[4943]: I0307 15:21:44.069109 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-index-z6wwq_08b0d828-0a74-4630-bffb-815ec50fda48/registry-server/0.log" Mar 07 15:21:44 crc kubenswrapper[4943]: I0307 15:21:44.266104 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8cwzdw8_487d337d-5286-4425-b303-e871dcec36e8/util/0.log" Mar 07 15:21:44 crc kubenswrapper[4943]: I0307 15:21:44.503237 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8cwzdw8_487d337d-5286-4425-b303-e871dcec36e8/util/0.log" Mar 07 15:21:44 crc kubenswrapper[4943]: I0307 15:21:44.508173 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8cwzdw8_487d337d-5286-4425-b303-e871dcec36e8/pull/0.log" Mar 07 15:21:44 crc kubenswrapper[4943]: I0307 15:21:44.508251 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8cwzdw8_487d337d-5286-4425-b303-e871dcec36e8/pull/0.log" Mar 07 15:21:44 crc kubenswrapper[4943]: I0307 15:21:44.753042 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8cwzdw8_487d337d-5286-4425-b303-e871dcec36e8/util/0.log" Mar 07 15:21:44 crc kubenswrapper[4943]: I0307 15:21:44.773541 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8cwzdw8_487d337d-5286-4425-b303-e871dcec36e8/extract/0.log" Mar 07 15:21:44 crc kubenswrapper[4943]: I0307 15:21:44.774160 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8cwzdw8_487d337d-5286-4425-b303-e871dcec36e8/pull/0.log" Mar 07 15:21:44 crc kubenswrapper[4943]: I0307 15:21:44.951981 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-6685f4bd5b-n5ksp_e9e5a357-affd-43b8-8463-214c981786fa/manager/0.log" Mar 07 15:21:45 crc kubenswrapper[4943]: I0307 15:21:45.036407 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-index-qzp5d_708bf537-729f-4ce7-8a31-ffd3f252cf43/registry-server/0.log" Mar 07 15:21:45 crc kubenswrapper[4943]: I0307 15:21:45.180569 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7d8dd5b6b9-q7chd_83c1e172-0efd-48eb-bfe3-b35302e82114/manager/0.log" Mar 07 15:21:45 crc kubenswrapper[4943]: I0307 15:21:45.292466 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-index-cnq8m_690e82f5-4a59-4926-95b5-b13362cb1a97/registry-server/0.log" Mar 07 15:21:45 crc kubenswrapper[4943]: I0307 15:21:45.415241 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-749f4cd94-8k4nn_9d6f004b-a409-4960-b377-8e8577597b99/manager/0.log" Mar 07 15:21:45 crc kubenswrapper[4943]: I0307 15:21:45.483633 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-index-qdw88_6562db67-0f59-4eab-876f-a60272522e0d/registry-server/0.log" Mar 07 15:21:45 crc kubenswrapper[4943]: I0307 15:21:45.621865 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-779fc9694b-42t4c_0b3e1865-67b8-4e70-ae06-e0c41256e303/operator/0.log" Mar 07 15:21:45 crc kubenswrapper[4943]: I0307 15:21:45.698066 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-index-jcj6r_541ac54d-ab66-40ac-a908-e38fc52e98cc/registry-server/0.log" Mar 07 15:21:45 crc kubenswrapper[4943]: I0307 15:21:45.885892 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9c8b7c9f6-wcwvx_125fcec4-ea6c-4199-b0cd-68bcd04c4580/manager/0.log" Mar 07 15:21:45 crc kubenswrapper[4943]: I0307 15:21:45.939887 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-index-zltk7_89b40fec-692e-44d3-9290-66c7a2931e40/registry-server/0.log" Mar 07 15:21:46 crc kubenswrapper[4943]: I0307 15:21:46.627129 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-675b7b5b56-95zhm_ebf90d0c-1550-42d8-96b6-695f161d5ace/manager/0.log" Mar 07 15:22:00 crc kubenswrapper[4943]: I0307 15:22:00.142249 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29548282-tblqn"] Mar 07 15:22:00 crc kubenswrapper[4943]: I0307 15:22:00.144354 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548282-tblqn" Mar 07 15:22:00 crc kubenswrapper[4943]: I0307 15:22:00.147760 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29548282-tblqn"] Mar 07 15:22:00 crc kubenswrapper[4943]: I0307 15:22:00.155890 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-qqlk5" Mar 07 15:22:00 crc kubenswrapper[4943]: I0307 15:22:00.156107 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 07 15:22:00 crc kubenswrapper[4943]: I0307 15:22:00.156235 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 07 15:22:00 crc kubenswrapper[4943]: I0307 15:22:00.250050 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvmv9\" (UniqueName: \"kubernetes.io/projected/43076e34-f66f-4bba-89dc-bc312cd547df-kube-api-access-qvmv9\") pod \"auto-csr-approver-29548282-tblqn\" (UID: \"43076e34-f66f-4bba-89dc-bc312cd547df\") " pod="openshift-infra/auto-csr-approver-29548282-tblqn" Mar 07 15:22:00 crc kubenswrapper[4943]: I0307 15:22:00.351539 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvmv9\" (UniqueName: \"kubernetes.io/projected/43076e34-f66f-4bba-89dc-bc312cd547df-kube-api-access-qvmv9\") pod \"auto-csr-approver-29548282-tblqn\" (UID: \"43076e34-f66f-4bba-89dc-bc312cd547df\") " pod="openshift-infra/auto-csr-approver-29548282-tblqn" Mar 07 15:22:00 crc kubenswrapper[4943]: I0307 15:22:00.379496 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvmv9\" (UniqueName: \"kubernetes.io/projected/43076e34-f66f-4bba-89dc-bc312cd547df-kube-api-access-qvmv9\") pod \"auto-csr-approver-29548282-tblqn\" (UID: \"43076e34-f66f-4bba-89dc-bc312cd547df\") " pod="openshift-infra/auto-csr-approver-29548282-tblqn" Mar 07 15:22:00 crc kubenswrapper[4943]: I0307 15:22:00.469492 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548282-tblqn" Mar 07 15:22:00 crc kubenswrapper[4943]: I0307 15:22:00.953470 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29548282-tblqn"] Mar 07 15:22:01 crc kubenswrapper[4943]: I0307 15:22:01.662827 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548282-tblqn" event={"ID":"43076e34-f66f-4bba-89dc-bc312cd547df","Type":"ContainerStarted","Data":"ea98caf63ededa8cdfd05756fdb35a757788a5650a8d7d0799e80aa3713e4c2c"} Mar 07 15:22:02 crc kubenswrapper[4943]: I0307 15:22:02.672027 4943 generic.go:334] "Generic (PLEG): container finished" podID="43076e34-f66f-4bba-89dc-bc312cd547df" containerID="630d9d3d51be97a913b4b7905769af6698dc4d1318961cefe8e43b6915becad6" exitCode=0 Mar 07 15:22:02 crc kubenswrapper[4943]: I0307 15:22:02.672085 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548282-tblqn" event={"ID":"43076e34-f66f-4bba-89dc-bc312cd547df","Type":"ContainerDied","Data":"630d9d3d51be97a913b4b7905769af6698dc4d1318961cefe8e43b6915becad6"} Mar 07 15:22:03 crc kubenswrapper[4943]: I0307 15:22:03.050509 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-pv7b4_fdcfa33a-5367-46db-8270-662184116079/control-plane-machine-set-operator/0.log" Mar 07 15:22:03 crc kubenswrapper[4943]: I0307 15:22:03.379870 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-bk5xc_c400fdf2-390c-4c48-8c3e-4854d2e0b58f/machine-api-operator/0.log" Mar 07 15:22:03 crc kubenswrapper[4943]: I0307 15:22:03.411624 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-bk5xc_c400fdf2-390c-4c48-8c3e-4854d2e0b58f/kube-rbac-proxy/0.log" Mar 07 15:22:03 crc kubenswrapper[4943]: I0307 15:22:03.941905 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548282-tblqn" Mar 07 15:22:04 crc kubenswrapper[4943]: I0307 15:22:04.026425 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qvmv9\" (UniqueName: \"kubernetes.io/projected/43076e34-f66f-4bba-89dc-bc312cd547df-kube-api-access-qvmv9\") pod \"43076e34-f66f-4bba-89dc-bc312cd547df\" (UID: \"43076e34-f66f-4bba-89dc-bc312cd547df\") " Mar 07 15:22:04 crc kubenswrapper[4943]: I0307 15:22:04.036816 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43076e34-f66f-4bba-89dc-bc312cd547df-kube-api-access-qvmv9" (OuterVolumeSpecName: "kube-api-access-qvmv9") pod "43076e34-f66f-4bba-89dc-bc312cd547df" (UID: "43076e34-f66f-4bba-89dc-bc312cd547df"). InnerVolumeSpecName "kube-api-access-qvmv9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:22:04 crc kubenswrapper[4943]: I0307 15:22:04.128369 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qvmv9\" (UniqueName: \"kubernetes.io/projected/43076e34-f66f-4bba-89dc-bc312cd547df-kube-api-access-qvmv9\") on node \"crc\" DevicePath \"\"" Mar 07 15:22:04 crc kubenswrapper[4943]: I0307 15:22:04.691756 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548282-tblqn" event={"ID":"43076e34-f66f-4bba-89dc-bc312cd547df","Type":"ContainerDied","Data":"ea98caf63ededa8cdfd05756fdb35a757788a5650a8d7d0799e80aa3713e4c2c"} Mar 07 15:22:04 crc kubenswrapper[4943]: I0307 15:22:04.691803 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ea98caf63ededa8cdfd05756fdb35a757788a5650a8d7d0799e80aa3713e4c2c" Mar 07 15:22:04 crc kubenswrapper[4943]: I0307 15:22:04.692377 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548282-tblqn" Mar 07 15:22:05 crc kubenswrapper[4943]: I0307 15:22:05.000853 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29548276-5qggb"] Mar 07 15:22:05 crc kubenswrapper[4943]: I0307 15:22:05.009220 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29548276-5qggb"] Mar 07 15:22:06 crc kubenswrapper[4943]: I0307 15:22:06.770651 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7215a3b-6a23-4b8d-bf59-de153d613605" path="/var/lib/kubelet/pods/c7215a3b-6a23-4b8d-bf59-de153d613605/volumes" Mar 07 15:22:25 crc kubenswrapper[4943]: I0307 15:22:25.330819 4943 scope.go:117] "RemoveContainer" containerID="6bf668a5979e8f1613e03fabd8ad361d86548ae73aafe064ef76ccb9eddbd1fd" Mar 07 15:22:25 crc kubenswrapper[4943]: I0307 15:22:25.385759 4943 scope.go:117] "RemoveContainer" containerID="b74fb102430ba96aef617e3e49ffdf5c1b424ac7bef18593d09befce2d7dd8f6" Mar 07 15:22:25 crc kubenswrapper[4943]: I0307 15:22:25.418356 4943 scope.go:117] "RemoveContainer" containerID="a736868f2888b612a98b97aaeb4cf05314e6e0f58df2907209bfa17f89dc7b4a" Mar 07 15:22:25 crc kubenswrapper[4943]: I0307 15:22:25.477121 4943 scope.go:117] "RemoveContainer" containerID="8769f5fcfacebac78a23380cbd6f9ea0f53a3990a721ef60a5f080181a664f5f" Mar 07 15:22:25 crc kubenswrapper[4943]: I0307 15:22:25.507469 4943 scope.go:117] "RemoveContainer" containerID="2691dce5a061b11d167965c7e2049be3d1bf60151c7e2349125fc3120468c7db" Mar 07 15:22:25 crc kubenswrapper[4943]: I0307 15:22:25.533197 4943 scope.go:117] "RemoveContainer" containerID="37428c462b8d4fed4c4147a76f6eff119d556b1a17f576bdc1717aa7b58b2d7a" Mar 07 15:22:25 crc kubenswrapper[4943]: I0307 15:22:25.574292 4943 scope.go:117] "RemoveContainer" containerID="4f6fa82d7c8fbcb5ffbe12a1b3c1454a3713c428d9888a404d38e918c2e10b37" Mar 07 15:22:25 crc kubenswrapper[4943]: I0307 15:22:25.597279 4943 scope.go:117] "RemoveContainer" containerID="0427dcf71f86ffa745eea10704d4fb1aacf6e0c845aa3d0cf085fb10961aedf2" Mar 07 15:22:25 crc kubenswrapper[4943]: I0307 15:22:25.620761 4943 scope.go:117] "RemoveContainer" containerID="725afe0d98faa0ad3972dd436d3ce97ef2226e2243b7710cda821e041b293b62" Mar 07 15:22:25 crc kubenswrapper[4943]: I0307 15:22:25.644504 4943 scope.go:117] "RemoveContainer" containerID="d486221e7cf6b2e38945d6f110d67cdb105d1da60f995808ad70faebec537173" Mar 07 15:22:35 crc kubenswrapper[4943]: I0307 15:22:35.433680 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-86ddb6bd46-wxgzg_31d1a300-cbac-4266-8182-300930377412/controller/0.log" Mar 07 15:22:35 crc kubenswrapper[4943]: I0307 15:22:35.441232 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-86ddb6bd46-wxgzg_31d1a300-cbac-4266-8182-300930377412/kube-rbac-proxy/0.log" Mar 07 15:22:35 crc kubenswrapper[4943]: I0307 15:22:35.722636 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ntxr6_19220a5d-633c-4051-87d4-246e167e6db5/cp-frr-files/0.log" Mar 07 15:22:35 crc kubenswrapper[4943]: I0307 15:22:35.820791 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ntxr6_19220a5d-633c-4051-87d4-246e167e6db5/cp-reloader/0.log" Mar 07 15:22:35 crc kubenswrapper[4943]: I0307 15:22:35.829908 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ntxr6_19220a5d-633c-4051-87d4-246e167e6db5/cp-metrics/0.log" Mar 07 15:22:35 crc kubenswrapper[4943]: I0307 15:22:35.856850 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ntxr6_19220a5d-633c-4051-87d4-246e167e6db5/cp-frr-files/0.log" Mar 07 15:22:35 crc kubenswrapper[4943]: I0307 15:22:35.934652 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ntxr6_19220a5d-633c-4051-87d4-246e167e6db5/cp-reloader/0.log" Mar 07 15:22:36 crc kubenswrapper[4943]: I0307 15:22:36.009983 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ntxr6_19220a5d-633c-4051-87d4-246e167e6db5/cp-frr-files/0.log" Mar 07 15:22:36 crc kubenswrapper[4943]: I0307 15:22:36.025048 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ntxr6_19220a5d-633c-4051-87d4-246e167e6db5/cp-reloader/0.log" Mar 07 15:22:36 crc kubenswrapper[4943]: I0307 15:22:36.058970 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ntxr6_19220a5d-633c-4051-87d4-246e167e6db5/cp-metrics/0.log" Mar 07 15:22:36 crc kubenswrapper[4943]: I0307 15:22:36.120671 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ntxr6_19220a5d-633c-4051-87d4-246e167e6db5/cp-metrics/0.log" Mar 07 15:22:36 crc kubenswrapper[4943]: I0307 15:22:36.264873 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ntxr6_19220a5d-633c-4051-87d4-246e167e6db5/cp-reloader/0.log" Mar 07 15:22:36 crc kubenswrapper[4943]: I0307 15:22:36.281956 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ntxr6_19220a5d-633c-4051-87d4-246e167e6db5/cp-frr-files/0.log" Mar 07 15:22:36 crc kubenswrapper[4943]: I0307 15:22:36.304628 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ntxr6_19220a5d-633c-4051-87d4-246e167e6db5/controller/0.log" Mar 07 15:22:36 crc kubenswrapper[4943]: I0307 15:22:36.330736 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ntxr6_19220a5d-633c-4051-87d4-246e167e6db5/cp-metrics/0.log" Mar 07 15:22:36 crc kubenswrapper[4943]: I0307 15:22:36.445587 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ntxr6_19220a5d-633c-4051-87d4-246e167e6db5/frr-metrics/0.log" Mar 07 15:22:36 crc kubenswrapper[4943]: I0307 15:22:36.459054 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ntxr6_19220a5d-633c-4051-87d4-246e167e6db5/kube-rbac-proxy/0.log" Mar 07 15:22:36 crc kubenswrapper[4943]: I0307 15:22:36.524841 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ntxr6_19220a5d-633c-4051-87d4-246e167e6db5/kube-rbac-proxy-frr/0.log" Mar 07 15:22:36 crc kubenswrapper[4943]: I0307 15:22:36.657031 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ntxr6_19220a5d-633c-4051-87d4-246e167e6db5/reloader/0.log" Mar 07 15:22:36 crc kubenswrapper[4943]: I0307 15:22:36.768381 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7f989f654f-h29zk_6e721eda-a709-4537-8f7a-4bdc31bf72f8/frr-k8s-webhook-server/0.log" Mar 07 15:22:36 crc kubenswrapper[4943]: I0307 15:22:36.948809 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-655d5479c5-wnsv8_75e7a24e-c89f-49b4-bae4-cf2627e32068/manager/0.log" Mar 07 15:22:37 crc kubenswrapper[4943]: I0307 15:22:37.117468 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-5f7c595484-khsdg_de0f2fd6-4595-498c-9e76-6f8635fe23f6/webhook-server/0.log" Mar 07 15:22:37 crc kubenswrapper[4943]: I0307 15:22:37.238030 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-22p8b_e8cc9b78-5dd4-43c6-a1fe-702f4f1c30c9/kube-rbac-proxy/0.log" Mar 07 15:22:37 crc kubenswrapper[4943]: I0307 15:22:37.507872 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-22p8b_e8cc9b78-5dd4-43c6-a1fe-702f4f1c30c9/speaker/0.log" Mar 07 15:22:37 crc kubenswrapper[4943]: I0307 15:22:37.820963 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ntxr6_19220a5d-633c-4051-87d4-246e167e6db5/frr/0.log" Mar 07 15:22:53 crc kubenswrapper[4943]: I0307 15:22:53.344210 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_barbican-api-57985b48cb-dqxs9_6f106573-1c59-4abf-aeb2-30bacbd53b04/barbican-api/0.log" Mar 07 15:22:53 crc kubenswrapper[4943]: I0307 15:22:53.438501 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_barbican-api-57985b48cb-dqxs9_6f106573-1c59-4abf-aeb2-30bacbd53b04/barbican-api-log/0.log" Mar 07 15:22:53 crc kubenswrapper[4943]: I0307 15:22:53.502619 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_barbican-db-sync-xtjgw_76563709-9e41-463e-a16f-d30e9cd4abf8/barbican-db-sync/0.log" Mar 07 15:22:53 crc kubenswrapper[4943]: I0307 15:22:53.604713 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_barbican-keystone-listener-54998b9498-bh574_d74297b6-b4e3-4dd8-8ee0-fd403bf79254/barbican-keystone-listener/0.log" Mar 07 15:22:53 crc kubenswrapper[4943]: I0307 15:22:53.641965 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_barbican-keystone-listener-54998b9498-bh574_d74297b6-b4e3-4dd8-8ee0-fd403bf79254/barbican-keystone-listener-log/0.log" Mar 07 15:22:53 crc kubenswrapper[4943]: I0307 15:22:53.715403 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_barbican-worker-8bf566dc7-r7nnf_2159c059-075c-4986-8244-bf6d57dc3cd9/barbican-worker/0.log" Mar 07 15:22:53 crc kubenswrapper[4943]: I0307 15:22:53.780436 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_barbican-worker-8bf566dc7-r7nnf_2159c059-075c-4986-8244-bf6d57dc3cd9/barbican-worker-log/0.log" Mar 07 15:22:53 crc kubenswrapper[4943]: I0307 15:22:53.978431 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_keystone-cron-29548261-jgm2m_d1b17bd5-bcfb-49b7-9a5a-9d718d864740/keystone-cron/0.log" Mar 07 15:22:54 crc kubenswrapper[4943]: I0307 15:22:54.084018 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_keystone-5c4bccddcc-5vjzw_4c58fc8e-3b29-49a6-b592-9455d49f89cd/keystone-api/0.log" Mar 07 15:22:54 crc kubenswrapper[4943]: I0307 15:22:54.234474 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_openstack-galera-0_92e3c4ec-dbf5-443c-b11a-7d1d961a92c0/mysql-bootstrap/0.log" Mar 07 15:22:54 crc kubenswrapper[4943]: I0307 15:22:54.446305 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_openstack-galera-0_92e3c4ec-dbf5-443c-b11a-7d1d961a92c0/mysql-bootstrap/0.log" Mar 07 15:22:54 crc kubenswrapper[4943]: I0307 15:22:54.467552 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_openstack-galera-0_92e3c4ec-dbf5-443c-b11a-7d1d961a92c0/galera/0.log" Mar 07 15:22:54 crc kubenswrapper[4943]: I0307 15:22:54.644436 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_openstack-galera-1_4f2c813e-73c6-41c7-ba8d-5498fe1d5671/mysql-bootstrap/0.log" Mar 07 15:22:54 crc kubenswrapper[4943]: I0307 15:22:54.783124 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_openstack-galera-1_4f2c813e-73c6-41c7-ba8d-5498fe1d5671/mysql-bootstrap/0.log" Mar 07 15:22:54 crc kubenswrapper[4943]: I0307 15:22:54.789106 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_openstack-galera-1_4f2c813e-73c6-41c7-ba8d-5498fe1d5671/galera/0.log" Mar 07 15:22:54 crc kubenswrapper[4943]: I0307 15:22:54.987167 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_openstack-galera-2_64dcacc7-fd15-41e0-ac82-fc37aa562331/mysql-bootstrap/0.log" Mar 07 15:22:55 crc kubenswrapper[4943]: I0307 15:22:55.175239 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_openstack-galera-2_64dcacc7-fd15-41e0-ac82-fc37aa562331/mysql-bootstrap/0.log" Mar 07 15:22:55 crc kubenswrapper[4943]: I0307 15:22:55.216428 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_openstack-galera-2_64dcacc7-fd15-41e0-ac82-fc37aa562331/galera/0.log" Mar 07 15:22:55 crc kubenswrapper[4943]: I0307 15:22:55.417585 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_rabbitmq-server-0_f73c32b4-3904-4bf3-9bb4-13750cb45e88/setup-container/0.log" Mar 07 15:22:55 crc kubenswrapper[4943]: I0307 15:22:55.566069 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_rabbitmq-server-0_f73c32b4-3904-4bf3-9bb4-13750cb45e88/rabbitmq/0.log" Mar 07 15:22:55 crc kubenswrapper[4943]: I0307 15:22:55.582157 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_rabbitmq-server-0_f73c32b4-3904-4bf3-9bb4-13750cb45e88/setup-container/0.log" Mar 07 15:22:55 crc kubenswrapper[4943]: I0307 15:22:55.799034 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-proxy-76c998454c-kngkw_32f9bc3c-6fe2-42ab-8485-d7fda4d10c78/proxy-httpd/0.log" Mar 07 15:22:55 crc kubenswrapper[4943]: I0307 15:22:55.883205 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-proxy-76c998454c-kngkw_32f9bc3c-6fe2-42ab-8485-d7fda4d10c78/proxy-server/0.log" Mar 07 15:22:56 crc kubenswrapper[4943]: I0307 15:22:56.023795 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-ring-rebalance-s5kqz_85e1e3e8-b9f8-4c3d-940e-4060262201ee/swift-ring-rebalance/0.log" Mar 07 15:22:56 crc kubenswrapper[4943]: I0307 15:22:56.175685 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_memcached-0_5d85dcb0-3ed6-4fcf-8ee9-49e7c8cf3385/memcached/0.log" Mar 07 15:22:56 crc kubenswrapper[4943]: I0307 15:22:56.216980 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-0_93b86e6f-5b8a-45fe-ac00-62b667d92a76/account-auditor/0.log" Mar 07 15:22:56 crc kubenswrapper[4943]: I0307 15:22:56.251448 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-0_93b86e6f-5b8a-45fe-ac00-62b667d92a76/account-reaper/0.log" Mar 07 15:22:56 crc kubenswrapper[4943]: I0307 15:22:56.317591 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-0_93b86e6f-5b8a-45fe-ac00-62b667d92a76/account-server/0.log" Mar 07 15:22:56 crc kubenswrapper[4943]: I0307 15:22:56.322849 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-0_93b86e6f-5b8a-45fe-ac00-62b667d92a76/account-replicator/0.log" Mar 07 15:22:56 crc kubenswrapper[4943]: I0307 15:22:56.516296 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-0_93b86e6f-5b8a-45fe-ac00-62b667d92a76/container-replicator/0.log" Mar 07 15:22:56 crc kubenswrapper[4943]: I0307 15:22:56.516664 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-0_93b86e6f-5b8a-45fe-ac00-62b667d92a76/container-auditor/0.log" Mar 07 15:22:56 crc kubenswrapper[4943]: I0307 15:22:56.602600 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-0_93b86e6f-5b8a-45fe-ac00-62b667d92a76/container-server/0.log" Mar 07 15:22:56 crc kubenswrapper[4943]: I0307 15:22:56.636884 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-0_93b86e6f-5b8a-45fe-ac00-62b667d92a76/container-updater/0.log" Mar 07 15:22:56 crc kubenswrapper[4943]: I0307 15:22:56.689377 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-0_93b86e6f-5b8a-45fe-ac00-62b667d92a76/object-auditor/0.log" Mar 07 15:22:56 crc kubenswrapper[4943]: I0307 15:22:56.701619 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-0_93b86e6f-5b8a-45fe-ac00-62b667d92a76/object-expirer/0.log" Mar 07 15:22:56 crc kubenswrapper[4943]: I0307 15:22:56.752275 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-0_93b86e6f-5b8a-45fe-ac00-62b667d92a76/object-replicator/0.log" Mar 07 15:22:56 crc kubenswrapper[4943]: I0307 15:22:56.847114 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-0_93b86e6f-5b8a-45fe-ac00-62b667d92a76/object-updater/0.log" Mar 07 15:22:56 crc kubenswrapper[4943]: I0307 15:22:56.853890 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-0_93b86e6f-5b8a-45fe-ac00-62b667d92a76/object-server/0.log" Mar 07 15:22:56 crc kubenswrapper[4943]: I0307 15:22:56.871298 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-0_93b86e6f-5b8a-45fe-ac00-62b667d92a76/rsync/0.log" Mar 07 15:22:56 crc kubenswrapper[4943]: I0307 15:22:56.949649 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-0_93b86e6f-5b8a-45fe-ac00-62b667d92a76/swift-recon-cron/0.log" Mar 07 15:22:56 crc kubenswrapper[4943]: I0307 15:22:56.998508 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-1_40da0e9b-e661-487c-8898-a54286149f85/account-reaper/0.log" Mar 07 15:22:57 crc kubenswrapper[4943]: I0307 15:22:57.043413 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-1_40da0e9b-e661-487c-8898-a54286149f85/account-auditor/0.log" Mar 07 15:22:57 crc kubenswrapper[4943]: I0307 15:22:57.060458 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-1_40da0e9b-e661-487c-8898-a54286149f85/account-replicator/0.log" Mar 07 15:22:57 crc kubenswrapper[4943]: I0307 15:22:57.119111 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-1_40da0e9b-e661-487c-8898-a54286149f85/account-server/0.log" Mar 07 15:22:57 crc kubenswrapper[4943]: I0307 15:22:57.170809 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-1_40da0e9b-e661-487c-8898-a54286149f85/container-auditor/0.log" Mar 07 15:22:57 crc kubenswrapper[4943]: I0307 15:22:57.197035 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-1_40da0e9b-e661-487c-8898-a54286149f85/container-server/0.log" Mar 07 15:22:57 crc kubenswrapper[4943]: I0307 15:22:57.212351 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-1_40da0e9b-e661-487c-8898-a54286149f85/container-replicator/0.log" Mar 07 15:22:57 crc kubenswrapper[4943]: I0307 15:22:57.236299 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-1_40da0e9b-e661-487c-8898-a54286149f85/container-updater/0.log" Mar 07 15:22:57 crc kubenswrapper[4943]: I0307 15:22:57.324559 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-1_40da0e9b-e661-487c-8898-a54286149f85/object-auditor/0.log" Mar 07 15:22:57 crc kubenswrapper[4943]: I0307 15:22:57.327212 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-1_40da0e9b-e661-487c-8898-a54286149f85/object-expirer/0.log" Mar 07 15:22:57 crc kubenswrapper[4943]: I0307 15:22:57.372004 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-1_40da0e9b-e661-487c-8898-a54286149f85/object-replicator/0.log" Mar 07 15:22:57 crc kubenswrapper[4943]: I0307 15:22:57.416195 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-1_40da0e9b-e661-487c-8898-a54286149f85/object-server/0.log" Mar 07 15:22:57 crc kubenswrapper[4943]: I0307 15:22:57.430232 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-1_40da0e9b-e661-487c-8898-a54286149f85/object-updater/0.log" Mar 07 15:22:57 crc kubenswrapper[4943]: I0307 15:22:57.512093 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-1_40da0e9b-e661-487c-8898-a54286149f85/swift-recon-cron/0.log" Mar 07 15:22:57 crc kubenswrapper[4943]: I0307 15:22:57.540823 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-1_40da0e9b-e661-487c-8898-a54286149f85/rsync/0.log" Mar 07 15:22:57 crc kubenswrapper[4943]: I0307 15:22:57.576702 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-2_c77eb68a-e396-4958-8657-8e62013b774e/account-reaper/0.log" Mar 07 15:22:57 crc kubenswrapper[4943]: I0307 15:22:57.596862 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-2_c77eb68a-e396-4958-8657-8e62013b774e/account-auditor/0.log" Mar 07 15:22:57 crc kubenswrapper[4943]: I0307 15:22:57.694270 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-2_c77eb68a-e396-4958-8657-8e62013b774e/account-server/0.log" Mar 07 15:22:57 crc kubenswrapper[4943]: I0307 15:22:57.696255 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-2_c77eb68a-e396-4958-8657-8e62013b774e/account-replicator/0.log" Mar 07 15:22:57 crc kubenswrapper[4943]: I0307 15:22:57.732581 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-2_c77eb68a-e396-4958-8657-8e62013b774e/container-auditor/0.log" Mar 07 15:22:57 crc kubenswrapper[4943]: I0307 15:22:57.773721 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-2_c77eb68a-e396-4958-8657-8e62013b774e/container-server/0.log" Mar 07 15:22:57 crc kubenswrapper[4943]: I0307 15:22:57.775049 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-2_c77eb68a-e396-4958-8657-8e62013b774e/container-replicator/0.log" Mar 07 15:22:57 crc kubenswrapper[4943]: I0307 15:22:57.867307 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-2_c77eb68a-e396-4958-8657-8e62013b774e/object-auditor/0.log" Mar 07 15:22:57 crc kubenswrapper[4943]: I0307 15:22:57.872984 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-2_c77eb68a-e396-4958-8657-8e62013b774e/container-updater/0.log" Mar 07 15:22:57 crc kubenswrapper[4943]: I0307 15:22:57.928836 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-2_c77eb68a-e396-4958-8657-8e62013b774e/object-expirer/0.log" Mar 07 15:22:57 crc kubenswrapper[4943]: I0307 15:22:57.938446 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-2_c77eb68a-e396-4958-8657-8e62013b774e/object-server/0.log" Mar 07 15:22:57 crc kubenswrapper[4943]: I0307 15:22:57.952298 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-2_c77eb68a-e396-4958-8657-8e62013b774e/object-replicator/0.log" Mar 07 15:22:58 crc kubenswrapper[4943]: I0307 15:22:58.035805 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-2_c77eb68a-e396-4958-8657-8e62013b774e/object-updater/0.log" Mar 07 15:22:58 crc kubenswrapper[4943]: I0307 15:22:58.044358 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-2_c77eb68a-e396-4958-8657-8e62013b774e/rsync/0.log" Mar 07 15:22:58 crc kubenswrapper[4943]: I0307 15:22:58.098971 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-2_c77eb68a-e396-4958-8657-8e62013b774e/swift-recon-cron/0.log" Mar 07 15:23:13 crc kubenswrapper[4943]: I0307 15:23:13.215012 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-4pbcq_0e0c280e-3497-4f16-bd4d-b26b9f196700/extract-utilities/0.log" Mar 07 15:23:13 crc kubenswrapper[4943]: I0307 15:23:13.391879 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-4pbcq_0e0c280e-3497-4f16-bd4d-b26b9f196700/extract-content/0.log" Mar 07 15:23:13 crc kubenswrapper[4943]: I0307 15:23:13.393327 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-4pbcq_0e0c280e-3497-4f16-bd4d-b26b9f196700/extract-content/0.log" Mar 07 15:23:13 crc kubenswrapper[4943]: I0307 15:23:13.417970 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-4pbcq_0e0c280e-3497-4f16-bd4d-b26b9f196700/extract-utilities/0.log" Mar 07 15:23:13 crc kubenswrapper[4943]: I0307 15:23:13.593166 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-4pbcq_0e0c280e-3497-4f16-bd4d-b26b9f196700/extract-utilities/0.log" Mar 07 15:23:13 crc kubenswrapper[4943]: I0307 15:23:13.621670 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-4pbcq_0e0c280e-3497-4f16-bd4d-b26b9f196700/extract-content/0.log" Mar 07 15:23:13 crc kubenswrapper[4943]: I0307 15:23:13.865443 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rhp5s_d52577a0-c952-477f-bcd3-9eaa06cb2575/extract-utilities/0.log" Mar 07 15:23:13 crc kubenswrapper[4943]: I0307 15:23:13.946829 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rhp5s_d52577a0-c952-477f-bcd3-9eaa06cb2575/extract-content/0.log" Mar 07 15:23:13 crc kubenswrapper[4943]: I0307 15:23:13.967293 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rhp5s_d52577a0-c952-477f-bcd3-9eaa06cb2575/extract-utilities/0.log" Mar 07 15:23:14 crc kubenswrapper[4943]: I0307 15:23:14.039971 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-4pbcq_0e0c280e-3497-4f16-bd4d-b26b9f196700/registry-server/0.log" Mar 07 15:23:14 crc kubenswrapper[4943]: I0307 15:23:14.105868 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rhp5s_d52577a0-c952-477f-bcd3-9eaa06cb2575/extract-content/0.log" Mar 07 15:23:14 crc kubenswrapper[4943]: I0307 15:23:14.200328 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rhp5s_d52577a0-c952-477f-bcd3-9eaa06cb2575/extract-utilities/0.log" Mar 07 15:23:14 crc kubenswrapper[4943]: I0307 15:23:14.212810 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rhp5s_d52577a0-c952-477f-bcd3-9eaa06cb2575/extract-content/0.log" Mar 07 15:23:14 crc kubenswrapper[4943]: I0307 15:23:14.415044 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f48wksn_dfbd5db0-a595-4698-85dd-6a6107bc23c5/util/0.log" Mar 07 15:23:14 crc kubenswrapper[4943]: I0307 15:23:14.583447 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f48wksn_dfbd5db0-a595-4698-85dd-6a6107bc23c5/pull/0.log" Mar 07 15:23:14 crc kubenswrapper[4943]: I0307 15:23:14.628090 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rhp5s_d52577a0-c952-477f-bcd3-9eaa06cb2575/registry-server/0.log" Mar 07 15:23:14 crc kubenswrapper[4943]: I0307 15:23:14.667872 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f48wksn_dfbd5db0-a595-4698-85dd-6a6107bc23c5/util/0.log" Mar 07 15:23:14 crc kubenswrapper[4943]: I0307 15:23:14.678328 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f48wksn_dfbd5db0-a595-4698-85dd-6a6107bc23c5/pull/0.log" Mar 07 15:23:14 crc kubenswrapper[4943]: I0307 15:23:14.844798 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f48wksn_dfbd5db0-a595-4698-85dd-6a6107bc23c5/extract/0.log" Mar 07 15:23:14 crc kubenswrapper[4943]: I0307 15:23:14.851162 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f48wksn_dfbd5db0-a595-4698-85dd-6a6107bc23c5/util/0.log" Mar 07 15:23:15 crc kubenswrapper[4943]: I0307 15:23:15.655496 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f48wksn_dfbd5db0-a595-4698-85dd-6a6107bc23c5/pull/0.log" Mar 07 15:23:15 crc kubenswrapper[4943]: I0307 15:23:15.952338 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-b2fvs_94d5996e-5894-4f62-961f-ce9f41375047/extract-utilities/0.log" Mar 07 15:23:15 crc kubenswrapper[4943]: I0307 15:23:15.962087 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-cpfgk_9ba51437-0e4b-4aca-87c7-1152f9c2a461/marketplace-operator/0.log" Mar 07 15:23:16 crc kubenswrapper[4943]: I0307 15:23:16.126856 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-b2fvs_94d5996e-5894-4f62-961f-ce9f41375047/extract-utilities/0.log" Mar 07 15:23:16 crc kubenswrapper[4943]: I0307 15:23:16.145640 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-b2fvs_94d5996e-5894-4f62-961f-ce9f41375047/extract-content/0.log" Mar 07 15:23:16 crc kubenswrapper[4943]: I0307 15:23:16.146457 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-b2fvs_94d5996e-5894-4f62-961f-ce9f41375047/extract-content/0.log" Mar 07 15:23:16 crc kubenswrapper[4943]: I0307 15:23:16.294171 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-b2fvs_94d5996e-5894-4f62-961f-ce9f41375047/extract-utilities/0.log" Mar 07 15:23:16 crc kubenswrapper[4943]: I0307 15:23:16.319434 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-b2fvs_94d5996e-5894-4f62-961f-ce9f41375047/extract-content/0.log" Mar 07 15:23:16 crc kubenswrapper[4943]: I0307 15:23:16.430227 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-b2fvs_94d5996e-5894-4f62-961f-ce9f41375047/registry-server/0.log" Mar 07 15:23:16 crc kubenswrapper[4943]: I0307 15:23:16.462894 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-k6478_db1d6561-f6c0-44fe-80f5-e5c6f9179c0a/extract-utilities/0.log" Mar 07 15:23:16 crc kubenswrapper[4943]: I0307 15:23:16.665563 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-k6478_db1d6561-f6c0-44fe-80f5-e5c6f9179c0a/extract-utilities/0.log" Mar 07 15:23:16 crc kubenswrapper[4943]: I0307 15:23:16.730525 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-k6478_db1d6561-f6c0-44fe-80f5-e5c6f9179c0a/extract-content/0.log" Mar 07 15:23:16 crc kubenswrapper[4943]: I0307 15:23:16.743239 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-k6478_db1d6561-f6c0-44fe-80f5-e5c6f9179c0a/extract-content/0.log" Mar 07 15:23:16 crc kubenswrapper[4943]: I0307 15:23:16.856644 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-k6478_db1d6561-f6c0-44fe-80f5-e5c6f9179c0a/extract-utilities/0.log" Mar 07 15:23:16 crc kubenswrapper[4943]: I0307 15:23:16.880479 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-k6478_db1d6561-f6c0-44fe-80f5-e5c6f9179c0a/extract-content/0.log" Mar 07 15:23:17 crc kubenswrapper[4943]: I0307 15:23:17.255458 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-k6478_db1d6561-f6c0-44fe-80f5-e5c6f9179c0a/registry-server/0.log" Mar 07 15:23:25 crc kubenswrapper[4943]: I0307 15:23:25.841606 4943 scope.go:117] "RemoveContainer" containerID="38cb674cbb0f8f9c160c67ea4a51102703e4b2c2261fffab9b7cca100ed589b1" Mar 07 15:23:25 crc kubenswrapper[4943]: I0307 15:23:25.869941 4943 scope.go:117] "RemoveContainer" containerID="17f36d0267ff2a0a309f993bbbb9d4fc4ab822fd9f800baa15fe969aaad3074c" Mar 07 15:23:25 crc kubenswrapper[4943]: I0307 15:23:25.918460 4943 scope.go:117] "RemoveContainer" containerID="61cd6cf3937822a98ccfe8d816ceac8893280a0ec691733b1fb03f468d36b867" Mar 07 15:23:25 crc kubenswrapper[4943]: I0307 15:23:25.959674 4943 scope.go:117] "RemoveContainer" containerID="17364057933ada0f1377538d14281f33b01873033659fcf72f6baf62dda3b21e" Mar 07 15:23:26 crc kubenswrapper[4943]: I0307 15:23:26.007245 4943 scope.go:117] "RemoveContainer" containerID="5b9e0bcb311794281eb3c56032cd33eacf7b7b4c013bb10cb04e7e7e66100c23" Mar 07 15:23:26 crc kubenswrapper[4943]: I0307 15:23:26.033697 4943 scope.go:117] "RemoveContainer" containerID="af2f7beea5f0bb15968e7e190112bbc6e1218ddef638e26cd31bfc3092e19963" Mar 07 15:23:26 crc kubenswrapper[4943]: I0307 15:23:26.061159 4943 scope.go:117] "RemoveContainer" containerID="3bd33192c0d7c38312fb29368fbc516d43db1bc3bf3816323cbacf54b8f2a1d2" Mar 07 15:23:26 crc kubenswrapper[4943]: I0307 15:23:26.086356 4943 scope.go:117] "RemoveContainer" containerID="d460a3d5d4eb2979f3dedb56c7984013529157d7cfdc0f799c927072a626b3bf" Mar 07 15:24:00 crc kubenswrapper[4943]: I0307 15:24:00.159505 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29548284-drw5v"] Mar 07 15:24:00 crc kubenswrapper[4943]: E0307 15:24:00.160649 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43076e34-f66f-4bba-89dc-bc312cd547df" containerName="oc" Mar 07 15:24:00 crc kubenswrapper[4943]: I0307 15:24:00.160670 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="43076e34-f66f-4bba-89dc-bc312cd547df" containerName="oc" Mar 07 15:24:00 crc kubenswrapper[4943]: I0307 15:24:00.160966 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="43076e34-f66f-4bba-89dc-bc312cd547df" containerName="oc" Mar 07 15:24:00 crc kubenswrapper[4943]: I0307 15:24:00.161753 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548284-drw5v" Mar 07 15:24:00 crc kubenswrapper[4943]: I0307 15:24:00.164360 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 07 15:24:00 crc kubenswrapper[4943]: I0307 15:24:00.164772 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 07 15:24:00 crc kubenswrapper[4943]: I0307 15:24:00.165098 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-qqlk5" Mar 07 15:24:00 crc kubenswrapper[4943]: I0307 15:24:00.169813 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29548284-drw5v"] Mar 07 15:24:00 crc kubenswrapper[4943]: I0307 15:24:00.353902 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2d55j\" (UniqueName: \"kubernetes.io/projected/21c23f2f-3292-4ad3-864d-59560d7aedc9-kube-api-access-2d55j\") pod \"auto-csr-approver-29548284-drw5v\" (UID: \"21c23f2f-3292-4ad3-864d-59560d7aedc9\") " pod="openshift-infra/auto-csr-approver-29548284-drw5v" Mar 07 15:24:00 crc kubenswrapper[4943]: I0307 15:24:00.456187 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2d55j\" (UniqueName: \"kubernetes.io/projected/21c23f2f-3292-4ad3-864d-59560d7aedc9-kube-api-access-2d55j\") pod \"auto-csr-approver-29548284-drw5v\" (UID: \"21c23f2f-3292-4ad3-864d-59560d7aedc9\") " pod="openshift-infra/auto-csr-approver-29548284-drw5v" Mar 07 15:24:00 crc kubenswrapper[4943]: I0307 15:24:00.494356 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2d55j\" (UniqueName: \"kubernetes.io/projected/21c23f2f-3292-4ad3-864d-59560d7aedc9-kube-api-access-2d55j\") pod \"auto-csr-approver-29548284-drw5v\" (UID: \"21c23f2f-3292-4ad3-864d-59560d7aedc9\") " pod="openshift-infra/auto-csr-approver-29548284-drw5v" Mar 07 15:24:00 crc kubenswrapper[4943]: I0307 15:24:00.522699 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548284-drw5v" Mar 07 15:24:00 crc kubenswrapper[4943]: I0307 15:24:00.808515 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29548284-drw5v"] Mar 07 15:24:00 crc kubenswrapper[4943]: I0307 15:24:00.823126 4943 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 07 15:24:01 crc kubenswrapper[4943]: I0307 15:24:01.080567 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548284-drw5v" event={"ID":"21c23f2f-3292-4ad3-864d-59560d7aedc9","Type":"ContainerStarted","Data":"61940b0450cf1da4cd8e862a3a29f223dbcc4a81b89e3fb9a41396732def6a87"} Mar 07 15:24:02 crc kubenswrapper[4943]: I0307 15:24:02.098322 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548284-drw5v" event={"ID":"21c23f2f-3292-4ad3-864d-59560d7aedc9","Type":"ContainerStarted","Data":"c6ada9d858c40263302cb6541cdc78fa658a3f8cbb4c159d15f49891d664a397"} Mar 07 15:24:02 crc kubenswrapper[4943]: I0307 15:24:02.124858 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29548284-drw5v" podStartSLOduration=1.199863127 podStartE2EDuration="2.124844318s" podCreationTimestamp="2026-03-07 15:24:00 +0000 UTC" firstStartedPulling="2026-03-07 15:24:00.822910779 +0000 UTC m=+2682.775047267" lastFinishedPulling="2026-03-07 15:24:01.74789196 +0000 UTC m=+2683.700028458" observedRunningTime="2026-03-07 15:24:02.12043777 +0000 UTC m=+2684.072574308" watchObservedRunningTime="2026-03-07 15:24:02.124844318 +0000 UTC m=+2684.076980806" Mar 07 15:24:03 crc kubenswrapper[4943]: I0307 15:24:03.116142 4943 generic.go:334] "Generic (PLEG): container finished" podID="21c23f2f-3292-4ad3-864d-59560d7aedc9" containerID="c6ada9d858c40263302cb6541cdc78fa658a3f8cbb4c159d15f49891d664a397" exitCode=0 Mar 07 15:24:03 crc kubenswrapper[4943]: I0307 15:24:03.116333 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548284-drw5v" event={"ID":"21c23f2f-3292-4ad3-864d-59560d7aedc9","Type":"ContainerDied","Data":"c6ada9d858c40263302cb6541cdc78fa658a3f8cbb4c159d15f49891d664a397"} Mar 07 15:24:04 crc kubenswrapper[4943]: I0307 15:24:04.516877 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548284-drw5v" Mar 07 15:24:04 crc kubenswrapper[4943]: I0307 15:24:04.625287 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d55j\" (UniqueName: \"kubernetes.io/projected/21c23f2f-3292-4ad3-864d-59560d7aedc9-kube-api-access-2d55j\") pod \"21c23f2f-3292-4ad3-864d-59560d7aedc9\" (UID: \"21c23f2f-3292-4ad3-864d-59560d7aedc9\") " Mar 07 15:24:04 crc kubenswrapper[4943]: I0307 15:24:04.643154 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21c23f2f-3292-4ad3-864d-59560d7aedc9-kube-api-access-2d55j" (OuterVolumeSpecName: "kube-api-access-2d55j") pod "21c23f2f-3292-4ad3-864d-59560d7aedc9" (UID: "21c23f2f-3292-4ad3-864d-59560d7aedc9"). InnerVolumeSpecName "kube-api-access-2d55j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:24:04 crc kubenswrapper[4943]: I0307 15:24:04.727299 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d55j\" (UniqueName: \"kubernetes.io/projected/21c23f2f-3292-4ad3-864d-59560d7aedc9-kube-api-access-2d55j\") on node \"crc\" DevicePath \"\"" Mar 07 15:24:05 crc kubenswrapper[4943]: I0307 15:24:05.140298 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548284-drw5v" event={"ID":"21c23f2f-3292-4ad3-864d-59560d7aedc9","Type":"ContainerDied","Data":"61940b0450cf1da4cd8e862a3a29f223dbcc4a81b89e3fb9a41396732def6a87"} Mar 07 15:24:05 crc kubenswrapper[4943]: I0307 15:24:05.140339 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="61940b0450cf1da4cd8e862a3a29f223dbcc4a81b89e3fb9a41396732def6a87" Mar 07 15:24:05 crc kubenswrapper[4943]: I0307 15:24:05.140345 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548284-drw5v" Mar 07 15:24:05 crc kubenswrapper[4943]: I0307 15:24:05.192254 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29548278-f5c2x"] Mar 07 15:24:05 crc kubenswrapper[4943]: I0307 15:24:05.201216 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29548278-f5c2x"] Mar 07 15:24:06 crc kubenswrapper[4943]: I0307 15:24:06.073646 4943 patch_prober.go:28] interesting pod/machine-config-daemon-cpdmm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 07 15:24:06 crc kubenswrapper[4943]: I0307 15:24:06.074070 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 07 15:24:06 crc kubenswrapper[4943]: I0307 15:24:06.771601 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe264b14-43cf-4101-9c71-ae8a5c77a3e5" path="/var/lib/kubelet/pods/fe264b14-43cf-4101-9c71-ae8a5c77a3e5/volumes" Mar 07 15:24:26 crc kubenswrapper[4943]: I0307 15:24:26.216760 4943 scope.go:117] "RemoveContainer" containerID="e61d64cba756b76315104336679b2c2dd348c08d8e95644f67c79dbfd8b55e13" Mar 07 15:24:26 crc kubenswrapper[4943]: I0307 15:24:26.281614 4943 scope.go:117] "RemoveContainer" containerID="58716c0f983cc31149b498b43d095ecc0c1da9ab57bc83f61ceeeb033fbe320e" Mar 07 15:24:26 crc kubenswrapper[4943]: I0307 15:24:26.353363 4943 scope.go:117] "RemoveContainer" containerID="d29f789a22230e7ae99d1c296f7d812000d0d2ffafa7c90a27fd516111a961e0" Mar 07 15:24:26 crc kubenswrapper[4943]: I0307 15:24:26.395840 4943 scope.go:117] "RemoveContainer" containerID="947668c3e7b1fdf5ff2eff97ead24071b8d501ac7fc0fdcbd0a21eb8e20c7eeb" Mar 07 15:24:26 crc kubenswrapper[4943]: I0307 15:24:26.432192 4943 scope.go:117] "RemoveContainer" containerID="1c4cc644d6231f8c43333e6d4105684175b14eebf2a5269c88dcd41990a21f54" Mar 07 15:24:26 crc kubenswrapper[4943]: I0307 15:24:26.473124 4943 scope.go:117] "RemoveContainer" containerID="53bd7b59c66104a6413ec02643d74763f6805fd1bea29090d815f0fb26a9908c" Mar 07 15:24:26 crc kubenswrapper[4943]: I0307 15:24:26.501404 4943 scope.go:117] "RemoveContainer" containerID="9cee618806c4822eec6839068f839c17ac0b9228435ce5aabeb5cf561850d0fe" Mar 07 15:24:26 crc kubenswrapper[4943]: I0307 15:24:26.524846 4943 scope.go:117] "RemoveContainer" containerID="d1ae9074ee81b7864d70c49b679509bbf9bf02c3d8d0fe7b9fac37707ed78614" Mar 07 15:24:32 crc kubenswrapper[4943]: I0307 15:24:32.442271 4943 generic.go:334] "Generic (PLEG): container finished" podID="112000fc-ea09-47ab-b27b-d70c6f37a8a0" containerID="ba5ca406c42a6e37bb34abce909357f13ae1eaf861c725d47e64e7f968a77817" exitCode=0 Mar 07 15:24:32 crc kubenswrapper[4943]: I0307 15:24:32.442396 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lxzm4/must-gather-lrs4f" event={"ID":"112000fc-ea09-47ab-b27b-d70c6f37a8a0","Type":"ContainerDied","Data":"ba5ca406c42a6e37bb34abce909357f13ae1eaf861c725d47e64e7f968a77817"} Mar 07 15:24:32 crc kubenswrapper[4943]: I0307 15:24:32.445170 4943 scope.go:117] "RemoveContainer" containerID="ba5ca406c42a6e37bb34abce909357f13ae1eaf861c725d47e64e7f968a77817" Mar 07 15:24:33 crc kubenswrapper[4943]: I0307 15:24:33.218906 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-lxzm4_must-gather-lrs4f_112000fc-ea09-47ab-b27b-d70c6f37a8a0/gather/0.log" Mar 07 15:24:36 crc kubenswrapper[4943]: I0307 15:24:36.074068 4943 patch_prober.go:28] interesting pod/machine-config-daemon-cpdmm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 07 15:24:36 crc kubenswrapper[4943]: I0307 15:24:36.074474 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 07 15:24:40 crc kubenswrapper[4943]: I0307 15:24:40.001689 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-lxzm4/must-gather-lrs4f"] Mar 07 15:24:40 crc kubenswrapper[4943]: I0307 15:24:40.002721 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-lxzm4/must-gather-lrs4f" podUID="112000fc-ea09-47ab-b27b-d70c6f37a8a0" containerName="copy" containerID="cri-o://f0b490bf5811307fe9ab134a8894c055a93b3e8cdb5783c6012ee854501f994c" gracePeriod=2 Mar 07 15:24:40 crc kubenswrapper[4943]: I0307 15:24:40.015127 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-lxzm4/must-gather-lrs4f"] Mar 07 15:24:40 crc kubenswrapper[4943]: I0307 15:24:40.447178 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-lxzm4_must-gather-lrs4f_112000fc-ea09-47ab-b27b-d70c6f37a8a0/copy/0.log" Mar 07 15:24:40 crc kubenswrapper[4943]: I0307 15:24:40.447831 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lxzm4/must-gather-lrs4f" Mar 07 15:24:40 crc kubenswrapper[4943]: I0307 15:24:40.524882 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-lxzm4_must-gather-lrs4f_112000fc-ea09-47ab-b27b-d70c6f37a8a0/copy/0.log" Mar 07 15:24:40 crc kubenswrapper[4943]: I0307 15:24:40.525280 4943 generic.go:334] "Generic (PLEG): container finished" podID="112000fc-ea09-47ab-b27b-d70c6f37a8a0" containerID="f0b490bf5811307fe9ab134a8894c055a93b3e8cdb5783c6012ee854501f994c" exitCode=143 Mar 07 15:24:40 crc kubenswrapper[4943]: I0307 15:24:40.525329 4943 scope.go:117] "RemoveContainer" containerID="f0b490bf5811307fe9ab134a8894c055a93b3e8cdb5783c6012ee854501f994c" Mar 07 15:24:40 crc kubenswrapper[4943]: I0307 15:24:40.525352 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lxzm4/must-gather-lrs4f" Mar 07 15:24:40 crc kubenswrapper[4943]: I0307 15:24:40.549711 4943 scope.go:117] "RemoveContainer" containerID="ba5ca406c42a6e37bb34abce909357f13ae1eaf861c725d47e64e7f968a77817" Mar 07 15:24:40 crc kubenswrapper[4943]: I0307 15:24:40.560299 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/112000fc-ea09-47ab-b27b-d70c6f37a8a0-must-gather-output\") pod \"112000fc-ea09-47ab-b27b-d70c6f37a8a0\" (UID: \"112000fc-ea09-47ab-b27b-d70c6f37a8a0\") " Mar 07 15:24:40 crc kubenswrapper[4943]: I0307 15:24:40.560510 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gllrc\" (UniqueName: \"kubernetes.io/projected/112000fc-ea09-47ab-b27b-d70c6f37a8a0-kube-api-access-gllrc\") pod \"112000fc-ea09-47ab-b27b-d70c6f37a8a0\" (UID: \"112000fc-ea09-47ab-b27b-d70c6f37a8a0\") " Mar 07 15:24:40 crc kubenswrapper[4943]: I0307 15:24:40.569395 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/112000fc-ea09-47ab-b27b-d70c6f37a8a0-kube-api-access-gllrc" (OuterVolumeSpecName: "kube-api-access-gllrc") pod "112000fc-ea09-47ab-b27b-d70c6f37a8a0" (UID: "112000fc-ea09-47ab-b27b-d70c6f37a8a0"). InnerVolumeSpecName "kube-api-access-gllrc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:24:40 crc kubenswrapper[4943]: I0307 15:24:40.601341 4943 scope.go:117] "RemoveContainer" containerID="f0b490bf5811307fe9ab134a8894c055a93b3e8cdb5783c6012ee854501f994c" Mar 07 15:24:40 crc kubenswrapper[4943]: E0307 15:24:40.602328 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0b490bf5811307fe9ab134a8894c055a93b3e8cdb5783c6012ee854501f994c\": container with ID starting with f0b490bf5811307fe9ab134a8894c055a93b3e8cdb5783c6012ee854501f994c not found: ID does not exist" containerID="f0b490bf5811307fe9ab134a8894c055a93b3e8cdb5783c6012ee854501f994c" Mar 07 15:24:40 crc kubenswrapper[4943]: I0307 15:24:40.602361 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0b490bf5811307fe9ab134a8894c055a93b3e8cdb5783c6012ee854501f994c"} err="failed to get container status \"f0b490bf5811307fe9ab134a8894c055a93b3e8cdb5783c6012ee854501f994c\": rpc error: code = NotFound desc = could not find container \"f0b490bf5811307fe9ab134a8894c055a93b3e8cdb5783c6012ee854501f994c\": container with ID starting with f0b490bf5811307fe9ab134a8894c055a93b3e8cdb5783c6012ee854501f994c not found: ID does not exist" Mar 07 15:24:40 crc kubenswrapper[4943]: I0307 15:24:40.602386 4943 scope.go:117] "RemoveContainer" containerID="ba5ca406c42a6e37bb34abce909357f13ae1eaf861c725d47e64e7f968a77817" Mar 07 15:24:40 crc kubenswrapper[4943]: E0307 15:24:40.602884 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba5ca406c42a6e37bb34abce909357f13ae1eaf861c725d47e64e7f968a77817\": container with ID starting with ba5ca406c42a6e37bb34abce909357f13ae1eaf861c725d47e64e7f968a77817 not found: ID does not exist" containerID="ba5ca406c42a6e37bb34abce909357f13ae1eaf861c725d47e64e7f968a77817" Mar 07 15:24:40 crc kubenswrapper[4943]: I0307 15:24:40.602924 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba5ca406c42a6e37bb34abce909357f13ae1eaf861c725d47e64e7f968a77817"} err="failed to get container status \"ba5ca406c42a6e37bb34abce909357f13ae1eaf861c725d47e64e7f968a77817\": rpc error: code = NotFound desc = could not find container \"ba5ca406c42a6e37bb34abce909357f13ae1eaf861c725d47e64e7f968a77817\": container with ID starting with ba5ca406c42a6e37bb34abce909357f13ae1eaf861c725d47e64e7f968a77817 not found: ID does not exist" Mar 07 15:24:40 crc kubenswrapper[4943]: I0307 15:24:40.641883 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/112000fc-ea09-47ab-b27b-d70c6f37a8a0-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "112000fc-ea09-47ab-b27b-d70c6f37a8a0" (UID: "112000fc-ea09-47ab-b27b-d70c6f37a8a0"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:24:40 crc kubenswrapper[4943]: I0307 15:24:40.662216 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gllrc\" (UniqueName: \"kubernetes.io/projected/112000fc-ea09-47ab-b27b-d70c6f37a8a0-kube-api-access-gllrc\") on node \"crc\" DevicePath \"\"" Mar 07 15:24:40 crc kubenswrapper[4943]: I0307 15:24:40.662274 4943 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/112000fc-ea09-47ab-b27b-d70c6f37a8a0-must-gather-output\") on node \"crc\" DevicePath \"\"" Mar 07 15:24:40 crc kubenswrapper[4943]: I0307 15:24:40.765758 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="112000fc-ea09-47ab-b27b-d70c6f37a8a0" path="/var/lib/kubelet/pods/112000fc-ea09-47ab-b27b-d70c6f37a8a0/volumes" Mar 07 15:24:49 crc kubenswrapper[4943]: I0307 15:24:49.979478 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hcm4l"] Mar 07 15:24:49 crc kubenswrapper[4943]: E0307 15:24:49.980170 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="112000fc-ea09-47ab-b27b-d70c6f37a8a0" containerName="gather" Mar 07 15:24:49 crc kubenswrapper[4943]: I0307 15:24:49.980181 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="112000fc-ea09-47ab-b27b-d70c6f37a8a0" containerName="gather" Mar 07 15:24:49 crc kubenswrapper[4943]: E0307 15:24:49.980207 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21c23f2f-3292-4ad3-864d-59560d7aedc9" containerName="oc" Mar 07 15:24:49 crc kubenswrapper[4943]: I0307 15:24:49.980214 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="21c23f2f-3292-4ad3-864d-59560d7aedc9" containerName="oc" Mar 07 15:24:49 crc kubenswrapper[4943]: E0307 15:24:49.980224 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="112000fc-ea09-47ab-b27b-d70c6f37a8a0" containerName="copy" Mar 07 15:24:49 crc kubenswrapper[4943]: I0307 15:24:49.980229 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="112000fc-ea09-47ab-b27b-d70c6f37a8a0" containerName="copy" Mar 07 15:24:49 crc kubenswrapper[4943]: I0307 15:24:49.980359 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="112000fc-ea09-47ab-b27b-d70c6f37a8a0" containerName="copy" Mar 07 15:24:49 crc kubenswrapper[4943]: I0307 15:24:49.980372 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="112000fc-ea09-47ab-b27b-d70c6f37a8a0" containerName="gather" Mar 07 15:24:49 crc kubenswrapper[4943]: I0307 15:24:49.980383 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="21c23f2f-3292-4ad3-864d-59560d7aedc9" containerName="oc" Mar 07 15:24:49 crc kubenswrapper[4943]: I0307 15:24:49.981270 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hcm4l" Mar 07 15:24:49 crc kubenswrapper[4943]: I0307 15:24:49.997575 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hcm4l"] Mar 07 15:24:50 crc kubenswrapper[4943]: I0307 15:24:50.101378 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmbkq\" (UniqueName: \"kubernetes.io/projected/a3cf2fba-57af-4b15-a687-467ccdc1b3c9-kube-api-access-hmbkq\") pod \"community-operators-hcm4l\" (UID: \"a3cf2fba-57af-4b15-a687-467ccdc1b3c9\") " pod="openshift-marketplace/community-operators-hcm4l" Mar 07 15:24:50 crc kubenswrapper[4943]: I0307 15:24:50.101459 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3cf2fba-57af-4b15-a687-467ccdc1b3c9-catalog-content\") pod \"community-operators-hcm4l\" (UID: \"a3cf2fba-57af-4b15-a687-467ccdc1b3c9\") " pod="openshift-marketplace/community-operators-hcm4l" Mar 07 15:24:50 crc kubenswrapper[4943]: I0307 15:24:50.101565 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3cf2fba-57af-4b15-a687-467ccdc1b3c9-utilities\") pod \"community-operators-hcm4l\" (UID: \"a3cf2fba-57af-4b15-a687-467ccdc1b3c9\") " pod="openshift-marketplace/community-operators-hcm4l" Mar 07 15:24:50 crc kubenswrapper[4943]: I0307 15:24:50.202985 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3cf2fba-57af-4b15-a687-467ccdc1b3c9-utilities\") pod \"community-operators-hcm4l\" (UID: \"a3cf2fba-57af-4b15-a687-467ccdc1b3c9\") " pod="openshift-marketplace/community-operators-hcm4l" Mar 07 15:24:50 crc kubenswrapper[4943]: I0307 15:24:50.203084 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmbkq\" (UniqueName: \"kubernetes.io/projected/a3cf2fba-57af-4b15-a687-467ccdc1b3c9-kube-api-access-hmbkq\") pod \"community-operators-hcm4l\" (UID: \"a3cf2fba-57af-4b15-a687-467ccdc1b3c9\") " pod="openshift-marketplace/community-operators-hcm4l" Mar 07 15:24:50 crc kubenswrapper[4943]: I0307 15:24:50.203196 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3cf2fba-57af-4b15-a687-467ccdc1b3c9-catalog-content\") pod \"community-operators-hcm4l\" (UID: \"a3cf2fba-57af-4b15-a687-467ccdc1b3c9\") " pod="openshift-marketplace/community-operators-hcm4l" Mar 07 15:24:50 crc kubenswrapper[4943]: I0307 15:24:50.203632 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3cf2fba-57af-4b15-a687-467ccdc1b3c9-utilities\") pod \"community-operators-hcm4l\" (UID: \"a3cf2fba-57af-4b15-a687-467ccdc1b3c9\") " pod="openshift-marketplace/community-operators-hcm4l" Mar 07 15:24:50 crc kubenswrapper[4943]: I0307 15:24:50.203745 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3cf2fba-57af-4b15-a687-467ccdc1b3c9-catalog-content\") pod \"community-operators-hcm4l\" (UID: \"a3cf2fba-57af-4b15-a687-467ccdc1b3c9\") " pod="openshift-marketplace/community-operators-hcm4l" Mar 07 15:24:50 crc kubenswrapper[4943]: I0307 15:24:50.227533 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmbkq\" (UniqueName: \"kubernetes.io/projected/a3cf2fba-57af-4b15-a687-467ccdc1b3c9-kube-api-access-hmbkq\") pod \"community-operators-hcm4l\" (UID: \"a3cf2fba-57af-4b15-a687-467ccdc1b3c9\") " pod="openshift-marketplace/community-operators-hcm4l" Mar 07 15:24:50 crc kubenswrapper[4943]: I0307 15:24:50.310617 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hcm4l" Mar 07 15:24:50 crc kubenswrapper[4943]: I0307 15:24:50.766701 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hcm4l"] Mar 07 15:24:51 crc kubenswrapper[4943]: I0307 15:24:51.624920 4943 generic.go:334] "Generic (PLEG): container finished" podID="a3cf2fba-57af-4b15-a687-467ccdc1b3c9" containerID="b011f925232065feadc544fcf61034af37901c4d8a62ae5bf536f75d02863cba" exitCode=0 Mar 07 15:24:51 crc kubenswrapper[4943]: I0307 15:24:51.625004 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hcm4l" event={"ID":"a3cf2fba-57af-4b15-a687-467ccdc1b3c9","Type":"ContainerDied","Data":"b011f925232065feadc544fcf61034af37901c4d8a62ae5bf536f75d02863cba"} Mar 07 15:24:51 crc kubenswrapper[4943]: I0307 15:24:51.625042 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hcm4l" event={"ID":"a3cf2fba-57af-4b15-a687-467ccdc1b3c9","Type":"ContainerStarted","Data":"097debe653a93385dad6c9c840f868a2a62c321c84abda08818a81ea25bec08f"} Mar 07 15:24:53 crc kubenswrapper[4943]: I0307 15:24:53.646442 4943 generic.go:334] "Generic (PLEG): container finished" podID="a3cf2fba-57af-4b15-a687-467ccdc1b3c9" containerID="2d9e0edb3666c367335fa8cc4633436fd14df4fa98ddbf5d2c98262a412ecfe8" exitCode=0 Mar 07 15:24:53 crc kubenswrapper[4943]: I0307 15:24:53.646559 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hcm4l" event={"ID":"a3cf2fba-57af-4b15-a687-467ccdc1b3c9","Type":"ContainerDied","Data":"2d9e0edb3666c367335fa8cc4633436fd14df4fa98ddbf5d2c98262a412ecfe8"} Mar 07 15:24:54 crc kubenswrapper[4943]: I0307 15:24:54.659466 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hcm4l" event={"ID":"a3cf2fba-57af-4b15-a687-467ccdc1b3c9","Type":"ContainerStarted","Data":"3e3497f413bec7b7d1da20dcb6fbe35cb9cbd38b1f511568d80cc0176c2758c9"} Mar 07 15:24:54 crc kubenswrapper[4943]: I0307 15:24:54.692874 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hcm4l" podStartSLOduration=3.260020892 podStartE2EDuration="5.692855699s" podCreationTimestamp="2026-03-07 15:24:49 +0000 UTC" firstStartedPulling="2026-03-07 15:24:51.627369176 +0000 UTC m=+2733.579505704" lastFinishedPulling="2026-03-07 15:24:54.060204013 +0000 UTC m=+2736.012340511" observedRunningTime="2026-03-07 15:24:54.687711063 +0000 UTC m=+2736.639847621" watchObservedRunningTime="2026-03-07 15:24:54.692855699 +0000 UTC m=+2736.644992207" Mar 07 15:24:59 crc kubenswrapper[4943]: I0307 15:24:59.591150 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xwkf5"] Mar 07 15:24:59 crc kubenswrapper[4943]: I0307 15:24:59.594761 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xwkf5" Mar 07 15:24:59 crc kubenswrapper[4943]: I0307 15:24:59.608344 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xwkf5"] Mar 07 15:24:59 crc kubenswrapper[4943]: I0307 15:24:59.748945 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6634c3f2-9aea-4561-aeaf-0a0f59f3d029-catalog-content\") pod \"certified-operators-xwkf5\" (UID: \"6634c3f2-9aea-4561-aeaf-0a0f59f3d029\") " pod="openshift-marketplace/certified-operators-xwkf5" Mar 07 15:24:59 crc kubenswrapper[4943]: I0307 15:24:59.749054 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lh5q4\" (UniqueName: \"kubernetes.io/projected/6634c3f2-9aea-4561-aeaf-0a0f59f3d029-kube-api-access-lh5q4\") pod \"certified-operators-xwkf5\" (UID: \"6634c3f2-9aea-4561-aeaf-0a0f59f3d029\") " pod="openshift-marketplace/certified-operators-xwkf5" Mar 07 15:24:59 crc kubenswrapper[4943]: I0307 15:24:59.749085 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6634c3f2-9aea-4561-aeaf-0a0f59f3d029-utilities\") pod \"certified-operators-xwkf5\" (UID: \"6634c3f2-9aea-4561-aeaf-0a0f59f3d029\") " pod="openshift-marketplace/certified-operators-xwkf5" Mar 07 15:24:59 crc kubenswrapper[4943]: I0307 15:24:59.850525 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lh5q4\" (UniqueName: \"kubernetes.io/projected/6634c3f2-9aea-4561-aeaf-0a0f59f3d029-kube-api-access-lh5q4\") pod \"certified-operators-xwkf5\" (UID: \"6634c3f2-9aea-4561-aeaf-0a0f59f3d029\") " pod="openshift-marketplace/certified-operators-xwkf5" Mar 07 15:24:59 crc kubenswrapper[4943]: I0307 15:24:59.850581 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6634c3f2-9aea-4561-aeaf-0a0f59f3d029-utilities\") pod \"certified-operators-xwkf5\" (UID: \"6634c3f2-9aea-4561-aeaf-0a0f59f3d029\") " pod="openshift-marketplace/certified-operators-xwkf5" Mar 07 15:24:59 crc kubenswrapper[4943]: I0307 15:24:59.851182 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6634c3f2-9aea-4561-aeaf-0a0f59f3d029-utilities\") pod \"certified-operators-xwkf5\" (UID: \"6634c3f2-9aea-4561-aeaf-0a0f59f3d029\") " pod="openshift-marketplace/certified-operators-xwkf5" Mar 07 15:24:59 crc kubenswrapper[4943]: I0307 15:24:59.852084 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6634c3f2-9aea-4561-aeaf-0a0f59f3d029-catalog-content\") pod \"certified-operators-xwkf5\" (UID: \"6634c3f2-9aea-4561-aeaf-0a0f59f3d029\") " pod="openshift-marketplace/certified-operators-xwkf5" Mar 07 15:24:59 crc kubenswrapper[4943]: I0307 15:24:59.852401 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6634c3f2-9aea-4561-aeaf-0a0f59f3d029-catalog-content\") pod \"certified-operators-xwkf5\" (UID: \"6634c3f2-9aea-4561-aeaf-0a0f59f3d029\") " pod="openshift-marketplace/certified-operators-xwkf5" Mar 07 15:24:59 crc kubenswrapper[4943]: I0307 15:24:59.875569 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lh5q4\" (UniqueName: \"kubernetes.io/projected/6634c3f2-9aea-4561-aeaf-0a0f59f3d029-kube-api-access-lh5q4\") pod \"certified-operators-xwkf5\" (UID: \"6634c3f2-9aea-4561-aeaf-0a0f59f3d029\") " pod="openshift-marketplace/certified-operators-xwkf5" Mar 07 15:24:59 crc kubenswrapper[4943]: I0307 15:24:59.921705 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xwkf5" Mar 07 15:25:00 crc kubenswrapper[4943]: I0307 15:25:00.310830 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hcm4l" Mar 07 15:25:00 crc kubenswrapper[4943]: I0307 15:25:00.311168 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hcm4l" Mar 07 15:25:00 crc kubenswrapper[4943]: I0307 15:25:00.370010 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hcm4l" Mar 07 15:25:00 crc kubenswrapper[4943]: I0307 15:25:00.787765 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hcm4l" Mar 07 15:25:01 crc kubenswrapper[4943]: I0307 15:25:01.182285 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xwkf5"] Mar 07 15:25:01 crc kubenswrapper[4943]: I0307 15:25:01.737497 4943 generic.go:334] "Generic (PLEG): container finished" podID="6634c3f2-9aea-4561-aeaf-0a0f59f3d029" containerID="dcded2899efb30f8ea8b658cfc7fb98b55595f42e18c33263a9ced74b34c8151" exitCode=0 Mar 07 15:25:01 crc kubenswrapper[4943]: I0307 15:25:01.737566 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xwkf5" event={"ID":"6634c3f2-9aea-4561-aeaf-0a0f59f3d029","Type":"ContainerDied","Data":"dcded2899efb30f8ea8b658cfc7fb98b55595f42e18c33263a9ced74b34c8151"} Mar 07 15:25:01 crc kubenswrapper[4943]: I0307 15:25:01.738186 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xwkf5" event={"ID":"6634c3f2-9aea-4561-aeaf-0a0f59f3d029","Type":"ContainerStarted","Data":"9119c722fe76c1c60d7e196d8c50e6b5165640ee23c605cd9f2992498fd77bc7"} Mar 07 15:25:02 crc kubenswrapper[4943]: I0307 15:25:02.981689 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hcm4l"] Mar 07 15:25:03 crc kubenswrapper[4943]: I0307 15:25:03.773821 4943 generic.go:334] "Generic (PLEG): container finished" podID="6634c3f2-9aea-4561-aeaf-0a0f59f3d029" containerID="7c985d0c6b4bd903a0cac2873eb8b4e1364e219c1f067e7838a5dcb5f96554f7" exitCode=0 Mar 07 15:25:03 crc kubenswrapper[4943]: I0307 15:25:03.774830 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hcm4l" podUID="a3cf2fba-57af-4b15-a687-467ccdc1b3c9" containerName="registry-server" containerID="cri-o://3e3497f413bec7b7d1da20dcb6fbe35cb9cbd38b1f511568d80cc0176c2758c9" gracePeriod=2 Mar 07 15:25:03 crc kubenswrapper[4943]: I0307 15:25:03.775824 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xwkf5" event={"ID":"6634c3f2-9aea-4561-aeaf-0a0f59f3d029","Type":"ContainerDied","Data":"7c985d0c6b4bd903a0cac2873eb8b4e1364e219c1f067e7838a5dcb5f96554f7"} Mar 07 15:25:03 crc kubenswrapper[4943]: E0307 15:25:03.921965 4943 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda3cf2fba_57af_4b15_a687_467ccdc1b3c9.slice/crio-3e3497f413bec7b7d1da20dcb6fbe35cb9cbd38b1f511568d80cc0176c2758c9.scope\": RecentStats: unable to find data in memory cache]" Mar 07 15:25:04 crc kubenswrapper[4943]: I0307 15:25:04.188379 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hcm4l" Mar 07 15:25:04 crc kubenswrapper[4943]: I0307 15:25:04.365559 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3cf2fba-57af-4b15-a687-467ccdc1b3c9-utilities\") pod \"a3cf2fba-57af-4b15-a687-467ccdc1b3c9\" (UID: \"a3cf2fba-57af-4b15-a687-467ccdc1b3c9\") " Mar 07 15:25:04 crc kubenswrapper[4943]: I0307 15:25:04.365724 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3cf2fba-57af-4b15-a687-467ccdc1b3c9-catalog-content\") pod \"a3cf2fba-57af-4b15-a687-467ccdc1b3c9\" (UID: \"a3cf2fba-57af-4b15-a687-467ccdc1b3c9\") " Mar 07 15:25:04 crc kubenswrapper[4943]: I0307 15:25:04.365759 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hmbkq\" (UniqueName: \"kubernetes.io/projected/a3cf2fba-57af-4b15-a687-467ccdc1b3c9-kube-api-access-hmbkq\") pod \"a3cf2fba-57af-4b15-a687-467ccdc1b3c9\" (UID: \"a3cf2fba-57af-4b15-a687-467ccdc1b3c9\") " Mar 07 15:25:04 crc kubenswrapper[4943]: I0307 15:25:04.366777 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3cf2fba-57af-4b15-a687-467ccdc1b3c9-utilities" (OuterVolumeSpecName: "utilities") pod "a3cf2fba-57af-4b15-a687-467ccdc1b3c9" (UID: "a3cf2fba-57af-4b15-a687-467ccdc1b3c9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:25:04 crc kubenswrapper[4943]: I0307 15:25:04.372457 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3cf2fba-57af-4b15-a687-467ccdc1b3c9-kube-api-access-hmbkq" (OuterVolumeSpecName: "kube-api-access-hmbkq") pod "a3cf2fba-57af-4b15-a687-467ccdc1b3c9" (UID: "a3cf2fba-57af-4b15-a687-467ccdc1b3c9"). InnerVolumeSpecName "kube-api-access-hmbkq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:25:04 crc kubenswrapper[4943]: I0307 15:25:04.412691 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3cf2fba-57af-4b15-a687-467ccdc1b3c9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a3cf2fba-57af-4b15-a687-467ccdc1b3c9" (UID: "a3cf2fba-57af-4b15-a687-467ccdc1b3c9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:25:04 crc kubenswrapper[4943]: I0307 15:25:04.467394 4943 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3cf2fba-57af-4b15-a687-467ccdc1b3c9-utilities\") on node \"crc\" DevicePath \"\"" Mar 07 15:25:04 crc kubenswrapper[4943]: I0307 15:25:04.467435 4943 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3cf2fba-57af-4b15-a687-467ccdc1b3c9-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 07 15:25:04 crc kubenswrapper[4943]: I0307 15:25:04.467453 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hmbkq\" (UniqueName: \"kubernetes.io/projected/a3cf2fba-57af-4b15-a687-467ccdc1b3c9-kube-api-access-hmbkq\") on node \"crc\" DevicePath \"\"" Mar 07 15:25:04 crc kubenswrapper[4943]: I0307 15:25:04.793346 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hcm4l" Mar 07 15:25:04 crc kubenswrapper[4943]: I0307 15:25:04.794242 4943 generic.go:334] "Generic (PLEG): container finished" podID="a3cf2fba-57af-4b15-a687-467ccdc1b3c9" containerID="3e3497f413bec7b7d1da20dcb6fbe35cb9cbd38b1f511568d80cc0176c2758c9" exitCode=0 Mar 07 15:25:04 crc kubenswrapper[4943]: I0307 15:25:04.794411 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hcm4l" event={"ID":"a3cf2fba-57af-4b15-a687-467ccdc1b3c9","Type":"ContainerDied","Data":"3e3497f413bec7b7d1da20dcb6fbe35cb9cbd38b1f511568d80cc0176c2758c9"} Mar 07 15:25:04 crc kubenswrapper[4943]: I0307 15:25:04.795628 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hcm4l" event={"ID":"a3cf2fba-57af-4b15-a687-467ccdc1b3c9","Type":"ContainerDied","Data":"097debe653a93385dad6c9c840f868a2a62c321c84abda08818a81ea25bec08f"} Mar 07 15:25:04 crc kubenswrapper[4943]: I0307 15:25:04.795672 4943 scope.go:117] "RemoveContainer" containerID="3e3497f413bec7b7d1da20dcb6fbe35cb9cbd38b1f511568d80cc0176c2758c9" Mar 07 15:25:04 crc kubenswrapper[4943]: I0307 15:25:04.807082 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xwkf5" event={"ID":"6634c3f2-9aea-4561-aeaf-0a0f59f3d029","Type":"ContainerStarted","Data":"430ded4e6d73bab28b7b9f137d984763bcbcece46a80d201833dfc759beba1d8"} Mar 07 15:25:04 crc kubenswrapper[4943]: I0307 15:25:04.833726 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xwkf5" podStartSLOduration=3.409211964 podStartE2EDuration="5.833707346s" podCreationTimestamp="2026-03-07 15:24:59 +0000 UTC" firstStartedPulling="2026-03-07 15:25:01.742572204 +0000 UTC m=+2743.694708702" lastFinishedPulling="2026-03-07 15:25:04.167067586 +0000 UTC m=+2746.119204084" observedRunningTime="2026-03-07 15:25:04.829071782 +0000 UTC m=+2746.781208310" watchObservedRunningTime="2026-03-07 15:25:04.833707346 +0000 UTC m=+2746.785843854" Mar 07 15:25:04 crc kubenswrapper[4943]: I0307 15:25:04.862766 4943 scope.go:117] "RemoveContainer" containerID="2d9e0edb3666c367335fa8cc4633436fd14df4fa98ddbf5d2c98262a412ecfe8" Mar 07 15:25:04 crc kubenswrapper[4943]: I0307 15:25:04.862887 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hcm4l"] Mar 07 15:25:04 crc kubenswrapper[4943]: I0307 15:25:04.878566 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hcm4l"] Mar 07 15:25:04 crc kubenswrapper[4943]: I0307 15:25:04.885849 4943 scope.go:117] "RemoveContainer" containerID="b011f925232065feadc544fcf61034af37901c4d8a62ae5bf536f75d02863cba" Mar 07 15:25:04 crc kubenswrapper[4943]: I0307 15:25:04.943546 4943 scope.go:117] "RemoveContainer" containerID="3e3497f413bec7b7d1da20dcb6fbe35cb9cbd38b1f511568d80cc0176c2758c9" Mar 07 15:25:04 crc kubenswrapper[4943]: E0307 15:25:04.944514 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e3497f413bec7b7d1da20dcb6fbe35cb9cbd38b1f511568d80cc0176c2758c9\": container with ID starting with 3e3497f413bec7b7d1da20dcb6fbe35cb9cbd38b1f511568d80cc0176c2758c9 not found: ID does not exist" containerID="3e3497f413bec7b7d1da20dcb6fbe35cb9cbd38b1f511568d80cc0176c2758c9" Mar 07 15:25:04 crc kubenswrapper[4943]: I0307 15:25:04.944571 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e3497f413bec7b7d1da20dcb6fbe35cb9cbd38b1f511568d80cc0176c2758c9"} err="failed to get container status \"3e3497f413bec7b7d1da20dcb6fbe35cb9cbd38b1f511568d80cc0176c2758c9\": rpc error: code = NotFound desc = could not find container \"3e3497f413bec7b7d1da20dcb6fbe35cb9cbd38b1f511568d80cc0176c2758c9\": container with ID starting with 3e3497f413bec7b7d1da20dcb6fbe35cb9cbd38b1f511568d80cc0176c2758c9 not found: ID does not exist" Mar 07 15:25:04 crc kubenswrapper[4943]: I0307 15:25:04.944626 4943 scope.go:117] "RemoveContainer" containerID="2d9e0edb3666c367335fa8cc4633436fd14df4fa98ddbf5d2c98262a412ecfe8" Mar 07 15:25:04 crc kubenswrapper[4943]: E0307 15:25:04.945087 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d9e0edb3666c367335fa8cc4633436fd14df4fa98ddbf5d2c98262a412ecfe8\": container with ID starting with 2d9e0edb3666c367335fa8cc4633436fd14df4fa98ddbf5d2c98262a412ecfe8 not found: ID does not exist" containerID="2d9e0edb3666c367335fa8cc4633436fd14df4fa98ddbf5d2c98262a412ecfe8" Mar 07 15:25:04 crc kubenswrapper[4943]: I0307 15:25:04.945127 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d9e0edb3666c367335fa8cc4633436fd14df4fa98ddbf5d2c98262a412ecfe8"} err="failed to get container status \"2d9e0edb3666c367335fa8cc4633436fd14df4fa98ddbf5d2c98262a412ecfe8\": rpc error: code = NotFound desc = could not find container \"2d9e0edb3666c367335fa8cc4633436fd14df4fa98ddbf5d2c98262a412ecfe8\": container with ID starting with 2d9e0edb3666c367335fa8cc4633436fd14df4fa98ddbf5d2c98262a412ecfe8 not found: ID does not exist" Mar 07 15:25:04 crc kubenswrapper[4943]: I0307 15:25:04.945142 4943 scope.go:117] "RemoveContainer" containerID="b011f925232065feadc544fcf61034af37901c4d8a62ae5bf536f75d02863cba" Mar 07 15:25:04 crc kubenswrapper[4943]: E0307 15:25:04.945474 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b011f925232065feadc544fcf61034af37901c4d8a62ae5bf536f75d02863cba\": container with ID starting with b011f925232065feadc544fcf61034af37901c4d8a62ae5bf536f75d02863cba not found: ID does not exist" containerID="b011f925232065feadc544fcf61034af37901c4d8a62ae5bf536f75d02863cba" Mar 07 15:25:04 crc kubenswrapper[4943]: I0307 15:25:04.945512 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b011f925232065feadc544fcf61034af37901c4d8a62ae5bf536f75d02863cba"} err="failed to get container status \"b011f925232065feadc544fcf61034af37901c4d8a62ae5bf536f75d02863cba\": rpc error: code = NotFound desc = could not find container \"b011f925232065feadc544fcf61034af37901c4d8a62ae5bf536f75d02863cba\": container with ID starting with b011f925232065feadc544fcf61034af37901c4d8a62ae5bf536f75d02863cba not found: ID does not exist" Mar 07 15:25:06 crc kubenswrapper[4943]: I0307 15:25:06.073586 4943 patch_prober.go:28] interesting pod/machine-config-daemon-cpdmm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 07 15:25:06 crc kubenswrapper[4943]: I0307 15:25:06.073642 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 07 15:25:06 crc kubenswrapper[4943]: I0307 15:25:06.073687 4943 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" Mar 07 15:25:06 crc kubenswrapper[4943]: I0307 15:25:06.074257 4943 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"600a051b71dd164e5cc8344d91e9c5c354cdc3c5514cd0b1a1a3e3062c02fe1d"} pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 07 15:25:06 crc kubenswrapper[4943]: I0307 15:25:06.074305 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerName="machine-config-daemon" containerID="cri-o://600a051b71dd164e5cc8344d91e9c5c354cdc3c5514cd0b1a1a3e3062c02fe1d" gracePeriod=600 Mar 07 15:25:06 crc kubenswrapper[4943]: I0307 15:25:06.769095 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3cf2fba-57af-4b15-a687-467ccdc1b3c9" path="/var/lib/kubelet/pods/a3cf2fba-57af-4b15-a687-467ccdc1b3c9/volumes" Mar 07 15:25:06 crc kubenswrapper[4943]: I0307 15:25:06.831067 4943 generic.go:334] "Generic (PLEG): container finished" podID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerID="600a051b71dd164e5cc8344d91e9c5c354cdc3c5514cd0b1a1a3e3062c02fe1d" exitCode=0 Mar 07 15:25:06 crc kubenswrapper[4943]: I0307 15:25:06.831118 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" event={"ID":"4c1f18ea-e557-41fc-95dd-ce4e1722a61e","Type":"ContainerDied","Data":"600a051b71dd164e5cc8344d91e9c5c354cdc3c5514cd0b1a1a3e3062c02fe1d"} Mar 07 15:25:06 crc kubenswrapper[4943]: I0307 15:25:06.831418 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" event={"ID":"4c1f18ea-e557-41fc-95dd-ce4e1722a61e","Type":"ContainerStarted","Data":"bf223df942d1cbe9d18e7fb9dee61069a08bdf260e19d98d98a046cc56a3d466"} Mar 07 15:25:06 crc kubenswrapper[4943]: I0307 15:25:06.831446 4943 scope.go:117] "RemoveContainer" containerID="448308ad6e34ac922d66b3c90e66d0927407a94dd28633ec350502d69e4ca8aa" Mar 07 15:25:09 crc kubenswrapper[4943]: I0307 15:25:09.922472 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xwkf5" Mar 07 15:25:09 crc kubenswrapper[4943]: I0307 15:25:09.923189 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xwkf5" Mar 07 15:25:10 crc kubenswrapper[4943]: I0307 15:25:10.007164 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xwkf5" Mar 07 15:25:10 crc kubenswrapper[4943]: I0307 15:25:10.929255 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xwkf5" Mar 07 15:25:11 crc kubenswrapper[4943]: I0307 15:25:11.584423 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vr2g2"] Mar 07 15:25:11 crc kubenswrapper[4943]: E0307 15:25:11.585185 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3cf2fba-57af-4b15-a687-467ccdc1b3c9" containerName="registry-server" Mar 07 15:25:11 crc kubenswrapper[4943]: I0307 15:25:11.585213 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3cf2fba-57af-4b15-a687-467ccdc1b3c9" containerName="registry-server" Mar 07 15:25:11 crc kubenswrapper[4943]: E0307 15:25:11.585280 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3cf2fba-57af-4b15-a687-467ccdc1b3c9" containerName="extract-utilities" Mar 07 15:25:11 crc kubenswrapper[4943]: I0307 15:25:11.585294 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3cf2fba-57af-4b15-a687-467ccdc1b3c9" containerName="extract-utilities" Mar 07 15:25:11 crc kubenswrapper[4943]: E0307 15:25:11.585327 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3cf2fba-57af-4b15-a687-467ccdc1b3c9" containerName="extract-content" Mar 07 15:25:11 crc kubenswrapper[4943]: I0307 15:25:11.585341 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3cf2fba-57af-4b15-a687-467ccdc1b3c9" containerName="extract-content" Mar 07 15:25:11 crc kubenswrapper[4943]: I0307 15:25:11.585632 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3cf2fba-57af-4b15-a687-467ccdc1b3c9" containerName="registry-server" Mar 07 15:25:11 crc kubenswrapper[4943]: I0307 15:25:11.587453 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vr2g2" Mar 07 15:25:11 crc kubenswrapper[4943]: I0307 15:25:11.597399 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vr2g2"] Mar 07 15:25:11 crc kubenswrapper[4943]: I0307 15:25:11.690372 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9qnd\" (UniqueName: \"kubernetes.io/projected/8370275c-6bbb-44bb-a001-92df7a96f713-kube-api-access-b9qnd\") pod \"redhat-marketplace-vr2g2\" (UID: \"8370275c-6bbb-44bb-a001-92df7a96f713\") " pod="openshift-marketplace/redhat-marketplace-vr2g2" Mar 07 15:25:11 crc kubenswrapper[4943]: I0307 15:25:11.690676 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8370275c-6bbb-44bb-a001-92df7a96f713-catalog-content\") pod \"redhat-marketplace-vr2g2\" (UID: \"8370275c-6bbb-44bb-a001-92df7a96f713\") " pod="openshift-marketplace/redhat-marketplace-vr2g2" Mar 07 15:25:11 crc kubenswrapper[4943]: I0307 15:25:11.690839 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8370275c-6bbb-44bb-a001-92df7a96f713-utilities\") pod \"redhat-marketplace-vr2g2\" (UID: \"8370275c-6bbb-44bb-a001-92df7a96f713\") " pod="openshift-marketplace/redhat-marketplace-vr2g2" Mar 07 15:25:11 crc kubenswrapper[4943]: I0307 15:25:11.792299 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8370275c-6bbb-44bb-a001-92df7a96f713-catalog-content\") pod \"redhat-marketplace-vr2g2\" (UID: \"8370275c-6bbb-44bb-a001-92df7a96f713\") " pod="openshift-marketplace/redhat-marketplace-vr2g2" Mar 07 15:25:11 crc kubenswrapper[4943]: I0307 15:25:11.792370 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8370275c-6bbb-44bb-a001-92df7a96f713-utilities\") pod \"redhat-marketplace-vr2g2\" (UID: \"8370275c-6bbb-44bb-a001-92df7a96f713\") " pod="openshift-marketplace/redhat-marketplace-vr2g2" Mar 07 15:25:11 crc kubenswrapper[4943]: I0307 15:25:11.792966 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8370275c-6bbb-44bb-a001-92df7a96f713-catalog-content\") pod \"redhat-marketplace-vr2g2\" (UID: \"8370275c-6bbb-44bb-a001-92df7a96f713\") " pod="openshift-marketplace/redhat-marketplace-vr2g2" Mar 07 15:25:11 crc kubenswrapper[4943]: I0307 15:25:11.792977 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8370275c-6bbb-44bb-a001-92df7a96f713-utilities\") pod \"redhat-marketplace-vr2g2\" (UID: \"8370275c-6bbb-44bb-a001-92df7a96f713\") " pod="openshift-marketplace/redhat-marketplace-vr2g2" Mar 07 15:25:11 crc kubenswrapper[4943]: I0307 15:25:11.793134 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9qnd\" (UniqueName: \"kubernetes.io/projected/8370275c-6bbb-44bb-a001-92df7a96f713-kube-api-access-b9qnd\") pod \"redhat-marketplace-vr2g2\" (UID: \"8370275c-6bbb-44bb-a001-92df7a96f713\") " pod="openshift-marketplace/redhat-marketplace-vr2g2" Mar 07 15:25:11 crc kubenswrapper[4943]: I0307 15:25:11.833329 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9qnd\" (UniqueName: \"kubernetes.io/projected/8370275c-6bbb-44bb-a001-92df7a96f713-kube-api-access-b9qnd\") pod \"redhat-marketplace-vr2g2\" (UID: \"8370275c-6bbb-44bb-a001-92df7a96f713\") " pod="openshift-marketplace/redhat-marketplace-vr2g2" Mar 07 15:25:11 crc kubenswrapper[4943]: I0307 15:25:11.963167 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vr2g2" Mar 07 15:25:12 crc kubenswrapper[4943]: W0307 15:25:12.422122 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8370275c_6bbb_44bb_a001_92df7a96f713.slice/crio-b74f8c9af35a4b07df62c7eac0d2e70da53629ae20220b4a681218315b574757 WatchSource:0}: Error finding container b74f8c9af35a4b07df62c7eac0d2e70da53629ae20220b4a681218315b574757: Status 404 returned error can't find the container with id b74f8c9af35a4b07df62c7eac0d2e70da53629ae20220b4a681218315b574757 Mar 07 15:25:12 crc kubenswrapper[4943]: I0307 15:25:12.422486 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vr2g2"] Mar 07 15:25:12 crc kubenswrapper[4943]: I0307 15:25:12.891994 4943 generic.go:334] "Generic (PLEG): container finished" podID="8370275c-6bbb-44bb-a001-92df7a96f713" containerID="0e0a5412c504ca54e6a510cfe1c27fd1419f4667ae79e7b35a590498ff1cca5b" exitCode=0 Mar 07 15:25:12 crc kubenswrapper[4943]: I0307 15:25:12.892060 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vr2g2" event={"ID":"8370275c-6bbb-44bb-a001-92df7a96f713","Type":"ContainerDied","Data":"0e0a5412c504ca54e6a510cfe1c27fd1419f4667ae79e7b35a590498ff1cca5b"} Mar 07 15:25:12 crc kubenswrapper[4943]: I0307 15:25:12.892105 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vr2g2" event={"ID":"8370275c-6bbb-44bb-a001-92df7a96f713","Type":"ContainerStarted","Data":"b74f8c9af35a4b07df62c7eac0d2e70da53629ae20220b4a681218315b574757"} Mar 07 15:25:15 crc kubenswrapper[4943]: I0307 15:25:15.943773 4943 generic.go:334] "Generic (PLEG): container finished" podID="8370275c-6bbb-44bb-a001-92df7a96f713" containerID="12c0c9d066a7f463f72a1a6773c05191a4633115de3713fbd54ff00ce0c62c1f" exitCode=0 Mar 07 15:25:15 crc kubenswrapper[4943]: I0307 15:25:15.943820 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vr2g2" event={"ID":"8370275c-6bbb-44bb-a001-92df7a96f713","Type":"ContainerDied","Data":"12c0c9d066a7f463f72a1a6773c05191a4633115de3713fbd54ff00ce0c62c1f"} Mar 07 15:25:16 crc kubenswrapper[4943]: I0307 15:25:16.954083 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vr2g2" event={"ID":"8370275c-6bbb-44bb-a001-92df7a96f713","Type":"ContainerStarted","Data":"0ff1100a027ad364dd091fe5ec0058dc65a914bec8d2765897009a628312fedf"} Mar 07 15:25:16 crc kubenswrapper[4943]: I0307 15:25:16.982340 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vr2g2" podStartSLOduration=2.542948555 podStartE2EDuration="5.982313602s" podCreationTimestamp="2026-03-07 15:25:11 +0000 UTC" firstStartedPulling="2026-03-07 15:25:12.895009674 +0000 UTC m=+2754.847146202" lastFinishedPulling="2026-03-07 15:25:16.334374721 +0000 UTC m=+2758.286511249" observedRunningTime="2026-03-07 15:25:16.97246094 +0000 UTC m=+2758.924597438" watchObservedRunningTime="2026-03-07 15:25:16.982313602 +0000 UTC m=+2758.934450120" Mar 07 15:25:17 crc kubenswrapper[4943]: I0307 15:25:17.978618 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xwkf5"] Mar 07 15:25:17 crc kubenswrapper[4943]: I0307 15:25:17.978920 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xwkf5" podUID="6634c3f2-9aea-4561-aeaf-0a0f59f3d029" containerName="registry-server" containerID="cri-o://430ded4e6d73bab28b7b9f137d984763bcbcece46a80d201833dfc759beba1d8" gracePeriod=2 Mar 07 15:25:18 crc kubenswrapper[4943]: I0307 15:25:18.496685 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xwkf5" Mar 07 15:25:18 crc kubenswrapper[4943]: I0307 15:25:18.601839 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6634c3f2-9aea-4561-aeaf-0a0f59f3d029-utilities\") pod \"6634c3f2-9aea-4561-aeaf-0a0f59f3d029\" (UID: \"6634c3f2-9aea-4561-aeaf-0a0f59f3d029\") " Mar 07 15:25:18 crc kubenswrapper[4943]: I0307 15:25:18.601965 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lh5q4\" (UniqueName: \"kubernetes.io/projected/6634c3f2-9aea-4561-aeaf-0a0f59f3d029-kube-api-access-lh5q4\") pod \"6634c3f2-9aea-4561-aeaf-0a0f59f3d029\" (UID: \"6634c3f2-9aea-4561-aeaf-0a0f59f3d029\") " Mar 07 15:25:18 crc kubenswrapper[4943]: I0307 15:25:18.602040 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6634c3f2-9aea-4561-aeaf-0a0f59f3d029-catalog-content\") pod \"6634c3f2-9aea-4561-aeaf-0a0f59f3d029\" (UID: \"6634c3f2-9aea-4561-aeaf-0a0f59f3d029\") " Mar 07 15:25:18 crc kubenswrapper[4943]: I0307 15:25:18.603897 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6634c3f2-9aea-4561-aeaf-0a0f59f3d029-utilities" (OuterVolumeSpecName: "utilities") pod "6634c3f2-9aea-4561-aeaf-0a0f59f3d029" (UID: "6634c3f2-9aea-4561-aeaf-0a0f59f3d029"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:25:18 crc kubenswrapper[4943]: I0307 15:25:18.610309 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6634c3f2-9aea-4561-aeaf-0a0f59f3d029-kube-api-access-lh5q4" (OuterVolumeSpecName: "kube-api-access-lh5q4") pod "6634c3f2-9aea-4561-aeaf-0a0f59f3d029" (UID: "6634c3f2-9aea-4561-aeaf-0a0f59f3d029"). InnerVolumeSpecName "kube-api-access-lh5q4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:25:18 crc kubenswrapper[4943]: I0307 15:25:18.673830 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6634c3f2-9aea-4561-aeaf-0a0f59f3d029-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6634c3f2-9aea-4561-aeaf-0a0f59f3d029" (UID: "6634c3f2-9aea-4561-aeaf-0a0f59f3d029"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:25:18 crc kubenswrapper[4943]: I0307 15:25:18.703861 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lh5q4\" (UniqueName: \"kubernetes.io/projected/6634c3f2-9aea-4561-aeaf-0a0f59f3d029-kube-api-access-lh5q4\") on node \"crc\" DevicePath \"\"" Mar 07 15:25:18 crc kubenswrapper[4943]: I0307 15:25:18.703900 4943 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6634c3f2-9aea-4561-aeaf-0a0f59f3d029-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 07 15:25:18 crc kubenswrapper[4943]: I0307 15:25:18.703910 4943 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6634c3f2-9aea-4561-aeaf-0a0f59f3d029-utilities\") on node \"crc\" DevicePath \"\"" Mar 07 15:25:18 crc kubenswrapper[4943]: I0307 15:25:18.977896 4943 generic.go:334] "Generic (PLEG): container finished" podID="6634c3f2-9aea-4561-aeaf-0a0f59f3d029" containerID="430ded4e6d73bab28b7b9f137d984763bcbcece46a80d201833dfc759beba1d8" exitCode=0 Mar 07 15:25:18 crc kubenswrapper[4943]: I0307 15:25:18.977968 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xwkf5" event={"ID":"6634c3f2-9aea-4561-aeaf-0a0f59f3d029","Type":"ContainerDied","Data":"430ded4e6d73bab28b7b9f137d984763bcbcece46a80d201833dfc759beba1d8"} Mar 07 15:25:18 crc kubenswrapper[4943]: I0307 15:25:18.978000 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xwkf5" event={"ID":"6634c3f2-9aea-4561-aeaf-0a0f59f3d029","Type":"ContainerDied","Data":"9119c722fe76c1c60d7e196d8c50e6b5165640ee23c605cd9f2992498fd77bc7"} Mar 07 15:25:18 crc kubenswrapper[4943]: I0307 15:25:18.978009 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xwkf5" Mar 07 15:25:18 crc kubenswrapper[4943]: I0307 15:25:18.978020 4943 scope.go:117] "RemoveContainer" containerID="430ded4e6d73bab28b7b9f137d984763bcbcece46a80d201833dfc759beba1d8" Mar 07 15:25:19 crc kubenswrapper[4943]: I0307 15:25:19.001372 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xwkf5"] Mar 07 15:25:19 crc kubenswrapper[4943]: I0307 15:25:19.010174 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xwkf5"] Mar 07 15:25:19 crc kubenswrapper[4943]: I0307 15:25:19.012754 4943 scope.go:117] "RemoveContainer" containerID="7c985d0c6b4bd903a0cac2873eb8b4e1364e219c1f067e7838a5dcb5f96554f7" Mar 07 15:25:19 crc kubenswrapper[4943]: I0307 15:25:19.042285 4943 scope.go:117] "RemoveContainer" containerID="dcded2899efb30f8ea8b658cfc7fb98b55595f42e18c33263a9ced74b34c8151" Mar 07 15:25:19 crc kubenswrapper[4943]: I0307 15:25:19.094289 4943 scope.go:117] "RemoveContainer" containerID="430ded4e6d73bab28b7b9f137d984763bcbcece46a80d201833dfc759beba1d8" Mar 07 15:25:19 crc kubenswrapper[4943]: E0307 15:25:19.094778 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"430ded4e6d73bab28b7b9f137d984763bcbcece46a80d201833dfc759beba1d8\": container with ID starting with 430ded4e6d73bab28b7b9f137d984763bcbcece46a80d201833dfc759beba1d8 not found: ID does not exist" containerID="430ded4e6d73bab28b7b9f137d984763bcbcece46a80d201833dfc759beba1d8" Mar 07 15:25:19 crc kubenswrapper[4943]: I0307 15:25:19.094823 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"430ded4e6d73bab28b7b9f137d984763bcbcece46a80d201833dfc759beba1d8"} err="failed to get container status \"430ded4e6d73bab28b7b9f137d984763bcbcece46a80d201833dfc759beba1d8\": rpc error: code = NotFound desc = could not find container \"430ded4e6d73bab28b7b9f137d984763bcbcece46a80d201833dfc759beba1d8\": container with ID starting with 430ded4e6d73bab28b7b9f137d984763bcbcece46a80d201833dfc759beba1d8 not found: ID does not exist" Mar 07 15:25:19 crc kubenswrapper[4943]: I0307 15:25:19.094847 4943 scope.go:117] "RemoveContainer" containerID="7c985d0c6b4bd903a0cac2873eb8b4e1364e219c1f067e7838a5dcb5f96554f7" Mar 07 15:25:19 crc kubenswrapper[4943]: E0307 15:25:19.095467 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c985d0c6b4bd903a0cac2873eb8b4e1364e219c1f067e7838a5dcb5f96554f7\": container with ID starting with 7c985d0c6b4bd903a0cac2873eb8b4e1364e219c1f067e7838a5dcb5f96554f7 not found: ID does not exist" containerID="7c985d0c6b4bd903a0cac2873eb8b4e1364e219c1f067e7838a5dcb5f96554f7" Mar 07 15:25:19 crc kubenswrapper[4943]: I0307 15:25:19.095634 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c985d0c6b4bd903a0cac2873eb8b4e1364e219c1f067e7838a5dcb5f96554f7"} err="failed to get container status \"7c985d0c6b4bd903a0cac2873eb8b4e1364e219c1f067e7838a5dcb5f96554f7\": rpc error: code = NotFound desc = could not find container \"7c985d0c6b4bd903a0cac2873eb8b4e1364e219c1f067e7838a5dcb5f96554f7\": container with ID starting with 7c985d0c6b4bd903a0cac2873eb8b4e1364e219c1f067e7838a5dcb5f96554f7 not found: ID does not exist" Mar 07 15:25:19 crc kubenswrapper[4943]: I0307 15:25:19.095799 4943 scope.go:117] "RemoveContainer" containerID="dcded2899efb30f8ea8b658cfc7fb98b55595f42e18c33263a9ced74b34c8151" Mar 07 15:25:19 crc kubenswrapper[4943]: E0307 15:25:19.096472 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dcded2899efb30f8ea8b658cfc7fb98b55595f42e18c33263a9ced74b34c8151\": container with ID starting with dcded2899efb30f8ea8b658cfc7fb98b55595f42e18c33263a9ced74b34c8151 not found: ID does not exist" containerID="dcded2899efb30f8ea8b658cfc7fb98b55595f42e18c33263a9ced74b34c8151" Mar 07 15:25:19 crc kubenswrapper[4943]: I0307 15:25:19.096542 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dcded2899efb30f8ea8b658cfc7fb98b55595f42e18c33263a9ced74b34c8151"} err="failed to get container status \"dcded2899efb30f8ea8b658cfc7fb98b55595f42e18c33263a9ced74b34c8151\": rpc error: code = NotFound desc = could not find container \"dcded2899efb30f8ea8b658cfc7fb98b55595f42e18c33263a9ced74b34c8151\": container with ID starting with dcded2899efb30f8ea8b658cfc7fb98b55595f42e18c33263a9ced74b34c8151 not found: ID does not exist" Mar 07 15:25:20 crc kubenswrapper[4943]: I0307 15:25:20.771295 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6634c3f2-9aea-4561-aeaf-0a0f59f3d029" path="/var/lib/kubelet/pods/6634c3f2-9aea-4561-aeaf-0a0f59f3d029/volumes" Mar 07 15:25:21 crc kubenswrapper[4943]: I0307 15:25:21.963629 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vr2g2" Mar 07 15:25:21 crc kubenswrapper[4943]: I0307 15:25:21.963977 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vr2g2" Mar 07 15:25:22 crc kubenswrapper[4943]: I0307 15:25:22.052737 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vr2g2" Mar 07 15:25:22 crc kubenswrapper[4943]: I0307 15:25:22.109349 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vr2g2" Mar 07 15:25:24 crc kubenswrapper[4943]: I0307 15:25:24.583109 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vr2g2"] Mar 07 15:25:24 crc kubenswrapper[4943]: I0307 15:25:24.583843 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-vr2g2" podUID="8370275c-6bbb-44bb-a001-92df7a96f713" containerName="registry-server" containerID="cri-o://0ff1100a027ad364dd091fe5ec0058dc65a914bec8d2765897009a628312fedf" gracePeriod=2 Mar 07 15:25:25 crc kubenswrapper[4943]: I0307 15:25:25.076800 4943 generic.go:334] "Generic (PLEG): container finished" podID="8370275c-6bbb-44bb-a001-92df7a96f713" containerID="0ff1100a027ad364dd091fe5ec0058dc65a914bec8d2765897009a628312fedf" exitCode=0 Mar 07 15:25:25 crc kubenswrapper[4943]: I0307 15:25:25.076872 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vr2g2" event={"ID":"8370275c-6bbb-44bb-a001-92df7a96f713","Type":"ContainerDied","Data":"0ff1100a027ad364dd091fe5ec0058dc65a914bec8d2765897009a628312fedf"} Mar 07 15:25:25 crc kubenswrapper[4943]: I0307 15:25:25.077178 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vr2g2" event={"ID":"8370275c-6bbb-44bb-a001-92df7a96f713","Type":"ContainerDied","Data":"b74f8c9af35a4b07df62c7eac0d2e70da53629ae20220b4a681218315b574757"} Mar 07 15:25:25 crc kubenswrapper[4943]: I0307 15:25:25.077191 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b74f8c9af35a4b07df62c7eac0d2e70da53629ae20220b4a681218315b574757" Mar 07 15:25:25 crc kubenswrapper[4943]: I0307 15:25:25.105366 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vr2g2" Mar 07 15:25:25 crc kubenswrapper[4943]: I0307 15:25:25.211147 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b9qnd\" (UniqueName: \"kubernetes.io/projected/8370275c-6bbb-44bb-a001-92df7a96f713-kube-api-access-b9qnd\") pod \"8370275c-6bbb-44bb-a001-92df7a96f713\" (UID: \"8370275c-6bbb-44bb-a001-92df7a96f713\") " Mar 07 15:25:25 crc kubenswrapper[4943]: I0307 15:25:25.211371 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8370275c-6bbb-44bb-a001-92df7a96f713-catalog-content\") pod \"8370275c-6bbb-44bb-a001-92df7a96f713\" (UID: \"8370275c-6bbb-44bb-a001-92df7a96f713\") " Mar 07 15:25:25 crc kubenswrapper[4943]: I0307 15:25:25.211475 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8370275c-6bbb-44bb-a001-92df7a96f713-utilities\") pod \"8370275c-6bbb-44bb-a001-92df7a96f713\" (UID: \"8370275c-6bbb-44bb-a001-92df7a96f713\") " Mar 07 15:25:25 crc kubenswrapper[4943]: I0307 15:25:25.213920 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8370275c-6bbb-44bb-a001-92df7a96f713-utilities" (OuterVolumeSpecName: "utilities") pod "8370275c-6bbb-44bb-a001-92df7a96f713" (UID: "8370275c-6bbb-44bb-a001-92df7a96f713"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:25:25 crc kubenswrapper[4943]: I0307 15:25:25.219218 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8370275c-6bbb-44bb-a001-92df7a96f713-kube-api-access-b9qnd" (OuterVolumeSpecName: "kube-api-access-b9qnd") pod "8370275c-6bbb-44bb-a001-92df7a96f713" (UID: "8370275c-6bbb-44bb-a001-92df7a96f713"). InnerVolumeSpecName "kube-api-access-b9qnd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:25:25 crc kubenswrapper[4943]: I0307 15:25:25.239701 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8370275c-6bbb-44bb-a001-92df7a96f713-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8370275c-6bbb-44bb-a001-92df7a96f713" (UID: "8370275c-6bbb-44bb-a001-92df7a96f713"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 07 15:25:25 crc kubenswrapper[4943]: I0307 15:25:25.313838 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b9qnd\" (UniqueName: \"kubernetes.io/projected/8370275c-6bbb-44bb-a001-92df7a96f713-kube-api-access-b9qnd\") on node \"crc\" DevicePath \"\"" Mar 07 15:25:25 crc kubenswrapper[4943]: I0307 15:25:25.313880 4943 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8370275c-6bbb-44bb-a001-92df7a96f713-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 07 15:25:25 crc kubenswrapper[4943]: I0307 15:25:25.313890 4943 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8370275c-6bbb-44bb-a001-92df7a96f713-utilities\") on node \"crc\" DevicePath \"\"" Mar 07 15:25:26 crc kubenswrapper[4943]: I0307 15:25:26.086313 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vr2g2" Mar 07 15:25:26 crc kubenswrapper[4943]: I0307 15:25:26.123404 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vr2g2"] Mar 07 15:25:26 crc kubenswrapper[4943]: I0307 15:25:26.129170 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-vr2g2"] Mar 07 15:25:26 crc kubenswrapper[4943]: I0307 15:25:26.701848 4943 scope.go:117] "RemoveContainer" containerID="368159a0e84aef45760627cc1b2cc7fd358c90db1a223bd795d6690286a4f335" Mar 07 15:25:26 crc kubenswrapper[4943]: I0307 15:25:26.758652 4943 scope.go:117] "RemoveContainer" containerID="4308b9d76705a4bd10464643ad4935fbd249c9d3db40036c52aa167d82157339" Mar 07 15:25:26 crc kubenswrapper[4943]: I0307 15:25:26.772200 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8370275c-6bbb-44bb-a001-92df7a96f713" path="/var/lib/kubelet/pods/8370275c-6bbb-44bb-a001-92df7a96f713/volumes" Mar 07 15:25:26 crc kubenswrapper[4943]: I0307 15:25:26.803534 4943 scope.go:117] "RemoveContainer" containerID="056e6ef8ca10962682d5005ded8199f65bab3faf1b66a8bfb7135c94cbd0cb1a" Mar 07 15:25:26 crc kubenswrapper[4943]: I0307 15:25:26.842838 4943 scope.go:117] "RemoveContainer" containerID="dba7383a12bbbc41263368a88607ad7cdea69d0d7123bff82132321a508f4ffc" Mar 07 15:25:26 crc kubenswrapper[4943]: I0307 15:25:26.875470 4943 scope.go:117] "RemoveContainer" containerID="9ef46553800be268c4887197ec8125dc8e3308f4f5928d408f0b7b257fc240f9" Mar 07 15:25:26 crc kubenswrapper[4943]: I0307 15:25:26.902516 4943 scope.go:117] "RemoveContainer" containerID="e955e20dafce7619e536a75178762f613fe0332cc33685e03107bd475fa4d6c0" Mar 07 15:25:26 crc kubenswrapper[4943]: I0307 15:25:26.943250 4943 scope.go:117] "RemoveContainer" containerID="05d36441cf236aeb49863eea9d387746e7ed71f540dd8f004d6b7dadb5fc62aa" Mar 07 15:25:26 crc kubenswrapper[4943]: I0307 15:25:26.969634 4943 scope.go:117] "RemoveContainer" containerID="dfd77d95691d128ef6b7a43d462d7203fe5ff87140cdd17cb672078fd27f5293" Mar 07 15:26:00 crc kubenswrapper[4943]: I0307 15:26:00.431598 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29548286-n7mpw"] Mar 07 15:26:00 crc kubenswrapper[4943]: E0307 15:26:00.432660 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6634c3f2-9aea-4561-aeaf-0a0f59f3d029" containerName="extract-content" Mar 07 15:26:00 crc kubenswrapper[4943]: I0307 15:26:00.432682 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="6634c3f2-9aea-4561-aeaf-0a0f59f3d029" containerName="extract-content" Mar 07 15:26:00 crc kubenswrapper[4943]: E0307 15:26:00.432705 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6634c3f2-9aea-4561-aeaf-0a0f59f3d029" containerName="registry-server" Mar 07 15:26:00 crc kubenswrapper[4943]: I0307 15:26:00.432718 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="6634c3f2-9aea-4561-aeaf-0a0f59f3d029" containerName="registry-server" Mar 07 15:26:00 crc kubenswrapper[4943]: E0307 15:26:00.432740 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8370275c-6bbb-44bb-a001-92df7a96f713" containerName="extract-utilities" Mar 07 15:26:00 crc kubenswrapper[4943]: I0307 15:26:00.432756 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="8370275c-6bbb-44bb-a001-92df7a96f713" containerName="extract-utilities" Mar 07 15:26:00 crc kubenswrapper[4943]: E0307 15:26:00.432790 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6634c3f2-9aea-4561-aeaf-0a0f59f3d029" containerName="extract-utilities" Mar 07 15:26:00 crc kubenswrapper[4943]: I0307 15:26:00.432805 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="6634c3f2-9aea-4561-aeaf-0a0f59f3d029" containerName="extract-utilities" Mar 07 15:26:00 crc kubenswrapper[4943]: E0307 15:26:00.432837 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8370275c-6bbb-44bb-a001-92df7a96f713" containerName="extract-content" Mar 07 15:26:00 crc kubenswrapper[4943]: I0307 15:26:00.432853 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="8370275c-6bbb-44bb-a001-92df7a96f713" containerName="extract-content" Mar 07 15:26:00 crc kubenswrapper[4943]: E0307 15:26:00.432905 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8370275c-6bbb-44bb-a001-92df7a96f713" containerName="registry-server" Mar 07 15:26:00 crc kubenswrapper[4943]: I0307 15:26:00.432921 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="8370275c-6bbb-44bb-a001-92df7a96f713" containerName="registry-server" Mar 07 15:26:00 crc kubenswrapper[4943]: I0307 15:26:00.433256 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="8370275c-6bbb-44bb-a001-92df7a96f713" containerName="registry-server" Mar 07 15:26:00 crc kubenswrapper[4943]: I0307 15:26:00.433297 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="6634c3f2-9aea-4561-aeaf-0a0f59f3d029" containerName="registry-server" Mar 07 15:26:00 crc kubenswrapper[4943]: I0307 15:26:00.434206 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548286-n7mpw" Mar 07 15:26:00 crc kubenswrapper[4943]: I0307 15:26:00.436239 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-qqlk5" Mar 07 15:26:00 crc kubenswrapper[4943]: I0307 15:26:00.436324 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 07 15:26:00 crc kubenswrapper[4943]: I0307 15:26:00.443670 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29548286-n7mpw"] Mar 07 15:26:00 crc kubenswrapper[4943]: I0307 15:26:00.446767 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 07 15:26:00 crc kubenswrapper[4943]: I0307 15:26:00.448320 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crp2s\" (UniqueName: \"kubernetes.io/projected/f80b3afc-423f-46ae-a1f0-ad3ed932bc40-kube-api-access-crp2s\") pod \"auto-csr-approver-29548286-n7mpw\" (UID: \"f80b3afc-423f-46ae-a1f0-ad3ed932bc40\") " pod="openshift-infra/auto-csr-approver-29548286-n7mpw" Mar 07 15:26:00 crc kubenswrapper[4943]: I0307 15:26:00.549883 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crp2s\" (UniqueName: \"kubernetes.io/projected/f80b3afc-423f-46ae-a1f0-ad3ed932bc40-kube-api-access-crp2s\") pod \"auto-csr-approver-29548286-n7mpw\" (UID: \"f80b3afc-423f-46ae-a1f0-ad3ed932bc40\") " pod="openshift-infra/auto-csr-approver-29548286-n7mpw" Mar 07 15:26:00 crc kubenswrapper[4943]: I0307 15:26:00.576416 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crp2s\" (UniqueName: \"kubernetes.io/projected/f80b3afc-423f-46ae-a1f0-ad3ed932bc40-kube-api-access-crp2s\") pod \"auto-csr-approver-29548286-n7mpw\" (UID: \"f80b3afc-423f-46ae-a1f0-ad3ed932bc40\") " pod="openshift-infra/auto-csr-approver-29548286-n7mpw" Mar 07 15:26:00 crc kubenswrapper[4943]: I0307 15:26:00.753070 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548286-n7mpw" Mar 07 15:26:01 crc kubenswrapper[4943]: I0307 15:26:01.117574 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29548286-n7mpw"] Mar 07 15:26:01 crc kubenswrapper[4943]: I0307 15:26:01.491797 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548286-n7mpw" event={"ID":"f80b3afc-423f-46ae-a1f0-ad3ed932bc40","Type":"ContainerStarted","Data":"55b7605f4133eee86df1c4715a21cdcc62eebd8dd7655c67b130bdb60ee662c4"} Mar 07 15:26:03 crc kubenswrapper[4943]: I0307 15:26:03.514375 4943 generic.go:334] "Generic (PLEG): container finished" podID="f80b3afc-423f-46ae-a1f0-ad3ed932bc40" containerID="810f4e5c95a1d92096398be7606f21c8029e2273999ddbef69b853b1507874fc" exitCode=0 Mar 07 15:26:03 crc kubenswrapper[4943]: I0307 15:26:03.514442 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548286-n7mpw" event={"ID":"f80b3afc-423f-46ae-a1f0-ad3ed932bc40","Type":"ContainerDied","Data":"810f4e5c95a1d92096398be7606f21c8029e2273999ddbef69b853b1507874fc"} Mar 07 15:26:04 crc kubenswrapper[4943]: I0307 15:26:04.938365 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548286-n7mpw" Mar 07 15:26:05 crc kubenswrapper[4943]: I0307 15:26:05.124406 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-crp2s\" (UniqueName: \"kubernetes.io/projected/f80b3afc-423f-46ae-a1f0-ad3ed932bc40-kube-api-access-crp2s\") pod \"f80b3afc-423f-46ae-a1f0-ad3ed932bc40\" (UID: \"f80b3afc-423f-46ae-a1f0-ad3ed932bc40\") " Mar 07 15:26:05 crc kubenswrapper[4943]: I0307 15:26:05.134354 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f80b3afc-423f-46ae-a1f0-ad3ed932bc40-kube-api-access-crp2s" (OuterVolumeSpecName: "kube-api-access-crp2s") pod "f80b3afc-423f-46ae-a1f0-ad3ed932bc40" (UID: "f80b3afc-423f-46ae-a1f0-ad3ed932bc40"). InnerVolumeSpecName "kube-api-access-crp2s". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:26:05 crc kubenswrapper[4943]: I0307 15:26:05.226178 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-crp2s\" (UniqueName: \"kubernetes.io/projected/f80b3afc-423f-46ae-a1f0-ad3ed932bc40-kube-api-access-crp2s\") on node \"crc\" DevicePath \"\"" Mar 07 15:26:05 crc kubenswrapper[4943]: I0307 15:26:05.542923 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548286-n7mpw" event={"ID":"f80b3afc-423f-46ae-a1f0-ad3ed932bc40","Type":"ContainerDied","Data":"55b7605f4133eee86df1c4715a21cdcc62eebd8dd7655c67b130bdb60ee662c4"} Mar 07 15:26:05 crc kubenswrapper[4943]: I0307 15:26:05.543346 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="55b7605f4133eee86df1c4715a21cdcc62eebd8dd7655c67b130bdb60ee662c4" Mar 07 15:26:05 crc kubenswrapper[4943]: I0307 15:26:05.542999 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548286-n7mpw" Mar 07 15:26:06 crc kubenswrapper[4943]: I0307 15:26:06.027387 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29548280-q7zxj"] Mar 07 15:26:06 crc kubenswrapper[4943]: I0307 15:26:06.036350 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29548280-q7zxj"] Mar 07 15:26:06 crc kubenswrapper[4943]: I0307 15:26:06.772057 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a63975e7-b7d9-4ab0-aee1-fcdd68e805cf" path="/var/lib/kubelet/pods/a63975e7-b7d9-4ab0-aee1-fcdd68e805cf/volumes" Mar 07 15:26:27 crc kubenswrapper[4943]: I0307 15:26:27.238005 4943 scope.go:117] "RemoveContainer" containerID="d442999f3351d46e800c4d2e91511ae83e9ab8d0dd60e769149db07b2ca7024e" Mar 07 15:26:27 crc kubenswrapper[4943]: I0307 15:26:27.306447 4943 scope.go:117] "RemoveContainer" containerID="96ee9016de2326c8c0501131b1ba67d2ea036f5a37256c42afb745bfadb03d25" Mar 07 15:26:27 crc kubenswrapper[4943]: I0307 15:26:27.353452 4943 scope.go:117] "RemoveContainer" containerID="e4cee84db2592c06ad43026174ee52161982e207a05d76fb520c5751c15115a7" Mar 07 15:26:27 crc kubenswrapper[4943]: I0307 15:26:27.396919 4943 scope.go:117] "RemoveContainer" containerID="27a0e2a50f061bddad7d687ea67d3df7313113490b76d8d044db002cc063ca08" Mar 07 15:26:27 crc kubenswrapper[4943]: I0307 15:26:27.422289 4943 scope.go:117] "RemoveContainer" containerID="a928e8c9e1805599cfcc0243ab9d559652e0cc68f084c5ad20b90e7af50975ca" Mar 07 15:26:27 crc kubenswrapper[4943]: I0307 15:26:27.458070 4943 scope.go:117] "RemoveContainer" containerID="a15c43c8f498a82a124d4b4e6c272172c48df774a1fa978357f13c98aa911040" Mar 07 15:26:27 crc kubenswrapper[4943]: I0307 15:26:27.497145 4943 scope.go:117] "RemoveContainer" containerID="9aa118624744f146b373f21bda642afdae46954f0c5a6428e17c76aaf87843a9" Mar 07 15:26:27 crc kubenswrapper[4943]: I0307 15:26:27.519491 4943 scope.go:117] "RemoveContainer" containerID="a87fab2fc350e01e1c8b6cb2c885db9037b0892ac597bcda669e2b0df98b51eb" Mar 07 15:27:06 crc kubenswrapper[4943]: I0307 15:27:06.073592 4943 patch_prober.go:28] interesting pod/machine-config-daemon-cpdmm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 07 15:27:06 crc kubenswrapper[4943]: I0307 15:27:06.074144 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 07 15:27:36 crc kubenswrapper[4943]: I0307 15:27:36.074072 4943 patch_prober.go:28] interesting pod/machine-config-daemon-cpdmm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 07 15:27:36 crc kubenswrapper[4943]: I0307 15:27:36.074737 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 07 15:28:00 crc kubenswrapper[4943]: I0307 15:28:00.156354 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29548288-69d2g"] Mar 07 15:28:00 crc kubenswrapper[4943]: E0307 15:28:00.158536 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f80b3afc-423f-46ae-a1f0-ad3ed932bc40" containerName="oc" Mar 07 15:28:00 crc kubenswrapper[4943]: I0307 15:28:00.158564 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="f80b3afc-423f-46ae-a1f0-ad3ed932bc40" containerName="oc" Mar 07 15:28:00 crc kubenswrapper[4943]: I0307 15:28:00.158817 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="f80b3afc-423f-46ae-a1f0-ad3ed932bc40" containerName="oc" Mar 07 15:28:00 crc kubenswrapper[4943]: I0307 15:28:00.159542 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548288-69d2g" Mar 07 15:28:00 crc kubenswrapper[4943]: I0307 15:28:00.162697 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-qqlk5" Mar 07 15:28:00 crc kubenswrapper[4943]: I0307 15:28:00.163143 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 07 15:28:00 crc kubenswrapper[4943]: I0307 15:28:00.164541 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 07 15:28:00 crc kubenswrapper[4943]: I0307 15:28:00.168303 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29548288-69d2g"] Mar 07 15:28:00 crc kubenswrapper[4943]: I0307 15:28:00.340652 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55j48\" (UniqueName: \"kubernetes.io/projected/9d10be01-3397-4d41-9c9e-c6c00c4728be-kube-api-access-55j48\") pod \"auto-csr-approver-29548288-69d2g\" (UID: \"9d10be01-3397-4d41-9c9e-c6c00c4728be\") " pod="openshift-infra/auto-csr-approver-29548288-69d2g" Mar 07 15:28:00 crc kubenswrapper[4943]: I0307 15:28:00.442501 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55j48\" (UniqueName: \"kubernetes.io/projected/9d10be01-3397-4d41-9c9e-c6c00c4728be-kube-api-access-55j48\") pod \"auto-csr-approver-29548288-69d2g\" (UID: \"9d10be01-3397-4d41-9c9e-c6c00c4728be\") " pod="openshift-infra/auto-csr-approver-29548288-69d2g" Mar 07 15:28:00 crc kubenswrapper[4943]: I0307 15:28:00.468367 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55j48\" (UniqueName: \"kubernetes.io/projected/9d10be01-3397-4d41-9c9e-c6c00c4728be-kube-api-access-55j48\") pod \"auto-csr-approver-29548288-69d2g\" (UID: \"9d10be01-3397-4d41-9c9e-c6c00c4728be\") " pod="openshift-infra/auto-csr-approver-29548288-69d2g" Mar 07 15:28:00 crc kubenswrapper[4943]: I0307 15:28:00.516316 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548288-69d2g" Mar 07 15:28:00 crc kubenswrapper[4943]: I0307 15:28:00.971597 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29548288-69d2g"] Mar 07 15:28:01 crc kubenswrapper[4943]: I0307 15:28:01.722032 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548288-69d2g" event={"ID":"9d10be01-3397-4d41-9c9e-c6c00c4728be","Type":"ContainerStarted","Data":"c8d2f65c1c1a2b81a76b60eee905bc1223ce50d328588ed7555980de864f50c1"} Mar 07 15:28:02 crc kubenswrapper[4943]: I0307 15:28:02.735453 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548288-69d2g" event={"ID":"9d10be01-3397-4d41-9c9e-c6c00c4728be","Type":"ContainerStarted","Data":"df7aef41e0729e1f3b1fbaffe039a9394c56266f485331ad304f3af676cf4301"} Mar 07 15:28:02 crc kubenswrapper[4943]: I0307 15:28:02.755904 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29548288-69d2g" podStartSLOduration=1.46764648 podStartE2EDuration="2.755881292s" podCreationTimestamp="2026-03-07 15:28:00 +0000 UTC" firstStartedPulling="2026-03-07 15:28:00.980674134 +0000 UTC m=+2922.932810672" lastFinishedPulling="2026-03-07 15:28:02.268908946 +0000 UTC m=+2924.221045484" observedRunningTime="2026-03-07 15:28:02.752781516 +0000 UTC m=+2924.704918054" watchObservedRunningTime="2026-03-07 15:28:02.755881292 +0000 UTC m=+2924.708017830" Mar 07 15:28:03 crc kubenswrapper[4943]: I0307 15:28:03.747612 4943 generic.go:334] "Generic (PLEG): container finished" podID="9d10be01-3397-4d41-9c9e-c6c00c4728be" containerID="df7aef41e0729e1f3b1fbaffe039a9394c56266f485331ad304f3af676cf4301" exitCode=0 Mar 07 15:28:03 crc kubenswrapper[4943]: I0307 15:28:03.747666 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548288-69d2g" event={"ID":"9d10be01-3397-4d41-9c9e-c6c00c4728be","Type":"ContainerDied","Data":"df7aef41e0729e1f3b1fbaffe039a9394c56266f485331ad304f3af676cf4301"} Mar 07 15:28:05 crc kubenswrapper[4943]: I0307 15:28:05.105314 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548288-69d2g" Mar 07 15:28:05 crc kubenswrapper[4943]: I0307 15:28:05.220194 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-55j48\" (UniqueName: \"kubernetes.io/projected/9d10be01-3397-4d41-9c9e-c6c00c4728be-kube-api-access-55j48\") pod \"9d10be01-3397-4d41-9c9e-c6c00c4728be\" (UID: \"9d10be01-3397-4d41-9c9e-c6c00c4728be\") " Mar 07 15:28:05 crc kubenswrapper[4943]: I0307 15:28:05.227331 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d10be01-3397-4d41-9c9e-c6c00c4728be-kube-api-access-55j48" (OuterVolumeSpecName: "kube-api-access-55j48") pod "9d10be01-3397-4d41-9c9e-c6c00c4728be" (UID: "9d10be01-3397-4d41-9c9e-c6c00c4728be"). InnerVolumeSpecName "kube-api-access-55j48". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 07 15:28:05 crc kubenswrapper[4943]: I0307 15:28:05.322528 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-55j48\" (UniqueName: \"kubernetes.io/projected/9d10be01-3397-4d41-9c9e-c6c00c4728be-kube-api-access-55j48\") on node \"crc\" DevicePath \"\"" Mar 07 15:28:05 crc kubenswrapper[4943]: I0307 15:28:05.771273 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29548288-69d2g" event={"ID":"9d10be01-3397-4d41-9c9e-c6c00c4728be","Type":"ContainerDied","Data":"c8d2f65c1c1a2b81a76b60eee905bc1223ce50d328588ed7555980de864f50c1"} Mar 07 15:28:05 crc kubenswrapper[4943]: I0307 15:28:05.771332 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c8d2f65c1c1a2b81a76b60eee905bc1223ce50d328588ed7555980de864f50c1" Mar 07 15:28:05 crc kubenswrapper[4943]: I0307 15:28:05.771356 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29548288-69d2g" Mar 07 15:28:05 crc kubenswrapper[4943]: I0307 15:28:05.840191 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29548282-tblqn"] Mar 07 15:28:05 crc kubenswrapper[4943]: I0307 15:28:05.847961 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29548282-tblqn"] Mar 07 15:28:06 crc kubenswrapper[4943]: I0307 15:28:06.073746 4943 patch_prober.go:28] interesting pod/machine-config-daemon-cpdmm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 07 15:28:06 crc kubenswrapper[4943]: I0307 15:28:06.073805 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 07 15:28:06 crc kubenswrapper[4943]: I0307 15:28:06.073850 4943 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" Mar 07 15:28:06 crc kubenswrapper[4943]: I0307 15:28:06.074481 4943 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bf223df942d1cbe9d18e7fb9dee61069a08bdf260e19d98d98a046cc56a3d466"} pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 07 15:28:06 crc kubenswrapper[4943]: I0307 15:28:06.074548 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerName="machine-config-daemon" containerID="cri-o://bf223df942d1cbe9d18e7fb9dee61069a08bdf260e19d98d98a046cc56a3d466" gracePeriod=600 Mar 07 15:28:06 crc kubenswrapper[4943]: E0307 15:28:06.206355 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cpdmm_openshift-machine-config-operator(4c1f18ea-e557-41fc-95dd-ce4e1722a61e)\"" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" Mar 07 15:28:06 crc kubenswrapper[4943]: I0307 15:28:06.773970 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43076e34-f66f-4bba-89dc-bc312cd547df" path="/var/lib/kubelet/pods/43076e34-f66f-4bba-89dc-bc312cd547df/volumes" Mar 07 15:28:06 crc kubenswrapper[4943]: I0307 15:28:06.783792 4943 generic.go:334] "Generic (PLEG): container finished" podID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" containerID="bf223df942d1cbe9d18e7fb9dee61069a08bdf260e19d98d98a046cc56a3d466" exitCode=0 Mar 07 15:28:06 crc kubenswrapper[4943]: I0307 15:28:06.783863 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" event={"ID":"4c1f18ea-e557-41fc-95dd-ce4e1722a61e","Type":"ContainerDied","Data":"bf223df942d1cbe9d18e7fb9dee61069a08bdf260e19d98d98a046cc56a3d466"} Mar 07 15:28:06 crc kubenswrapper[4943]: I0307 15:28:06.784104 4943 scope.go:117] "RemoveContainer" containerID="600a051b71dd164e5cc8344d91e9c5c354cdc3c5514cd0b1a1a3e3062c02fe1d" Mar 07 15:28:06 crc kubenswrapper[4943]: I0307 15:28:06.784906 4943 scope.go:117] "RemoveContainer" containerID="bf223df942d1cbe9d18e7fb9dee61069a08bdf260e19d98d98a046cc56a3d466" Mar 07 15:28:06 crc kubenswrapper[4943]: E0307 15:28:06.785389 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cpdmm_openshift-machine-config-operator(4c1f18ea-e557-41fc-95dd-ce4e1722a61e)\"" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" Mar 07 15:28:17 crc kubenswrapper[4943]: I0307 15:28:17.755867 4943 scope.go:117] "RemoveContainer" containerID="bf223df942d1cbe9d18e7fb9dee61069a08bdf260e19d98d98a046cc56a3d466" Mar 07 15:28:17 crc kubenswrapper[4943]: E0307 15:28:17.756843 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cpdmm_openshift-machine-config-operator(4c1f18ea-e557-41fc-95dd-ce4e1722a61e)\"" pod="openshift-machine-config-operator/machine-config-daemon-cpdmm" podUID="4c1f18ea-e557-41fc-95dd-ce4e1722a61e" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515153042045024444 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015153042046017362 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015153033746016514 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015153033746015464 5ustar corecore